[ 466.208815] env[61923]: DEBUG os_vif [-] Loaded VIF plugin class '' with name 'linux_bridge' {{(pid=61923) initialize /opt/stack/data/venv/lib/python3.10/site-packages/os_vif/__init__.py:44}} [ 466.209190] env[61923]: DEBUG os_vif [-] Loaded VIF plugin class '' with name 'noop' {{(pid=61923) initialize /opt/stack/data/venv/lib/python3.10/site-packages/os_vif/__init__.py:44}} [ 466.209271] env[61923]: DEBUG os_vif [-] Loaded VIF plugin class '' with name 'ovs' {{(pid=61923) initialize /opt/stack/data/venv/lib/python3.10/site-packages/os_vif/__init__.py:44}} [ 466.209609] env[61923]: INFO os_vif [-] Loaded VIF plugins: linux_bridge, noop, ovs [ 466.302604] env[61923]: DEBUG oslo_concurrency.processutils [-] Running cmd (subprocess): grep -F node.session.scan /sbin/iscsiadm {{(pid=61923) execute /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/processutils.py:390}} [ 466.312869] env[61923]: DEBUG oslo_concurrency.processutils [-] CMD "grep -F node.session.scan /sbin/iscsiadm" returned: 0 in 0.010s {{(pid=61923) execute /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/processutils.py:428}} [ 466.912444] env[61923]: INFO nova.virt.driver [None req-e0bfa645-9735-4523-88aa-7318f0f0f3d6 None None] Loading compute driver 'vmwareapi.VMwareVCDriver' [ 466.983314] env[61923]: DEBUG oslo_concurrency.lockutils [-] Acquiring lock "oslo_vmware_api_lock" by "oslo_vmware.api.VMwareAPISession._create_session" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 466.983455] env[61923]: DEBUG oslo_concurrency.lockutils [-] Lock "oslo_vmware_api_lock" acquired by "oslo_vmware.api.VMwareAPISession._create_session" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 466.983560] env[61923]: DEBUG oslo_vmware.service [-] Creating suds client with soap_url='https://vc1.osci.c.eu-de-1.cloud.sap:443/sdk' and wsdl_url='https://vc1.osci.c.eu-de-1.cloud.sap:443/sdk/vimService.wsdl' {{(pid=61923) __init__ /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:242}} [ 470.129163] env[61923]: DEBUG oslo_vmware.service [-] Invoking ServiceInstance.RetrieveServiceContent with opID=oslo.vmware-ebf8753f-93e1-40a6-9dd6-0f53ad837347 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 470.145248] env[61923]: DEBUG oslo_vmware.api [-] Logging into host: vc1.osci.c.eu-de-1.cloud.sap. {{(pid=61923) _create_session /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:242}} [ 470.145426] env[61923]: DEBUG oslo_vmware.service [-] Invoking SessionManager.Login with opID=oslo.vmware-f60f7986-3d6e-42c9-adb5-ca7cf1f6c940 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 470.188686] env[61923]: INFO oslo_vmware.api [-] Successfully established new session; session ID is 45381. [ 470.188834] env[61923]: DEBUG oslo_concurrency.lockutils [-] Lock "oslo_vmware_api_lock" "released" by "oslo_vmware.api.VMwareAPISession._create_session" :: held 3.205s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 470.189371] env[61923]: INFO nova.virt.vmwareapi.driver [None req-e0bfa645-9735-4523-88aa-7318f0f0f3d6 None None] VMware vCenter version: 7.0.3 [ 470.192796] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-132f96db-d1e1-4de0-9e03-0afe3b5a1dac {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 470.209948] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5378ab7-3e32-4209-810b-27bdf2057929 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 470.215662] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9205b270-795f-4c2e-b946-f504e0d6893a {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 470.221997] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81d071b0-734f-4eeb-b7c8-87732f520952 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 470.235637] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47420eac-a595-4499-b2e8-fc7477069359 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 470.241453] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ff5c4fd-8b5c-44db-905d-d6603ce9104d {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 470.270668] env[61923]: DEBUG oslo_vmware.service [-] Invoking ExtensionManager.FindExtension with opID=oslo.vmware-ab7dba14-f99b-458c-a4b7-bc114149dc16 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 470.275619] env[61923]: DEBUG nova.virt.vmwareapi.driver [None req-e0bfa645-9735-4523-88aa-7318f0f0f3d6 None None] Extension org.openstack.compute already exists. {{(pid=61923) _register_openstack_extension /opt/stack/nova/nova/virt/vmwareapi/driver.py:226}} [ 470.278371] env[61923]: INFO nova.compute.provider_config [None req-e0bfa645-9735-4523-88aa-7318f0f0f3d6 None None] No provider configs found in /etc/nova/provider_config/. If files are present, ensure the Nova process has access. [ 470.782070] env[61923]: DEBUG nova.context [None req-e0bfa645-9735-4523-88aa-7318f0f0f3d6 None None] Found 2 cells: 00000000-0000-0000-0000-000000000000(cell0),7b162bc2-542e-4fdd-be60-fad18f19221d(cell1) {{(pid=61923) load_cells /opt/stack/nova/nova/context.py:464}} [ 470.785029] env[61923]: DEBUG oslo_concurrency.lockutils [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 470.785029] env[61923]: DEBUG oslo_concurrency.lockutils [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 470.785203] env[61923]: DEBUG oslo_concurrency.lockutils [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.001s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 470.785543] env[61923]: DEBUG oslo_concurrency.lockutils [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] Acquiring lock "7b162bc2-542e-4fdd-be60-fad18f19221d" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 470.785753] env[61923]: DEBUG oslo_concurrency.lockutils [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] Lock "7b162bc2-542e-4fdd-be60-fad18f19221d" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 470.786785] env[61923]: DEBUG oslo_concurrency.lockutils [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] Lock "7b162bc2-542e-4fdd-be60-fad18f19221d" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.001s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 470.806693] env[61923]: INFO dbcounter [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] Registered counter for database nova_cell0 [ 470.815112] env[61923]: INFO dbcounter [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] Registered counter for database nova_cell1 [ 470.819053] env[61923]: DEBUG oslo_db.sqlalchemy.engines [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] MySQL server mode set to STRICT_TRANS_TABLES,STRICT_ALL_TABLES,NO_ZERO_IN_DATE,NO_ZERO_DATE,ERROR_FOR_DIVISION_BY_ZERO,TRADITIONAL,NO_AUTO_CREATE_USER,NO_ENGINE_SUBSTITUTION {{(pid=61923) _check_effective_sql_mode /opt/stack/data/venv/lib/python3.10/site-packages/oslo_db/sqlalchemy/engines.py:342}} [ 470.819418] env[61923]: DEBUG oslo_db.sqlalchemy.engines [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] MySQL server mode set to STRICT_TRANS_TABLES,STRICT_ALL_TABLES,NO_ZERO_IN_DATE,NO_ZERO_DATE,ERROR_FOR_DIVISION_BY_ZERO,TRADITIONAL,NO_AUTO_CREATE_USER,NO_ENGINE_SUBSTITUTION {{(pid=61923) _check_effective_sql_mode /opt/stack/data/venv/lib/python3.10/site-packages/oslo_db/sqlalchemy/engines.py:342}} [ 470.823966] env[61923]: ERROR nova.db.main.api [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] No DB access allowed in nova-compute: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 470.823966] env[61923]: result = function(*args, **kwargs) [ 470.823966] env[61923]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 470.823966] env[61923]: return func(*args, **kwargs) [ 470.823966] env[61923]: File "/opt/stack/nova/nova/context.py", line 422, in gather_result [ 470.823966] env[61923]: result = fn(*args, **kwargs) [ 470.823966] env[61923]: File "/opt/stack/nova/nova/db/main/api.py", line 179, in wrapper [ 470.823966] env[61923]: return f(*args, **kwargs) [ 470.823966] env[61923]: File "/opt/stack/nova/nova/objects/service.py", line 554, in _db_service_get_minimum_version [ 470.823966] env[61923]: return db.service_get_minimum_version(context, binaries) [ 470.823966] env[61923]: File "/opt/stack/nova/nova/db/main/api.py", line 238, in wrapper [ 470.823966] env[61923]: _check_db_access() [ 470.823966] env[61923]: File "/opt/stack/nova/nova/db/main/api.py", line 188, in _check_db_access [ 470.823966] env[61923]: stacktrace = ''.join(traceback.format_stack()) [ 470.823966] env[61923]: [ 470.825292] env[61923]: ERROR nova.db.main.api [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] No DB access allowed in nova-compute: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 470.825292] env[61923]: result = function(*args, **kwargs) [ 470.825292] env[61923]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 470.825292] env[61923]: return func(*args, **kwargs) [ 470.825292] env[61923]: File "/opt/stack/nova/nova/context.py", line 422, in gather_result [ 470.825292] env[61923]: result = fn(*args, **kwargs) [ 470.825292] env[61923]: File "/opt/stack/nova/nova/db/main/api.py", line 179, in wrapper [ 470.825292] env[61923]: return f(*args, **kwargs) [ 470.825292] env[61923]: File "/opt/stack/nova/nova/objects/service.py", line 554, in _db_service_get_minimum_version [ 470.825292] env[61923]: return db.service_get_minimum_version(context, binaries) [ 470.825292] env[61923]: File "/opt/stack/nova/nova/db/main/api.py", line 238, in wrapper [ 470.825292] env[61923]: _check_db_access() [ 470.825292] env[61923]: File "/opt/stack/nova/nova/db/main/api.py", line 188, in _check_db_access [ 470.825292] env[61923]: stacktrace = ''.join(traceback.format_stack()) [ 470.825292] env[61923]: [ 470.825718] env[61923]: WARNING nova.objects.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] Failed to get minimum service version for cell 7b162bc2-542e-4fdd-be60-fad18f19221d [ 470.825859] env[61923]: WARNING nova.objects.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] Failed to get minimum service version for cell 00000000-0000-0000-0000-000000000000 [ 470.826309] env[61923]: DEBUG oslo_concurrency.lockutils [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] Acquiring lock "singleton_lock" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 470.826472] env[61923]: DEBUG oslo_concurrency.lockutils [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] Acquired lock "singleton_lock" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 470.826747] env[61923]: DEBUG oslo_concurrency.lockutils [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] Releasing lock "singleton_lock" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 470.827082] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] Full set of CONF: {{(pid=61923) _wait_for_exit_or_signal /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/service.py:363}} [ 470.827234] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] ******************************************************************************** {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2806}} [ 470.827362] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] Configuration options gathered from: {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2807}} [ 470.827499] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] command line args: ['--config-file', '/etc/nova/nova.conf', '--config-file', '/etc/nova/nova-cpu-common.conf', '--config-file', '/etc/nova/nova-cpu-1.conf'] {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2808}} [ 470.827694] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] config files: ['/etc/nova/nova.conf', '/etc/nova/nova-cpu-common.conf', '/etc/nova/nova-cpu-1.conf'] {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2809}} [ 470.827825] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] ================================================================================ {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2811}} [ 470.828048] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] allow_resize_to_same_host = True {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.828229] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] arq_binding_timeout = 300 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.828366] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] backdoor_port = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.828495] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] backdoor_socket = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.828665] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] block_device_allocate_retries = 60 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.828830] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] block_device_allocate_retries_interval = 3 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.829007] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] cert = self.pem {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.829187] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] compute_driver = vmwareapi.VMwareVCDriver {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.829360] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] compute_monitors = [] {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.829535] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] config_dir = [] {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.829697] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] config_drive_format = iso9660 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.829831] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] config_file = ['/etc/nova/nova.conf', '/etc/nova/nova-cpu-common.conf', '/etc/nova/nova-cpu-1.conf'] {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.829995] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] config_source = [] {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.830172] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] console_host = devstack {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.830339] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] control_exchange = nova {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.830500] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] cpu_allocation_ratio = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.830662] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] daemon = False {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.830830] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] debug = True {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.830990] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] default_access_ip_network_name = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.831169] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] default_availability_zone = nova {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.831326] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] default_ephemeral_format = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.831488] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] default_green_pool_size = 1000 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.831726] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] default_log_levels = ['amqp=WARN', 'amqplib=WARN', 'boto=WARN', 'qpid=WARN', 'sqlalchemy=WARN', 'suds=INFO', 'oslo.messaging=INFO', 'oslo_messaging=INFO', 'iso8601=WARN', 'requests.packages.urllib3.connectionpool=WARN', 'urllib3.connectionpool=WARN', 'websocket=WARN', 'requests.packages.urllib3.util.retry=WARN', 'urllib3.util.retry=WARN', 'keystonemiddleware=WARN', 'routes.middleware=WARN', 'stevedore=WARN', 'taskflow=WARN', 'keystoneauth=WARN', 'oslo.cache=INFO', 'oslo_policy=INFO', 'dogpile.core.dogpile=INFO', 'glanceclient=WARN', 'oslo.privsep.daemon=INFO'] {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.831892] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] default_schedule_zone = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.832062] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] disk_allocation_ratio = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.832228] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] enable_new_services = True {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.832408] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] enabled_apis = ['osapi_compute'] {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.832573] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] enabled_ssl_apis = [] {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.832734] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] flat_injected = False {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.832890] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] force_config_drive = False {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.833064] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] force_raw_images = True {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.833244] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] graceful_shutdown_timeout = 5 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.833437] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] heal_instance_info_cache_interval = 60 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.833663] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] host = cpu-1 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.833846] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] initial_cpu_allocation_ratio = 4.0 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.834025] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] initial_disk_allocation_ratio = 1.0 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.834199] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] initial_ram_allocation_ratio = 1.0 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.834413] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] injected_network_template = /opt/stack/nova/nova/virt/interfaces.template {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.834582] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] instance_build_timeout = 0 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.834746] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] instance_delete_interval = 300 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.834912] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] instance_format = [instance: %(uuid)s] {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.835090] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] instance_name_template = instance-%08x {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.835258] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] instance_usage_audit = False {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.835445] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] instance_usage_audit_period = month {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.835631] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] instance_uuid_format = [instance: %(uuid)s] {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.835801] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] instances_path = /opt/stack/data/nova/instances {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.835969] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] internal_service_availability_zone = internal {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.836145] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] key = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.836322] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] live_migration_retry_count = 30 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.836473] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] log_color = False {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.836690] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] log_config_append = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.836873] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] log_date_format = %Y-%m-%d %H:%M:%S {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.837045] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] log_dir = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.837211] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] log_file = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.837342] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] log_options = True {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.837504] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] log_rotate_interval = 1 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.837672] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] log_rotate_interval_type = days {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.837835] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] log_rotation_type = none {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.837966] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] logging_context_format_string = %(color)s%(levelname)s %(name)s [%(global_request_id)s %(request_id)s %(project_name)s %(user_name)s%(color)s] %(instance)s%(color)s%(message)s {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.838106] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] logging_debug_format_suffix = {{(pid=%(process)d) %(funcName)s %(pathname)s:%(lineno)d}} {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.838283] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] logging_default_format_string = %(color)s%(levelname)s %(name)s [-%(color)s] %(instance)s%(color)s%(message)s {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.838447] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] logging_exception_prefix = ERROR %(name)s %(instance)s {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.838577] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] logging_user_identity_format = %(user)s %(project)s %(domain)s %(system_scope)s %(user_domain)s %(project_domain)s {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.838750] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] long_rpc_timeout = 1800 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.838913] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] max_concurrent_builds = 10 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.839090] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] max_concurrent_live_migrations = 1 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.839255] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] max_concurrent_snapshots = 5 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.839419] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] max_local_block_devices = 3 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.839582] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] max_logfile_count = 30 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.839745] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] max_logfile_size_mb = 200 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.839906] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] maximum_instance_delete_attempts = 5 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.840084] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] metadata_listen = 0.0.0.0 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.840266] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] metadata_listen_port = 8775 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.840421] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] metadata_workers = 2 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.840584] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] migrate_max_retries = -1 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.840748] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] mkisofs_cmd = genisoimage {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.840955] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] my_block_storage_ip = 10.180.1.21 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.841103] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] my_ip = 10.180.1.21 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.841275] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] network_allocate_retries = 0 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.841453] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] non_inheritable_image_properties = ['cache_in_nova', 'bittorrent'] {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.841622] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] osapi_compute_listen = 0.0.0.0 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.841784] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] osapi_compute_listen_port = 8774 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.841951] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] osapi_compute_unique_server_name_scope = {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.842132] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] osapi_compute_workers = 2 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.842299] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] password_length = 12 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.842459] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] periodic_enable = True {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.842624] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] periodic_fuzzy_delay = 60 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.842786] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] pointer_model = usbtablet {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.842947] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] preallocate_images = none {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.843119] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] publish_errors = False {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.843248] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] pybasedir = /opt/stack/nova {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.843433] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] ram_allocation_ratio = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.843607] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] rate_limit_burst = 0 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.843773] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] rate_limit_except_level = CRITICAL {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.843935] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] rate_limit_interval = 0 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.844110] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] reboot_timeout = 0 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.844273] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] reclaim_instance_interval = 0 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.844429] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] record = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.844601] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] reimage_timeout_per_gb = 60 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.844767] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] report_interval = 120 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.844929] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] rescue_timeout = 0 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.845101] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] reserved_host_cpus = 0 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.845265] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] reserved_host_disk_mb = 0 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.845450] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] reserved_host_memory_mb = 512 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.845623] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] reserved_huge_pages = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.845789] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] resize_confirm_window = 0 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.845949] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] resize_fs_using_block_device = False {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.846124] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] resume_guests_state_on_host_boot = False {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.846294] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] rootwrap_config = /etc/nova/rootwrap.conf {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.846462] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] rpc_response_timeout = 60 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.846646] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] run_external_periodic_tasks = True {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.846820] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] running_deleted_instance_action = reap {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.846985] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] running_deleted_instance_poll_interval = 1800 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.847160] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] running_deleted_instance_timeout = 0 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.847320] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] scheduler_instance_sync_interval = 120 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.847489] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] service_down_time = 720 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.847660] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] servicegroup_driver = db {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.847818] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] shell_completion = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.847979] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] shelved_offload_time = 0 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.848156] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] shelved_poll_interval = 3600 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.848327] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] shutdown_timeout = 0 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.848492] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] source_is_ipv6 = False {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.848651] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] ssl_only = False {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.848896] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] state_path = /opt/stack/data/n-cpu-1 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.849080] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] sync_power_state_interval = 600 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.849248] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] sync_power_state_pool_size = 1000 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.849417] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] syslog_log_facility = LOG_USER {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.849577] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] tempdir = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.849739] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] timeout_nbd = 10 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.849907] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] transport_url = **** {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.850082] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] update_resources_interval = 0 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.850246] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] use_cow_images = True {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.850407] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] use_eventlog = False {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.850569] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] use_journal = False {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.850728] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] use_json = False {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.850885] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] use_rootwrap_daemon = False {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.851054] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] use_stderr = False {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.851218] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] use_syslog = False {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.851373] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] vcpu_pin_set = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.851538] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] vif_plugging_is_fatal = True {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.851706] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] vif_plugging_timeout = 300 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.851871] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] virt_mkfs = [] {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.852047] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] volume_usage_poll_interval = 0 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.852212] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] watch_log_file = False {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.852380] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] web = /usr/share/spice-html5 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 470.852568] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] os_brick.lock_path = /opt/stack/data/n-cpu-1 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.852742] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] os_brick.wait_mpath_device_attempts = 4 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.852907] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] os_brick.wait_mpath_device_interval = 1 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.853090] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] oslo_concurrency.disable_process_locking = False {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.853689] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] oslo_concurrency.lock_path = /opt/stack/data/n-cpu-1 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.853890] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] oslo_messaging_metrics.metrics_buffer_size = 1000 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.854082] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] oslo_messaging_metrics.metrics_enabled = False {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.854275] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] oslo_messaging_metrics.metrics_process_name = {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.854452] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] oslo_messaging_metrics.metrics_socket_file = /var/tmp/metrics_collector.sock {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.854625] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] oslo_messaging_metrics.metrics_thread_stop_timeout = 10 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.854811] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] api.auth_strategy = keystone {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.854983] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] api.compute_link_prefix = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.855178] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] api.config_drive_skip_versions = 1.0 2007-01-19 2007-03-01 2007-08-29 2007-10-10 2007-12-15 2008-02-01 2008-09-01 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.855358] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] api.dhcp_domain = novalocal {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.855570] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] api.enable_instance_password = True {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.855755] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] api.glance_link_prefix = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.855932] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] api.instance_list_cells_batch_fixed_size = 100 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.856124] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] api.instance_list_cells_batch_strategy = distributed {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.856295] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] api.instance_list_per_project_cells = False {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.856484] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] api.list_records_by_skipping_down_cells = True {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.856694] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] api.local_metadata_per_cell = False {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.856874] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] api.max_limit = 1000 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.857064] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] api.metadata_cache_expiration = 15 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.857244] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] api.neutron_default_tenant_id = default {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.857419] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] api.response_validation = warn {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.857595] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] api.use_neutron_default_nets = False {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.857769] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] api.vendordata_dynamic_connect_timeout = 5 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.857935] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] api.vendordata_dynamic_failure_fatal = False {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.858122] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] api.vendordata_dynamic_read_timeout = 5 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.858301] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] api.vendordata_dynamic_ssl_certfile = {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.858478] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] api.vendordata_dynamic_targets = [] {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.858646] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] api.vendordata_jsonfile_path = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.859081] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] api.vendordata_providers = ['StaticJSON'] {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.859081] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] cache.backend = dogpile.cache.memcached {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.859191] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] cache.backend_argument = **** {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.859358] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] cache.config_prefix = cache.oslo {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.859531] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] cache.dead_timeout = 60.0 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.859701] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] cache.debug_cache_backend = False {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.859866] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] cache.enable_retry_client = False {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.860040] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] cache.enable_socket_keepalive = False {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.860322] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] cache.enabled = True {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.860382] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] cache.enforce_fips_mode = False {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.860542] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] cache.expiration_time = 600 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.860712] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] cache.hashclient_retry_attempts = 2 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.860878] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] cache.hashclient_retry_delay = 1.0 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.861056] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] cache.memcache_dead_retry = 300 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.861221] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] cache.memcache_password = **** {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.861391] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] cache.memcache_pool_connection_get_timeout = 10 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.861555] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] cache.memcache_pool_flush_on_reconnect = False {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.861722] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] cache.memcache_pool_maxsize = 10 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.861884] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] cache.memcache_pool_unused_timeout = 60 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.862061] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] cache.memcache_sasl_enabled = False {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.862246] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] cache.memcache_servers = ['localhost:11211'] {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.862417] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] cache.memcache_socket_timeout = 1.0 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.862580] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] cache.memcache_username = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.862751] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] cache.proxies = [] {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.862916] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] cache.redis_db = 0 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.863094] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] cache.redis_password = **** {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.863269] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] cache.redis_sentinel_service_name = mymaster {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.863480] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] cache.redis_sentinels = ['localhost:26379'] {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.863659] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] cache.redis_server = localhost:6379 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.863827] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] cache.redis_socket_timeout = 1.0 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.863987] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] cache.redis_username = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.864169] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] cache.retry_attempts = 2 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.864335] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] cache.retry_delay = 0.0 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.864502] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] cache.socket_keepalive_count = 1 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.864667] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] cache.socket_keepalive_idle = 1 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.864833] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] cache.socket_keepalive_interval = 1 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.864992] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] cache.tls_allowed_ciphers = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.865170] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] cache.tls_cafile = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.865330] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] cache.tls_certfile = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.865524] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] cache.tls_enabled = False {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.865695] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] cache.tls_keyfile = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.865875] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] cinder.auth_section = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.866124] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] cinder.auth_type = password {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.866303] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] cinder.cafile = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.866484] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] cinder.catalog_info = volumev3::publicURL {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.866678] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] cinder.certfile = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.866849] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] cinder.collect_timing = False {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.867029] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] cinder.cross_az_attach = True {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.867196] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] cinder.debug = False {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.867357] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] cinder.endpoint_template = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.867525] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] cinder.http_retries = 3 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.867690] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] cinder.insecure = False {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.867849] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] cinder.keyfile = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.868029] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] cinder.os_region_name = RegionOne {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.868202] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] cinder.split_loggers = False {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.868368] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] cinder.timeout = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.868541] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] compute.consecutive_build_service_disable_threshold = 10 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.868705] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] compute.cpu_dedicated_set = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.868865] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] compute.cpu_shared_set = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.869045] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] compute.image_type_exclude_list = [] {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.869216] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] compute.live_migration_wait_for_vif_plug = True {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.869383] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] compute.max_concurrent_disk_ops = 0 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.869547] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] compute.max_disk_devices_to_attach = -1 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.869715] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] compute.packing_host_numa_cells_allocation_strategy = False {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.869883] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] compute.provider_config_location = /etc/nova/provider_config/ {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.870058] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] compute.resource_provider_association_refresh = 300 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.870228] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] compute.sharing_providers_max_uuids_per_request = 200 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.870393] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] compute.shutdown_retry_interval = 10 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.870577] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] compute.vmdk_allowed_types = ['streamOptimized', 'monolithicSparse'] {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.870761] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] conductor.workers = 2 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.870941] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] console.allowed_origins = [] {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.871120] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] console.ssl_ciphers = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.871296] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] console.ssl_minimum_version = default {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.871467] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] consoleauth.enforce_session_timeout = False {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.871640] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] consoleauth.token_ttl = 600 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.871814] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] cyborg.cafile = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.871973] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] cyborg.certfile = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.872154] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] cyborg.collect_timing = False {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.872316] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] cyborg.connect_retries = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.872477] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] cyborg.connect_retry_delay = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.872638] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] cyborg.endpoint_override = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.872799] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] cyborg.insecure = False {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.872959] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] cyborg.keyfile = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.873134] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] cyborg.max_version = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.873297] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] cyborg.min_version = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.873487] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] cyborg.region_name = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.873656] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] cyborg.retriable_status_codes = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.873816] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] cyborg.service_name = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.873991] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] cyborg.service_type = accelerator {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.874167] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] cyborg.split_loggers = False {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.874332] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] cyborg.status_code_retries = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.874492] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] cyborg.status_code_retry_delay = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.874655] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] cyborg.timeout = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.874836] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] cyborg.valid_interfaces = ['internal', 'public'] {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.874998] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] cyborg.version = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.875197] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] database.backend = sqlalchemy {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.875373] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] database.connection = **** {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.875574] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] database.connection_debug = 0 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.875755] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] database.connection_parameters = {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.875925] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] database.connection_recycle_time = 3600 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.876103] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] database.connection_trace = False {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.876271] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] database.db_inc_retry_interval = True {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.876439] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] database.db_max_retries = 20 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.876629] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] database.db_max_retry_interval = 10 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.876804] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] database.db_retry_interval = 1 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.876969] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] database.max_overflow = 50 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.877149] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] database.max_pool_size = 5 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.877316] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] database.max_retries = 10 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.877488] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] database.mysql_sql_mode = TRADITIONAL {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.877650] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] database.mysql_wsrep_sync_wait = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.877810] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] database.pool_timeout = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.877975] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] database.retry_interval = 10 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.878150] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] database.slave_connection = **** {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.878317] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] database.sqlite_synchronous = True {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.878477] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] database.use_db_reconnect = False {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.878655] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] api_database.backend = sqlalchemy {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.878825] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] api_database.connection = **** {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.878992] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] api_database.connection_debug = 0 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.879179] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] api_database.connection_parameters = {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.879346] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] api_database.connection_recycle_time = 3600 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.879512] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] api_database.connection_trace = False {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.879676] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] api_database.db_inc_retry_interval = True {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.879843] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] api_database.db_max_retries = 20 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.880020] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] api_database.db_max_retry_interval = 10 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.880188] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] api_database.db_retry_interval = 1 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.880354] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] api_database.max_overflow = 50 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.880520] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] api_database.max_pool_size = 5 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.880688] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] api_database.max_retries = 10 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.880860] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] api_database.mysql_sql_mode = TRADITIONAL {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.881035] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] api_database.mysql_wsrep_sync_wait = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.881205] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] api_database.pool_timeout = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.881370] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] api_database.retry_interval = 10 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.881530] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] api_database.slave_connection = **** {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.881695] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] api_database.sqlite_synchronous = True {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.881871] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] devices.enabled_mdev_types = [] {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.882062] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] ephemeral_storage_encryption.cipher = aes-xts-plain64 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.882241] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] ephemeral_storage_encryption.default_format = luks {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.882406] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] ephemeral_storage_encryption.enabled = False {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.882574] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] ephemeral_storage_encryption.key_size = 512 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.882753] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] glance.api_servers = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.882979] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] glance.cafile = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.883176] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] glance.certfile = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.883368] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] glance.collect_timing = False {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.883549] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] glance.connect_retries = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.883729] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] glance.connect_retry_delay = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.883880] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] glance.debug = False {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.884057] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] glance.default_trusted_certificate_ids = [] {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.884231] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] glance.enable_certificate_validation = False {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.884396] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] glance.enable_rbd_download = False {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.884567] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] glance.endpoint_override = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.884749] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] glance.insecure = False {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.884915] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] glance.keyfile = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.885092] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] glance.max_version = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.885258] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] glance.min_version = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.885439] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] glance.num_retries = 3 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.885636] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] glance.rbd_ceph_conf = {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.885807] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] glance.rbd_connect_timeout = 5 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.885974] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] glance.rbd_pool = {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.886160] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] glance.rbd_user = {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.886325] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] glance.region_name = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.886488] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] glance.retriable_status_codes = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.886670] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] glance.service_name = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.886843] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] glance.service_type = image {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.887016] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] glance.split_loggers = False {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.887192] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] glance.status_code_retries = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.887354] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] glance.status_code_retry_delay = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.887516] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] glance.timeout = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.887700] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] glance.valid_interfaces = ['internal', 'public'] {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.887867] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] glance.verify_glance_signatures = False {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.888041] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] glance.version = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.888215] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] guestfs.debug = False {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.888383] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] mks.enabled = False {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.888746] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] mks.mksproxy_base_url = http://127.0.0.1:6090/ {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.888940] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] image_cache.manager_interval = 2400 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.889129] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] image_cache.precache_concurrency = 1 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.889303] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] image_cache.remove_unused_base_images = True {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.889475] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] image_cache.remove_unused_original_minimum_age_seconds = 86400 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.889649] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] image_cache.remove_unused_resized_minimum_age_seconds = 3600 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.889829] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] image_cache.subdirectory_name = _base {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.890015] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] ironic.api_max_retries = 60 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.890197] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] ironic.api_retry_interval = 2 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.890361] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] ironic.auth_section = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.890526] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] ironic.auth_type = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.890692] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] ironic.cafile = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.890846] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] ironic.certfile = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.891015] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] ironic.collect_timing = False {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.891188] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] ironic.conductor_group = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.891352] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] ironic.connect_retries = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.891518] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] ironic.connect_retry_delay = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.891679] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] ironic.endpoint_override = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.891844] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] ironic.insecure = False {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.892014] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] ironic.keyfile = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.892183] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] ironic.max_version = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.892345] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] ironic.min_version = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.892515] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] ironic.peer_list = [] {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.892679] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] ironic.region_name = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.892841] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] ironic.retriable_status_codes = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.893015] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] ironic.serial_console_state_timeout = 10 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.893190] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] ironic.service_name = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.893387] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] ironic.service_type = baremetal {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.893567] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] ironic.shard = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.893735] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] ironic.split_loggers = False {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.893896] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] ironic.status_code_retries = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.894072] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] ironic.status_code_retry_delay = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.894242] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] ironic.timeout = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.894424] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] ironic.valid_interfaces = ['internal', 'public'] {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.894595] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] ironic.version = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.894775] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] key_manager.backend = nova.keymgr.conf_key_mgr.ConfKeyManager {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.894950] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] key_manager.fixed_key = **** {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.895151] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] barbican.auth_endpoint = http://localhost/identity/v3 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.895319] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] barbican.barbican_api_version = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.895502] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] barbican.barbican_endpoint = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.895689] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] barbican.barbican_endpoint_type = public {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.895852] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] barbican.barbican_region_name = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.896019] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] barbican.cafile = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.896189] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] barbican.certfile = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.896353] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] barbican.collect_timing = False {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.896526] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] barbican.insecure = False {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.896719] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] barbican.keyfile = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.896891] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] barbican.number_of_retries = 60 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.897068] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] barbican.retry_delay = 1 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.897239] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] barbican.send_service_user_token = False {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.897402] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] barbican.split_loggers = False {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.897568] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] barbican.timeout = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.897728] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] barbican.verify_ssl = True {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.897890] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] barbican.verify_ssl_path = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.898062] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] barbican_service_user.auth_section = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.898233] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] barbican_service_user.auth_type = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.898394] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] barbican_service_user.cafile = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.898551] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] barbican_service_user.certfile = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.898717] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] barbican_service_user.collect_timing = False {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.898879] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] barbican_service_user.insecure = False {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.899073] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] barbican_service_user.keyfile = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.899245] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] barbican_service_user.split_loggers = False {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.899409] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] barbican_service_user.timeout = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.899579] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] vault.approle_role_id = **** {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.899739] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] vault.approle_secret_id = **** {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.899908] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] vault.kv_mountpoint = secret {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.900081] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] vault.kv_path = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.900253] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] vault.kv_version = 2 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.900413] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] vault.namespace = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.900577] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] vault.root_token_id = **** {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.900733] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] vault.ssl_ca_crt_file = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.900903] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] vault.timeout = 60.0 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.901078] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] vault.use_ssl = False {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.901259] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] vault.vault_url = http://127.0.0.1:8200 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.901435] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] keystone.auth_section = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.901603] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] keystone.auth_type = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.901765] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] keystone.cafile = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.901927] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] keystone.certfile = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.902106] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] keystone.collect_timing = False {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.902271] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] keystone.connect_retries = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.902432] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] keystone.connect_retry_delay = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.902595] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] keystone.endpoint_override = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.902758] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] keystone.insecure = False {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.902917] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] keystone.keyfile = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.903089] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] keystone.max_version = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.903253] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] keystone.min_version = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.903446] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] keystone.region_name = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.903620] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] keystone.retriable_status_codes = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.903779] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] keystone.service_name = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.903948] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] keystone.service_type = identity {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.904129] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] keystone.split_loggers = False {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.904294] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] keystone.status_code_retries = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.904454] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] keystone.status_code_retry_delay = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.904616] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] keystone.timeout = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.904798] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] keystone.valid_interfaces = ['internal', 'public'] {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.904962] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] keystone.version = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.905183] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] libvirt.connection_uri = {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.905348] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] libvirt.cpu_mode = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.905542] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] libvirt.cpu_model_extra_flags = [] {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.905729] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] libvirt.cpu_models = [] {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.905904] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] libvirt.cpu_power_governor_high = performance {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.906086] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] libvirt.cpu_power_governor_low = powersave {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.906257] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] libvirt.cpu_power_management = False {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.906435] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] libvirt.cpu_power_management_strategy = cpu_state {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.906626] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] libvirt.device_detach_attempts = 8 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.906800] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] libvirt.device_detach_timeout = 20 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.906969] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] libvirt.disk_cachemodes = [] {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.907145] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] libvirt.disk_prefix = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.907315] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] libvirt.enabled_perf_events = [] {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.907482] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] libvirt.file_backed_memory = 0 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.907653] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] libvirt.gid_maps = [] {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.907817] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] libvirt.hw_disk_discard = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.907975] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] libvirt.hw_machine_type = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.908162] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] libvirt.images_rbd_ceph_conf = {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.908336] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] libvirt.images_rbd_glance_copy_poll_interval = 15 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.908504] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] libvirt.images_rbd_glance_copy_timeout = 600 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.908676] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] libvirt.images_rbd_glance_store_name = {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.908847] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] libvirt.images_rbd_pool = rbd {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.909029] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] libvirt.images_type = default {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.909198] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] libvirt.images_volume_group = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.909367] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] libvirt.inject_key = False {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.909531] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] libvirt.inject_partition = -2 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.909696] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] libvirt.inject_password = False {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.909861] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] libvirt.iscsi_iface = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.910034] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] libvirt.iser_use_multipath = False {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.910208] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] libvirt.live_migration_bandwidth = 0 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.910376] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] libvirt.live_migration_completion_timeout = 800 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.910543] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] libvirt.live_migration_downtime = 500 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.910709] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] libvirt.live_migration_downtime_delay = 75 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.910877] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] libvirt.live_migration_downtime_steps = 10 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.911048] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] libvirt.live_migration_inbound_addr = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.911218] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] libvirt.live_migration_permit_auto_converge = False {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.911386] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] libvirt.live_migration_permit_post_copy = False {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.911550] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] libvirt.live_migration_scheme = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.911735] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] libvirt.live_migration_timeout_action = abort {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.911904] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] libvirt.live_migration_tunnelled = False {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.912085] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] libvirt.live_migration_uri = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.912256] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] libvirt.live_migration_with_native_tls = False {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.912422] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] libvirt.max_queues = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.912589] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] libvirt.mem_stats_period_seconds = 10 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.912826] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] libvirt.migration_inbound_addr = 10.180.1.21 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.912992] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] libvirt.nfs_mount_options = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.913313] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] libvirt.nfs_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.913516] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] libvirt.num_aoe_discover_tries = 3 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.913694] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] libvirt.num_iser_scan_tries = 5 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.913861] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] libvirt.num_memory_encrypted_guests = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.914058] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] libvirt.num_nvme_discover_tries = 5 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.914243] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] libvirt.num_pcie_ports = 0 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.914416] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] libvirt.num_volume_scan_tries = 5 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.914588] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] libvirt.pmem_namespaces = [] {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.914751] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] libvirt.quobyte_client_cfg = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.915059] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] libvirt.quobyte_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.915247] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] libvirt.rbd_connect_timeout = 5 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.915434] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] libvirt.rbd_destroy_volume_retries = 12 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.915614] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] libvirt.rbd_destroy_volume_retry_interval = 5 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.915781] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] libvirt.rbd_secret_uuid = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.915942] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] libvirt.rbd_user = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.916126] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] libvirt.realtime_scheduler_priority = 1 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.916301] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] libvirt.remote_filesystem_transport = ssh {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.916466] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] libvirt.rescue_image_id = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.916652] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] libvirt.rescue_kernel_id = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.916819] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] libvirt.rescue_ramdisk_id = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.916987] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] libvirt.rng_dev_path = /dev/urandom {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.917168] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] libvirt.rx_queue_size = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.917339] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] libvirt.smbfs_mount_options = {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.917618] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] libvirt.smbfs_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.917794] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] libvirt.snapshot_compression = False {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.917957] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] libvirt.snapshot_image_format = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.918208] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] libvirt.snapshots_directory = /opt/stack/data/nova/instances/snapshots {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.918357] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] libvirt.sparse_logical_volumes = False {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.918519] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] libvirt.swtpm_enabled = False {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.918692] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] libvirt.swtpm_group = tss {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.918859] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] libvirt.swtpm_user = tss {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.919048] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] libvirt.sysinfo_serial = unique {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.919206] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] libvirt.tb_cache_size = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.919368] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] libvirt.tx_queue_size = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.919539] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] libvirt.uid_maps = [] {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.919729] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] libvirt.use_virtio_for_bridges = True {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.919911] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] libvirt.virt_type = kvm {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.920094] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] libvirt.volume_clear = zero {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.920264] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] libvirt.volume_clear_size = 0 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.920433] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] libvirt.volume_use_multipath = False {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.920593] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] libvirt.vzstorage_cache_path = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.920762] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] libvirt.vzstorage_log_path = /var/log/vstorage/%(cluster_name)s/nova.log.gz {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.920927] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] libvirt.vzstorage_mount_group = qemu {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.921114] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] libvirt.vzstorage_mount_opts = [] {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.921288] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] libvirt.vzstorage_mount_perms = 0770 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.921562] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] libvirt.vzstorage_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.921742] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] libvirt.vzstorage_mount_user = stack {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.921911] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] libvirt.wait_soft_reboot_seconds = 120 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.922100] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] neutron.auth_section = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.922282] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] neutron.auth_type = password {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.922444] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] neutron.cafile = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.922605] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] neutron.certfile = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.922768] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] neutron.collect_timing = False {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.922927] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] neutron.connect_retries = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.923099] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] neutron.connect_retry_delay = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.923271] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] neutron.default_floating_pool = public {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.923456] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] neutron.endpoint_override = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.923661] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] neutron.extension_sync_interval = 600 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.923836] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] neutron.http_retries = 3 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.924045] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] neutron.insecure = False {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.924225] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] neutron.keyfile = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.924390] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] neutron.max_version = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.924559] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] neutron.metadata_proxy_shared_secret = **** {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.924720] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] neutron.min_version = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.924887] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] neutron.ovs_bridge = br-int {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.925065] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] neutron.physnets = [] {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.925239] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] neutron.region_name = RegionOne {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.925399] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] neutron.retriable_status_codes = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.925589] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] neutron.service_metadata_proxy = True {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.925754] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] neutron.service_name = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.925922] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] neutron.service_type = network {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.926099] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] neutron.split_loggers = False {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.926264] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] neutron.status_code_retries = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.926423] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] neutron.status_code_retry_delay = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.926606] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] neutron.timeout = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.926849] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] neutron.valid_interfaces = ['internal', 'public'] {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.927081] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] neutron.version = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.927287] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] notifications.bdms_in_notifications = False {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.927473] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] notifications.default_level = INFO {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.927652] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] notifications.notification_format = unversioned {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.927834] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] notifications.notify_on_state_change = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.928052] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] notifications.versioned_notifications_topics = ['versioned_notifications'] {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.928315] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] pci.alias = [] {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.928433] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] pci.device_spec = [] {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.928600] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] pci.report_in_placement = False {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.928772] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] placement.auth_section = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.928946] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] placement.auth_type = password {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.929130] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] placement.auth_url = http://10.180.1.21/identity {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.929291] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] placement.cafile = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.929449] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] placement.certfile = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.929623] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] placement.collect_timing = False {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.929773] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] placement.connect_retries = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.929932] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] placement.connect_retry_delay = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.930103] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] placement.default_domain_id = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.930266] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] placement.default_domain_name = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.930425] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] placement.domain_id = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.930584] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] placement.domain_name = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.930740] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] placement.endpoint_override = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.930901] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] placement.insecure = False {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.931069] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] placement.keyfile = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.931231] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] placement.max_version = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.931387] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] placement.min_version = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.931553] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] placement.password = **** {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.931713] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] placement.project_domain_id = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.931877] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] placement.project_domain_name = Default {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.932059] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] placement.project_id = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.932238] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] placement.project_name = service {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.932406] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] placement.region_name = RegionOne {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.932570] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] placement.retriable_status_codes = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.932728] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] placement.service_name = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.932895] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] placement.service_type = placement {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.933069] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] placement.split_loggers = False {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.933229] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] placement.status_code_retries = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.933417] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] placement.status_code_retry_delay = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.933589] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] placement.system_scope = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.933752] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] placement.timeout = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.933910] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] placement.trust_id = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.934081] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] placement.user_domain_id = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.934279] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] placement.user_domain_name = Default {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.934444] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] placement.user_id = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.934636] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] placement.username = nova {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.934835] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] placement.valid_interfaces = ['internal', 'public'] {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.935019] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] placement.version = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.935199] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] quota.cores = 20 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.935364] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] quota.count_usage_from_placement = False {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.935562] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] quota.driver = nova.quota.DbQuotaDriver {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.935749] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] quota.injected_file_content_bytes = 10240 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.935918] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] quota.injected_file_path_length = 255 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.936100] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] quota.injected_files = 5 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.936272] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] quota.instances = 10 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.936441] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] quota.key_pairs = 100 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.936638] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] quota.metadata_items = 128 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.936816] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] quota.ram = 51200 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.936982] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] quota.recheck_quota = True {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.937170] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] quota.server_group_members = 10 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.937340] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] quota.server_groups = 10 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.937516] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] scheduler.discover_hosts_in_cells_interval = -1 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.937686] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] scheduler.enable_isolated_aggregate_filtering = False {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.937850] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] scheduler.image_metadata_prefilter = False {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.938021] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] scheduler.limit_tenants_to_placement_aggregate = False {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.938195] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] scheduler.max_attempts = 3 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.938366] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] scheduler.max_placement_results = 1000 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.938529] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] scheduler.placement_aggregate_required_for_tenants = False {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.938692] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] scheduler.query_placement_for_image_type_support = False {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.938853] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] scheduler.query_placement_for_routed_network_aggregates = False {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.939041] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] scheduler.workers = 2 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.939222] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] filter_scheduler.aggregate_image_properties_isolation_namespace = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.939396] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] filter_scheduler.aggregate_image_properties_isolation_separator = . {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.939578] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] filter_scheduler.available_filters = ['nova.scheduler.filters.all_filters'] {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.939752] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] filter_scheduler.build_failure_weight_multiplier = 1000000.0 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.939917] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] filter_scheduler.cpu_weight_multiplier = 1.0 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.940096] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] filter_scheduler.cross_cell_move_weight_multiplier = 1000000.0 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.940266] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] filter_scheduler.disk_weight_multiplier = 1.0 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.940457] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] filter_scheduler.enabled_filters = ['ComputeFilter', 'ComputeCapabilitiesFilter', 'ImagePropertiesFilter', 'ServerGroupAntiAffinityFilter', 'ServerGroupAffinityFilter', 'SameHostFilter', 'DifferentHostFilter'] {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.940631] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] filter_scheduler.host_subset_size = 1 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.940800] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] filter_scheduler.hypervisor_version_weight_multiplier = 1.0 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.940960] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] filter_scheduler.image_properties_default_architecture = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.941138] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] filter_scheduler.io_ops_weight_multiplier = -1.0 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.941308] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] filter_scheduler.isolated_hosts = [] {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.941470] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] filter_scheduler.isolated_images = [] {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.941633] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] filter_scheduler.max_instances_per_host = 50 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.941795] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] filter_scheduler.max_io_ops_per_host = 8 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.941956] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] filter_scheduler.num_instances_weight_multiplier = 0.0 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.942131] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] filter_scheduler.pci_in_placement = False {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.942296] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] filter_scheduler.pci_weight_multiplier = 1.0 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.942459] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] filter_scheduler.ram_weight_multiplier = 1.0 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.942620] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] filter_scheduler.restrict_isolated_hosts_to_isolated_images = True {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.942777] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] filter_scheduler.shuffle_best_same_weighed_hosts = False {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.942945] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] filter_scheduler.soft_affinity_weight_multiplier = 1.0 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.943123] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] filter_scheduler.soft_anti_affinity_weight_multiplier = 1.0 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.943292] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] filter_scheduler.track_instance_changes = True {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.943502] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] filter_scheduler.weight_classes = ['nova.scheduler.weights.all_weighers'] {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.943680] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] metrics.required = True {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.943847] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] metrics.weight_multiplier = 1.0 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.944018] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] metrics.weight_of_unavailable = -10000.0 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.944193] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] metrics.weight_setting = [] {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.944515] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] serial_console.base_url = ws://127.0.0.1:6083/ {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.944694] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] serial_console.enabled = False {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.944871] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] serial_console.port_range = 10000:20000 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.945061] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] serial_console.proxyclient_address = 127.0.0.1 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.945235] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] serial_console.serialproxy_host = 0.0.0.0 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.945407] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] serial_console.serialproxy_port = 6083 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.945599] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] service_user.auth_section = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.945779] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] service_user.auth_type = password {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.945941] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] service_user.cafile = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.946114] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] service_user.certfile = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.946281] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] service_user.collect_timing = False {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.946440] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] service_user.insecure = False {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.946621] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] service_user.keyfile = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.946798] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] service_user.send_service_user_token = True {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.946961] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] service_user.split_loggers = False {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.947152] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] service_user.timeout = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.947326] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] spice.agent_enabled = True {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.947490] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] spice.enabled = False {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.947809] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] spice.html5proxy_base_url = http://127.0.0.1:6082/spice_auto.html {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.948009] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] spice.html5proxy_host = 0.0.0.0 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.948192] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] spice.html5proxy_port = 6082 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.948355] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] spice.image_compression = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.948515] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] spice.jpeg_compression = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.948675] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] spice.playback_compression = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.948834] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] spice.require_secure = False {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.949012] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] spice.server_listen = 127.0.0.1 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.949184] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] spice.server_proxyclient_address = 127.0.0.1 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.949343] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] spice.streaming_mode = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.949500] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] spice.zlib_compression = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.949667] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] upgrade_levels.baseapi = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.949836] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] upgrade_levels.compute = auto {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.949995] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] upgrade_levels.conductor = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.950167] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] upgrade_levels.scheduler = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.950333] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] vendordata_dynamic_auth.auth_section = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.950497] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] vendordata_dynamic_auth.auth_type = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.950665] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] vendordata_dynamic_auth.cafile = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.950826] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] vendordata_dynamic_auth.certfile = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.950988] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] vendordata_dynamic_auth.collect_timing = False {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.951162] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] vendordata_dynamic_auth.insecure = False {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.951344] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] vendordata_dynamic_auth.keyfile = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.951480] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] vendordata_dynamic_auth.split_loggers = False {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.951641] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] vendordata_dynamic_auth.timeout = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.951815] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] vmware.api_retry_count = 10 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.951973] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] vmware.ca_file = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.952156] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] vmware.cache_prefix = devstack-image-cache {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.952323] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] vmware.cluster_name = testcl1 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.952489] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] vmware.connection_pool_size = 10 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.952650] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] vmware.console_delay_seconds = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.952816] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] vmware.datastore_regex = ^datastore.* {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.953040] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] vmware.host_ip = vc1.osci.c.eu-de-1.cloud.sap {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.953218] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] vmware.host_password = **** {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.953408] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] vmware.host_port = 443 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.953590] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] vmware.host_username = administrator@vsphere.local {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.953758] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] vmware.insecure = True {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.953922] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] vmware.integration_bridge = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.954104] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] vmware.maximum_objects = 100 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.954268] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] vmware.pbm_default_policy = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.954431] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] vmware.pbm_enabled = False {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.954594] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] vmware.pbm_wsdl_location = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.954759] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] vmware.serial_log_dir = /opt/vmware/vspc {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.954918] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] vmware.serial_port_proxy_uri = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.955089] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] vmware.serial_port_service_uri = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.955258] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] vmware.task_poll_interval = 0.5 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.955458] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] vmware.use_linked_clone = False {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.955634] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] vmware.vnc_keymap = en-us {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.955803] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] vmware.vnc_port = 5900 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.955970] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] vmware.vnc_port_total = 10000 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.956172] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] vnc.auth_schemes = ['none'] {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.956348] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] vnc.enabled = False {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.956666] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] vnc.novncproxy_base_url = http://127.0.0.1:6080/vnc_auto.html {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.956860] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] vnc.novncproxy_host = 0.0.0.0 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.957046] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] vnc.novncproxy_port = 6080 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.957231] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] vnc.server_listen = 127.0.0.1 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.957406] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] vnc.server_proxyclient_address = 127.0.0.1 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.957569] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] vnc.vencrypt_ca_certs = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.957727] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] vnc.vencrypt_client_cert = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.957885] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] vnc.vencrypt_client_key = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.958082] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] workarounds.disable_compute_service_check_for_ffu = False {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.958251] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] workarounds.disable_deep_image_inspection = False {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.958417] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] workarounds.disable_fallback_pcpu_query = False {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.958578] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] workarounds.disable_group_policy_check_upcall = False {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.958740] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] workarounds.disable_libvirt_livesnapshot = False {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.958902] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] workarounds.disable_rootwrap = False {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.959096] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] workarounds.enable_numa_live_migration = False {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.959281] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] workarounds.enable_qemu_monitor_announce_self = False {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.959449] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] workarounds.ensure_libvirt_rbd_instance_dir_cleanup = False {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.959612] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] workarounds.handle_virt_lifecycle_events = True {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.959772] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] workarounds.libvirt_disable_apic = False {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.959932] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] workarounds.never_download_image_if_on_rbd = False {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.960128] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] workarounds.qemu_monitor_announce_self_count = 3 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.960300] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] workarounds.qemu_monitor_announce_self_interval = 1 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.960465] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] workarounds.reserve_disk_resource_for_image_cache = False {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.960629] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] workarounds.skip_cpu_compare_at_startup = False {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.960792] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] workarounds.skip_cpu_compare_on_dest = False {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.960954] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] workarounds.skip_hypervisor_version_check_on_lm = False {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.961132] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] workarounds.skip_reserve_in_use_ironic_nodes = False {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.961294] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] workarounds.unified_limits_count_pcpu_as_vcpu = False {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.961460] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] workarounds.wait_for_vif_plugged_event_during_hard_reboot = [] {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.961645] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] wsgi.api_paste_config = /etc/nova/api-paste.ini {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.961817] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] wsgi.client_socket_timeout = 900 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.961986] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] wsgi.default_pool_size = 1000 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.962170] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] wsgi.keep_alive = True {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.962343] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] wsgi.max_header_line = 16384 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.962506] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] wsgi.secure_proxy_ssl_header = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.962669] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] wsgi.ssl_ca_file = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.962831] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] wsgi.ssl_cert_file = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.962991] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] wsgi.ssl_key_file = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.963173] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] wsgi.tcp_keepidle = 600 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.963379] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] wsgi.wsgi_log_format = %(client_ip)s "%(request_line)s" status: %(status_code)s len: %(body_length)s time: %(wall_seconds).7f {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.963565] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] zvm.ca_file = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.963732] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] zvm.cloud_connector_url = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.964038] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] zvm.image_tmp_path = /opt/stack/data/n-cpu-1/images {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.964224] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] zvm.reachable_timeout = 300 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.964412] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] oslo_policy.enforce_new_defaults = True {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.964805] env[61923]: WARNING oslo_config.cfg [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] Deprecated: Option "enforce_scope" from group "oslo_policy" is deprecated for removal (This configuration was added temporarily to facilitate a smooth transition to the new RBAC. OpenStack will always enforce scope checks. This configuration option is deprecated and will be removed in the 2025.2 cycle.). Its value may be silently ignored in the future. [ 470.964994] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] oslo_policy.enforce_scope = True {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.965190] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] oslo_policy.policy_default_rule = default {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.965377] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] oslo_policy.policy_dirs = ['policy.d'] {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.965599] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] oslo_policy.policy_file = policy.yaml {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.965813] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] oslo_policy.remote_content_type = application/x-www-form-urlencoded {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.965952] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] oslo_policy.remote_ssl_ca_crt_file = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.966133] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] oslo_policy.remote_ssl_client_crt_file = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.966298] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] oslo_policy.remote_ssl_client_key_file = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.966482] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] oslo_policy.remote_ssl_verify_server_crt = False {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.966703] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] oslo_versionedobjects.fatal_exception_format_errors = False {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.966893] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] oslo_middleware.http_basic_auth_user_file = /etc/htpasswd {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.967090] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] profiler.connection_string = messaging:// {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.967266] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] profiler.enabled = False {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.967439] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] profiler.es_doc_type = notification {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.967608] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] profiler.es_scroll_size = 10000 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.967777] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] profiler.es_scroll_time = 2m {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.967940] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] profiler.filter_error_trace = False {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.968126] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] profiler.hmac_keys = **** {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.968297] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] profiler.sentinel_service_name = mymaster {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.968464] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] profiler.socket_timeout = 0.1 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.968629] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] profiler.trace_requests = False {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.968825] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] profiler.trace_sqlalchemy = False {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.968970] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] profiler_jaeger.process_tags = {} {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.969145] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] profiler_jaeger.service_name_prefix = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.969313] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] profiler_otlp.service_name_prefix = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.969484] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] remote_debug.host = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.969648] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] remote_debug.port = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.969828] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] oslo_messaging_rabbit.amqp_auto_delete = False {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.969990] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] oslo_messaging_rabbit.amqp_durable_queues = False {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.970170] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] oslo_messaging_rabbit.conn_pool_min_size = 2 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.970332] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] oslo_messaging_rabbit.conn_pool_ttl = 1200 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.970494] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] oslo_messaging_rabbit.direct_mandatory_flag = True {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.970655] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] oslo_messaging_rabbit.enable_cancel_on_failover = False {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.970815] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] oslo_messaging_rabbit.heartbeat_in_pthread = False {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.970977] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] oslo_messaging_rabbit.heartbeat_rate = 3 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.971157] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] oslo_messaging_rabbit.heartbeat_timeout_threshold = 60 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.971330] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] oslo_messaging_rabbit.hostname = devstack {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.971535] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] oslo_messaging_rabbit.kombu_compression = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.971660] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] oslo_messaging_rabbit.kombu_failover_strategy = round-robin {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.971832] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] oslo_messaging_rabbit.kombu_missing_consumer_retry_timeout = 60 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.972023] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] oslo_messaging_rabbit.kombu_reconnect_delay = 1.0 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.972194] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] oslo_messaging_rabbit.processname = nova-compute {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.972364] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] oslo_messaging_rabbit.rabbit_ha_queues = False {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.972528] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] oslo_messaging_rabbit.rabbit_interval_max = 30 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.972702] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] oslo_messaging_rabbit.rabbit_login_method = AMQPLAIN {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.972867] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] oslo_messaging_rabbit.rabbit_qos_prefetch_count = 0 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.973043] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] oslo_messaging_rabbit.rabbit_quorum_delivery_limit = 0 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.973219] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] oslo_messaging_rabbit.rabbit_quorum_max_memory_bytes = 0 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.973409] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] oslo_messaging_rabbit.rabbit_quorum_max_memory_length = 0 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.973589] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] oslo_messaging_rabbit.rabbit_quorum_queue = False {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.973760] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] oslo_messaging_rabbit.rabbit_retry_backoff = 2 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.973926] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] oslo_messaging_rabbit.rabbit_retry_interval = 1 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.974104] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] oslo_messaging_rabbit.rabbit_stream_fanout = False {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.974275] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] oslo_messaging_rabbit.rabbit_transient_queues_ttl = 1800 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.974437] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] oslo_messaging_rabbit.rabbit_transient_quorum_queue = False {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.974608] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] oslo_messaging_rabbit.rpc_conn_pool_size = 30 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.974777] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] oslo_messaging_rabbit.ssl = False {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.974951] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] oslo_messaging_rabbit.ssl_ca_file = {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.975135] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] oslo_messaging_rabbit.ssl_cert_file = {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.975302] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] oslo_messaging_rabbit.ssl_enforce_fips_mode = False {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.975499] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] oslo_messaging_rabbit.ssl_key_file = {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.975681] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] oslo_messaging_rabbit.ssl_version = {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.975846] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] oslo_messaging_rabbit.use_queue_manager = False {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.976052] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] oslo_messaging_notifications.driver = ['messagingv2'] {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.976228] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] oslo_messaging_notifications.retry = -1 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.976414] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] oslo_messaging_notifications.topics = ['notifications'] {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.976617] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] oslo_messaging_notifications.transport_url = **** {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.976800] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] oslo_limit.auth_section = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.976970] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] oslo_limit.auth_type = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.977151] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] oslo_limit.cafile = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.977315] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] oslo_limit.certfile = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.977481] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] oslo_limit.collect_timing = False {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.977645] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] oslo_limit.connect_retries = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.977806] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] oslo_limit.connect_retry_delay = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.977964] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] oslo_limit.endpoint_id = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.978138] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] oslo_limit.endpoint_override = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.978303] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] oslo_limit.insecure = False {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.978463] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] oslo_limit.keyfile = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.978658] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] oslo_limit.max_version = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.978832] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] oslo_limit.min_version = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.978992] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] oslo_limit.region_name = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.979174] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] oslo_limit.retriable_status_codes = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.979335] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] oslo_limit.service_name = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.979495] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] oslo_limit.service_type = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.979659] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] oslo_limit.split_loggers = False {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.979819] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] oslo_limit.status_code_retries = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.979976] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] oslo_limit.status_code_retry_delay = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.980151] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] oslo_limit.timeout = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.980314] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] oslo_limit.valid_interfaces = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.980473] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] oslo_limit.version = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.980644] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] oslo_reports.file_event_handler = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.980812] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] oslo_reports.file_event_handler_interval = 1 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.980972] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] oslo_reports.log_dir = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.981163] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] vif_plug_linux_bridge_privileged.capabilities = [12] {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.981324] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] vif_plug_linux_bridge_privileged.group = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.981483] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] vif_plug_linux_bridge_privileged.helper_command = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.981649] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] vif_plug_linux_bridge_privileged.logger_name = oslo_privsep.daemon {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.981814] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] vif_plug_linux_bridge_privileged.thread_pool_size = 8 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.981973] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] vif_plug_linux_bridge_privileged.user = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.982162] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] vif_plug_ovs_privileged.capabilities = [12, 1] {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.982340] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] vif_plug_ovs_privileged.group = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.982600] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] vif_plug_ovs_privileged.helper_command = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.982802] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] vif_plug_ovs_privileged.logger_name = oslo_privsep.daemon {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.982977] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] vif_plug_ovs_privileged.thread_pool_size = 8 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.983160] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] vif_plug_ovs_privileged.user = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.983355] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] os_vif_linux_bridge.flat_interface = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.983554] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] os_vif_linux_bridge.forward_bridge_interface = ['all'] {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.983732] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] os_vif_linux_bridge.iptables_bottom_regex = {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.983903] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] os_vif_linux_bridge.iptables_drop_action = DROP {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.984085] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] os_vif_linux_bridge.iptables_top_regex = {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.984261] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] os_vif_linux_bridge.network_device_mtu = 1500 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.984429] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] os_vif_linux_bridge.use_ipv6 = False {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.984604] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] os_vif_linux_bridge.vlan_interface = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.984794] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] os_vif_ovs.default_qos_type = linux-noop {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.984967] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] os_vif_ovs.isolate_vif = False {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.985155] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] os_vif_ovs.network_device_mtu = 1500 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.985325] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] os_vif_ovs.ovs_vsctl_timeout = 120 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.985522] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] os_vif_ovs.ovsdb_connection = tcp:127.0.0.1:6640 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.985708] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] os_vif_ovs.ovsdb_interface = native {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.985876] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] os_vif_ovs.per_port_bridge = False {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.986066] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] privsep_osbrick.capabilities = [21] {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.986278] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] privsep_osbrick.group = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.986471] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] privsep_osbrick.helper_command = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.986648] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] privsep_osbrick.logger_name = os_brick.privileged {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.986818] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] privsep_osbrick.thread_pool_size = 8 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.986982] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] privsep_osbrick.user = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.987177] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] nova_sys_admin.capabilities = [0, 1, 2, 3, 12, 21] {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.987341] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] nova_sys_admin.group = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.987503] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] nova_sys_admin.helper_command = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.987670] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] nova_sys_admin.logger_name = oslo_privsep.daemon {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.987835] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] nova_sys_admin.thread_pool_size = 8 {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.987994] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] nova_sys_admin.user = None {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 470.988141] env[61923]: DEBUG oslo_service.service [None req-d0f2efbd-dcc0-4b04-b9e2-e4d44b94adfe None None] ******************************************************************************** {{(pid=61923) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2830}} [ 470.988629] env[61923]: INFO nova.service [-] Starting compute node (version 0.0.1) [ 471.494017] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-48415267-b006-4918-b654-5aaf236b3ebc None None] Getting list of instances from cluster (obj){ [ 471.494017] env[61923]: value = "domain-c8" [ 471.494017] env[61923]: _type = "ClusterComputeResource" [ 471.494017] env[61923]: } {{(pid=61923) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2116}} [ 471.494017] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80acc752-04a0-4eef-a702-05353e14e5a0 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 471.502375] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-48415267-b006-4918-b654-5aaf236b3ebc None None] Got total of 0 instances {{(pid=61923) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2125}} [ 471.503044] env[61923]: WARNING nova.virt.vmwareapi.driver [None req-48415267-b006-4918-b654-5aaf236b3ebc None None] The vmwareapi driver is not tested by the OpenStack project nor does it have clear maintainer(s) and thus its quality can not be ensured. It should be considered experimental and may be removed in a future release. If you are using the driver in production please let us know via the openstack-discuss mailing list. [ 471.503702] env[61923]: INFO nova.virt.node [None req-48415267-b006-4918-b654-5aaf236b3ebc None None] Generated node identity f81803f2-f7f8-4939-a757-a77d34a1d0a2 [ 471.504080] env[61923]: INFO nova.virt.node [None req-48415267-b006-4918-b654-5aaf236b3ebc None None] Wrote node identity f81803f2-f7f8-4939-a757-a77d34a1d0a2 to /opt/stack/data/n-cpu-1/compute_id [ 472.007053] env[61923]: WARNING nova.compute.manager [None req-48415267-b006-4918-b654-5aaf236b3ebc None None] Compute nodes ['f81803f2-f7f8-4939-a757-a77d34a1d0a2'] for host cpu-1 were not found in the database. If this is the first time this service is starting on this host, then you can ignore this warning. [ 473.012687] env[61923]: INFO nova.compute.manager [None req-48415267-b006-4918-b654-5aaf236b3ebc None None] Looking for unclaimed instances stuck in BUILDING status for nodes managed by this host [ 474.018357] env[61923]: WARNING nova.compute.manager [None req-48415267-b006-4918-b654-5aaf236b3ebc None None] No compute node record found for host cpu-1. If this is the first time this service is starting on this host, then you can ignore this warning.: nova.exception_Remote.ComputeHostNotFound_Remote: Compute host cpu-1 could not be found. [ 474.018769] env[61923]: DEBUG oslo_concurrency.lockutils [None req-48415267-b006-4918-b654-5aaf236b3ebc None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 474.018843] env[61923]: DEBUG oslo_concurrency.lockutils [None req-48415267-b006-4918-b654-5aaf236b3ebc None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 474.018934] env[61923]: DEBUG oslo_concurrency.lockutils [None req-48415267-b006-4918-b654-5aaf236b3ebc None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 474.019111] env[61923]: DEBUG nova.compute.resource_tracker [None req-48415267-b006-4918-b654-5aaf236b3ebc None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61923) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 474.020087] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-160081dc-d444-41cc-8f02-6fb94779f761 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 474.028666] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96140f7e-b0ca-433e-82c9-bef709e428d3 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 474.041997] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25083117-6395-4f86-b0f3-394058a09136 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 474.048472] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b8d2938-cb27-4c9c-a44a-1a578fb5ee49 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 474.077144] env[61923]: DEBUG nova.compute.resource_tracker [None req-48415267-b006-4918-b654-5aaf236b3ebc None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181462MB free_disk=178GB free_vcpus=48 pci_devices=None {{(pid=61923) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 474.077296] env[61923]: DEBUG oslo_concurrency.lockutils [None req-48415267-b006-4918-b654-5aaf236b3ebc None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 474.077463] env[61923]: DEBUG oslo_concurrency.lockutils [None req-48415267-b006-4918-b654-5aaf236b3ebc None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 474.579792] env[61923]: WARNING nova.compute.resource_tracker [None req-48415267-b006-4918-b654-5aaf236b3ebc None None] No compute node record for cpu-1:f81803f2-f7f8-4939-a757-a77d34a1d0a2: nova.exception_Remote.ComputeHostNotFound_Remote: Compute host f81803f2-f7f8-4939-a757-a77d34a1d0a2 could not be found. [ 475.083490] env[61923]: INFO nova.compute.resource_tracker [None req-48415267-b006-4918-b654-5aaf236b3ebc None None] Compute node record created for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 with uuid: f81803f2-f7f8-4939-a757-a77d34a1d0a2 [ 476.591677] env[61923]: DEBUG nova.compute.resource_tracker [None req-48415267-b006-4918-b654-5aaf236b3ebc None None] Total usable vcpus: 48, total allocated vcpus: 0 {{(pid=61923) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 476.592126] env[61923]: DEBUG nova.compute.resource_tracker [None req-48415267-b006-4918-b654-5aaf236b3ebc None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=512MB phys_disk=200GB used_disk=0GB total_vcpus=48 used_vcpus=0 pci_stats=[] {{(pid=61923) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 476.745874] env[61923]: INFO nova.scheduler.client.report [None req-48415267-b006-4918-b654-5aaf236b3ebc None None] [req-4a96ddf5-258c-45ee-a23b-9f27eeffaf19] Created resource provider record via placement API for resource provider with UUID f81803f2-f7f8-4939-a757-a77d34a1d0a2 and name domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28. [ 476.760251] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7cd6eeea-a06a-4756-8f4f-dde056231674 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 476.767985] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38d85d2b-50a7-49e0-b22a-89322b2d8097 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 476.797706] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5eece64-c8cf-422a-9201-dff8e18f8824 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 476.804786] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e19014f1-b1a1-4915-9ff7-adc8e58938eb {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 476.817651] env[61923]: DEBUG nova.compute.provider_tree [None req-48415267-b006-4918-b654-5aaf236b3ebc None None] Updating inventory in ProviderTree for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 477.356185] env[61923]: DEBUG nova.scheduler.client.report [None req-48415267-b006-4918-b654-5aaf236b3ebc None None] Updated inventory for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 with generation 0 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 477.356410] env[61923]: DEBUG nova.compute.provider_tree [None req-48415267-b006-4918-b654-5aaf236b3ebc None None] Updating resource provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 generation from 0 to 1 during operation: update_inventory {{(pid=61923) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 477.356574] env[61923]: DEBUG nova.compute.provider_tree [None req-48415267-b006-4918-b654-5aaf236b3ebc None None] Updating inventory in ProviderTree for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 477.404987] env[61923]: DEBUG nova.compute.provider_tree [None req-48415267-b006-4918-b654-5aaf236b3ebc None None] Updating resource provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 generation from 1 to 2 during operation: update_traits {{(pid=61923) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 477.909424] env[61923]: DEBUG nova.compute.resource_tracker [None req-48415267-b006-4918-b654-5aaf236b3ebc None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61923) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 477.909789] env[61923]: DEBUG oslo_concurrency.lockutils [None req-48415267-b006-4918-b654-5aaf236b3ebc None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 3.832s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 477.909789] env[61923]: DEBUG nova.service [None req-48415267-b006-4918-b654-5aaf236b3ebc None None] Creating RPC server for service compute {{(pid=61923) start /opt/stack/nova/nova/service.py:186}} [ 477.922478] env[61923]: DEBUG nova.service [None req-48415267-b006-4918-b654-5aaf236b3ebc None None] Join ServiceGroup membership for this service compute {{(pid=61923) start /opt/stack/nova/nova/service.py:203}} [ 477.922686] env[61923]: DEBUG nova.servicegroup.drivers.db [None req-48415267-b006-4918-b654-5aaf236b3ebc None None] DB_Driver: join new ServiceGroup member cpu-1 to the compute group, service = {{(pid=61923) join /opt/stack/nova/nova/servicegroup/drivers/db.py:44}} [ 498.926643] env[61923]: DEBUG oslo_service.periodic_task [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Running periodic task ComputeManager._sync_power_states {{(pid=61923) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 499.429680] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Getting list of instances from cluster (obj){ [ 499.429680] env[61923]: value = "domain-c8" [ 499.429680] env[61923]: _type = "ClusterComputeResource" [ 499.429680] env[61923]: } {{(pid=61923) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2116}} [ 499.431225] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9b05e8c-7a14-42a2-b750-1a7d4fc7631c {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 499.439423] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Got total of 0 instances {{(pid=61923) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2125}} [ 499.439680] env[61923]: DEBUG oslo_service.periodic_task [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Running periodic task ComputeManager._cleanup_running_deleted_instances {{(pid=61923) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 499.439980] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Getting list of instances from cluster (obj){ [ 499.439980] env[61923]: value = "domain-c8" [ 499.439980] env[61923]: _type = "ClusterComputeResource" [ 499.439980] env[61923]: } {{(pid=61923) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2116}} [ 499.440853] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38c72445-1767-425a-bce7-ec49fb345818 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 499.448636] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Got total of 0 instances {{(pid=61923) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2125}} [ 510.658937] env[61923]: DEBUG oslo_concurrency.lockutils [None req-dc345cdf-ff58-4514-9baa-d2d8d13601af tempest-MigrationsAdminTest-1311988845 tempest-MigrationsAdminTest-1311988845-project-member] Acquiring lock "242f5cab-3c0c-4f8d-82a8-0cae8d3d419e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 510.658937] env[61923]: DEBUG oslo_concurrency.lockutils [None req-dc345cdf-ff58-4514-9baa-d2d8d13601af tempest-MigrationsAdminTest-1311988845 tempest-MigrationsAdminTest-1311988845-project-member] Lock "242f5cab-3c0c-4f8d-82a8-0cae8d3d419e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 511.161789] env[61923]: DEBUG nova.compute.manager [None req-dc345cdf-ff58-4514-9baa-d2d8d13601af tempest-MigrationsAdminTest-1311988845 tempest-MigrationsAdminTest-1311988845-project-member] [instance: 242f5cab-3c0c-4f8d-82a8-0cae8d3d419e] Starting instance... {{(pid=61923) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 511.715451] env[61923]: DEBUG oslo_concurrency.lockutils [None req-dc345cdf-ff58-4514-9baa-d2d8d13601af tempest-MigrationsAdminTest-1311988845 tempest-MigrationsAdminTest-1311988845-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 511.715934] env[61923]: DEBUG oslo_concurrency.lockutils [None req-dc345cdf-ff58-4514-9baa-d2d8d13601af tempest-MigrationsAdminTest-1311988845 tempest-MigrationsAdminTest-1311988845-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.001s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 511.718797] env[61923]: INFO nova.compute.claims [None req-dc345cdf-ff58-4514-9baa-d2d8d13601af tempest-MigrationsAdminTest-1311988845 tempest-MigrationsAdminTest-1311988845-project-member] [instance: 242f5cab-3c0c-4f8d-82a8-0cae8d3d419e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 512.794949] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d46d150-2ecc-4f76-babf-10a4f2bd49a2 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 512.810947] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30690682-f77a-4310-ac50-11d6ef52bd71 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 512.855859] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c26d1d8-6f62-4ccf-a7ee-aa99e78cd44d {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 512.866484] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74424a64-8e29-4fcf-a081-fa7d3244b842 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 512.885109] env[61923]: DEBUG nova.compute.provider_tree [None req-dc345cdf-ff58-4514-9baa-d2d8d13601af tempest-MigrationsAdminTest-1311988845 tempest-MigrationsAdminTest-1311988845-project-member] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 513.388337] env[61923]: DEBUG nova.scheduler.client.report [None req-dc345cdf-ff58-4514-9baa-d2d8d13601af tempest-MigrationsAdminTest-1311988845 tempest-MigrationsAdminTest-1311988845-project-member] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 513.440473] env[61923]: DEBUG oslo_concurrency.lockutils [None req-a2070053-8e36-4abe-a723-0b93480b3283 tempest-ServerDiagnosticsTest-1095156200 tempest-ServerDiagnosticsTest-1095156200-project-member] Acquiring lock "a134ab6d-b261-4df6-b440-d32a3e578645" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 513.441320] env[61923]: DEBUG oslo_concurrency.lockutils [None req-a2070053-8e36-4abe-a723-0b93480b3283 tempest-ServerDiagnosticsTest-1095156200 tempest-ServerDiagnosticsTest-1095156200-project-member] Lock "a134ab6d-b261-4df6-b440-d32a3e578645" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 513.895976] env[61923]: DEBUG oslo_concurrency.lockutils [None req-dc345cdf-ff58-4514-9baa-d2d8d13601af tempest-MigrationsAdminTest-1311988845 tempest-MigrationsAdminTest-1311988845-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.180s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 513.896526] env[61923]: DEBUG nova.compute.manager [None req-dc345cdf-ff58-4514-9baa-d2d8d13601af tempest-MigrationsAdminTest-1311988845 tempest-MigrationsAdminTest-1311988845-project-member] [instance: 242f5cab-3c0c-4f8d-82a8-0cae8d3d419e] Start building networks asynchronously for instance. {{(pid=61923) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 513.945607] env[61923]: DEBUG nova.compute.manager [None req-a2070053-8e36-4abe-a723-0b93480b3283 tempest-ServerDiagnosticsTest-1095156200 tempest-ServerDiagnosticsTest-1095156200-project-member] [instance: a134ab6d-b261-4df6-b440-d32a3e578645] Starting instance... {{(pid=61923) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 514.404548] env[61923]: DEBUG nova.compute.utils [None req-dc345cdf-ff58-4514-9baa-d2d8d13601af tempest-MigrationsAdminTest-1311988845 tempest-MigrationsAdminTest-1311988845-project-member] Using /dev/sd instead of None {{(pid=61923) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 514.407366] env[61923]: DEBUG nova.compute.manager [None req-dc345cdf-ff58-4514-9baa-d2d8d13601af tempest-MigrationsAdminTest-1311988845 tempest-MigrationsAdminTest-1311988845-project-member] [instance: 242f5cab-3c0c-4f8d-82a8-0cae8d3d419e] Allocating IP information in the background. {{(pid=61923) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 514.407366] env[61923]: DEBUG nova.network.neutron [None req-dc345cdf-ff58-4514-9baa-d2d8d13601af tempest-MigrationsAdminTest-1311988845 tempest-MigrationsAdminTest-1311988845-project-member] [instance: 242f5cab-3c0c-4f8d-82a8-0cae8d3d419e] allocate_for_instance() {{(pid=61923) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 514.490112] env[61923]: DEBUG oslo_concurrency.lockutils [None req-a2070053-8e36-4abe-a723-0b93480b3283 tempest-ServerDiagnosticsTest-1095156200 tempest-ServerDiagnosticsTest-1095156200-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 514.490112] env[61923]: DEBUG oslo_concurrency.lockutils [None req-a2070053-8e36-4abe-a723-0b93480b3283 tempest-ServerDiagnosticsTest-1095156200 tempest-ServerDiagnosticsTest-1095156200-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.001s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 514.492140] env[61923]: INFO nova.compute.claims [None req-a2070053-8e36-4abe-a723-0b93480b3283 tempest-ServerDiagnosticsTest-1095156200 tempest-ServerDiagnosticsTest-1095156200-project-member] [instance: a134ab6d-b261-4df6-b440-d32a3e578645] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 514.824934] env[61923]: DEBUG nova.policy [None req-dc345cdf-ff58-4514-9baa-d2d8d13601af tempest-MigrationsAdminTest-1311988845 tempest-MigrationsAdminTest-1311988845-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5eaeac65cba14de5b8cd759cf5fba8f4', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'feb670247d4a4baabfb19df3f9399223', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61923) authorize /opt/stack/nova/nova/policy.py:201}} [ 514.925040] env[61923]: DEBUG nova.compute.manager [None req-dc345cdf-ff58-4514-9baa-d2d8d13601af tempest-MigrationsAdminTest-1311988845 tempest-MigrationsAdminTest-1311988845-project-member] [instance: 242f5cab-3c0c-4f8d-82a8-0cae8d3d419e] Start building block device mappings for instance. {{(pid=61923) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 515.571230] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2354b0e-41f9-4f81-9546-bca372582a4f {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 515.582902] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63fdd756-64fd-490d-bf54-c1e5f8711c56 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 515.625888] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32f905e0-fa88-44ba-9912-ed06e620a82d {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 515.633405] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-865c9059-4d89-4425-9486-28eecc7ab558 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 515.647625] env[61923]: DEBUG nova.compute.provider_tree [None req-a2070053-8e36-4abe-a723-0b93480b3283 tempest-ServerDiagnosticsTest-1095156200 tempest-ServerDiagnosticsTest-1095156200-project-member] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 515.939241] env[61923]: DEBUG nova.compute.manager [None req-dc345cdf-ff58-4514-9baa-d2d8d13601af tempest-MigrationsAdminTest-1311988845 tempest-MigrationsAdminTest-1311988845-project-member] [instance: 242f5cab-3c0c-4f8d-82a8-0cae8d3d419e] Start spawning the instance on the hypervisor. {{(pid=61923) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 515.975091] env[61923]: DEBUG nova.virt.hardware [None req-dc345cdf-ff58-4514-9baa-d2d8d13601af tempest-MigrationsAdminTest-1311988845 tempest-MigrationsAdminTest-1311988845-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-29T20:07:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-29T20:07:35Z,direct_url=,disk_format='vmdk',id=0f153f63-ae0a-45b1-b7f5-7f9b673c947f,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='4e7b5e3b3c3443c8bd5c70f8a15739f5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-29T20:07:36Z,virtual_size=,visibility=), allow threads: False {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 515.977116] env[61923]: DEBUG nova.virt.hardware [None req-dc345cdf-ff58-4514-9baa-d2d8d13601af tempest-MigrationsAdminTest-1311988845 tempest-MigrationsAdminTest-1311988845-project-member] Flavor limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 515.977116] env[61923]: DEBUG nova.virt.hardware [None req-dc345cdf-ff58-4514-9baa-d2d8d13601af tempest-MigrationsAdminTest-1311988845 tempest-MigrationsAdminTest-1311988845-project-member] Image limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 515.977116] env[61923]: DEBUG nova.virt.hardware [None req-dc345cdf-ff58-4514-9baa-d2d8d13601af tempest-MigrationsAdminTest-1311988845 tempest-MigrationsAdminTest-1311988845-project-member] Flavor pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 515.977116] env[61923]: DEBUG nova.virt.hardware [None req-dc345cdf-ff58-4514-9baa-d2d8d13601af tempest-MigrationsAdminTest-1311988845 tempest-MigrationsAdminTest-1311988845-project-member] Image pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 515.977116] env[61923]: DEBUG nova.virt.hardware [None req-dc345cdf-ff58-4514-9baa-d2d8d13601af tempest-MigrationsAdminTest-1311988845 tempest-MigrationsAdminTest-1311988845-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 515.977482] env[61923]: DEBUG nova.virt.hardware [None req-dc345cdf-ff58-4514-9baa-d2d8d13601af tempest-MigrationsAdminTest-1311988845 tempest-MigrationsAdminTest-1311988845-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 515.977482] env[61923]: DEBUG nova.virt.hardware [None req-dc345cdf-ff58-4514-9baa-d2d8d13601af tempest-MigrationsAdminTest-1311988845 tempest-MigrationsAdminTest-1311988845-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 515.977482] env[61923]: DEBUG nova.virt.hardware [None req-dc345cdf-ff58-4514-9baa-d2d8d13601af tempest-MigrationsAdminTest-1311988845 tempest-MigrationsAdminTest-1311988845-project-member] Got 1 possible topologies {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 515.977482] env[61923]: DEBUG nova.virt.hardware [None req-dc345cdf-ff58-4514-9baa-d2d8d13601af tempest-MigrationsAdminTest-1311988845 tempest-MigrationsAdminTest-1311988845-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 515.977482] env[61923]: DEBUG nova.virt.hardware [None req-dc345cdf-ff58-4514-9baa-d2d8d13601af tempest-MigrationsAdminTest-1311988845 tempest-MigrationsAdminTest-1311988845-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 515.978321] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83ce61a1-a757-4a5e-8111-d929509e38b9 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 515.988908] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0b17525-ec74-4c14-a18b-8ddc2c5f969e {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 516.010011] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d994036-a286-4fb2-9a88-b3c95cf661a6 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 516.138649] env[61923]: DEBUG nova.network.neutron [None req-dc345cdf-ff58-4514-9baa-d2d8d13601af tempest-MigrationsAdminTest-1311988845 tempest-MigrationsAdminTest-1311988845-project-member] [instance: 242f5cab-3c0c-4f8d-82a8-0cae8d3d419e] Successfully created port: 88c39c48-cf39-4b87-abda-5dbe55ea29c8 {{(pid=61923) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 516.153114] env[61923]: DEBUG nova.scheduler.client.report [None req-a2070053-8e36-4abe-a723-0b93480b3283 tempest-ServerDiagnosticsTest-1095156200 tempest-ServerDiagnosticsTest-1095156200-project-member] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 516.662208] env[61923]: DEBUG oslo_concurrency.lockutils [None req-a2070053-8e36-4abe-a723-0b93480b3283 tempest-ServerDiagnosticsTest-1095156200 tempest-ServerDiagnosticsTest-1095156200-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.172s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 516.663932] env[61923]: DEBUG nova.compute.manager [None req-a2070053-8e36-4abe-a723-0b93480b3283 tempest-ServerDiagnosticsTest-1095156200 tempest-ServerDiagnosticsTest-1095156200-project-member] [instance: a134ab6d-b261-4df6-b440-d32a3e578645] Start building networks asynchronously for instance. {{(pid=61923) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 517.174405] env[61923]: DEBUG nova.compute.utils [None req-a2070053-8e36-4abe-a723-0b93480b3283 tempest-ServerDiagnosticsTest-1095156200 tempest-ServerDiagnosticsTest-1095156200-project-member] Using /dev/sd instead of None {{(pid=61923) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 517.175897] env[61923]: DEBUG nova.compute.manager [None req-a2070053-8e36-4abe-a723-0b93480b3283 tempest-ServerDiagnosticsTest-1095156200 tempest-ServerDiagnosticsTest-1095156200-project-member] [instance: a134ab6d-b261-4df6-b440-d32a3e578645] Allocating IP information in the background. {{(pid=61923) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 517.177169] env[61923]: DEBUG nova.network.neutron [None req-a2070053-8e36-4abe-a723-0b93480b3283 tempest-ServerDiagnosticsTest-1095156200 tempest-ServerDiagnosticsTest-1095156200-project-member] [instance: a134ab6d-b261-4df6-b440-d32a3e578645] allocate_for_instance() {{(pid=61923) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 517.217121] env[61923]: DEBUG oslo_concurrency.lockutils [None req-464a16f8-1845-4d4d-b3ea-fd8ba00de8bb tempest-ServersAdminNegativeTestJSON-1828705899 tempest-ServersAdminNegativeTestJSON-1828705899-project-member] Acquiring lock "1ade1b37-8319-4d74-9544-6b83e8738498" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 517.217852] env[61923]: DEBUG oslo_concurrency.lockutils [None req-464a16f8-1845-4d4d-b3ea-fd8ba00de8bb tempest-ServersAdminNegativeTestJSON-1828705899 tempest-ServersAdminNegativeTestJSON-1828705899-project-member] Lock "1ade1b37-8319-4d74-9544-6b83e8738498" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 517.219932] env[61923]: DEBUG oslo_concurrency.lockutils [None req-35a5fc5c-7e44-4806-b390-80f5a8a499d8 tempest-DeleteServersAdminTestJSON-590116201 tempest-DeleteServersAdminTestJSON-590116201-project-member] Acquiring lock "9ea2600a-d439-4fc3-a131-9bb93250d8cd" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 517.220141] env[61923]: DEBUG oslo_concurrency.lockutils [None req-35a5fc5c-7e44-4806-b390-80f5a8a499d8 tempest-DeleteServersAdminTestJSON-590116201 tempest-DeleteServersAdminTestJSON-590116201-project-member] Lock "9ea2600a-d439-4fc3-a131-9bb93250d8cd" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 517.316575] env[61923]: DEBUG nova.policy [None req-a2070053-8e36-4abe-a723-0b93480b3283 tempest-ServerDiagnosticsTest-1095156200 tempest-ServerDiagnosticsTest-1095156200-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ef8253d96d0d4dd592e890bbff6e1e60', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e44edd2a605d4531bf3696c9de32a639', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61923) authorize /opt/stack/nova/nova/policy.py:201}} [ 517.683087] env[61923]: DEBUG nova.compute.manager [None req-a2070053-8e36-4abe-a723-0b93480b3283 tempest-ServerDiagnosticsTest-1095156200 tempest-ServerDiagnosticsTest-1095156200-project-member] [instance: a134ab6d-b261-4df6-b440-d32a3e578645] Start building block device mappings for instance. {{(pid=61923) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 517.725829] env[61923]: DEBUG nova.compute.manager [None req-464a16f8-1845-4d4d-b3ea-fd8ba00de8bb tempest-ServersAdminNegativeTestJSON-1828705899 tempest-ServersAdminNegativeTestJSON-1828705899-project-member] [instance: 1ade1b37-8319-4d74-9544-6b83e8738498] Starting instance... {{(pid=61923) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 517.732307] env[61923]: DEBUG nova.compute.manager [None req-35a5fc5c-7e44-4806-b390-80f5a8a499d8 tempest-DeleteServersAdminTestJSON-590116201 tempest-DeleteServersAdminTestJSON-590116201-project-member] [instance: 9ea2600a-d439-4fc3-a131-9bb93250d8cd] Starting instance... {{(pid=61923) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 517.873542] env[61923]: DEBUG oslo_concurrency.lockutils [None req-8a6b6bc2-54c9-4389-ab30-057158d86227 tempest-ServerDiagnosticsNegativeTest-664316905 tempest-ServerDiagnosticsNegativeTest-664316905-project-member] Acquiring lock "b0a45314-eed8-4d0a-bea2-d4b47dc0bc80" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 517.874092] env[61923]: DEBUG oslo_concurrency.lockutils [None req-8a6b6bc2-54c9-4389-ab30-057158d86227 tempest-ServerDiagnosticsNegativeTest-664316905 tempest-ServerDiagnosticsNegativeTest-664316905-project-member] Lock "b0a45314-eed8-4d0a-bea2-d4b47dc0bc80" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 518.048078] env[61923]: DEBUG nova.network.neutron [None req-a2070053-8e36-4abe-a723-0b93480b3283 tempest-ServerDiagnosticsTest-1095156200 tempest-ServerDiagnosticsTest-1095156200-project-member] [instance: a134ab6d-b261-4df6-b440-d32a3e578645] Successfully created port: 6aa16849-cb00-49ad-9abe-b5ca6484171d {{(pid=61923) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 518.260632] env[61923]: DEBUG oslo_concurrency.lockutils [None req-35a5fc5c-7e44-4806-b390-80f5a8a499d8 tempest-DeleteServersAdminTestJSON-590116201 tempest-DeleteServersAdminTestJSON-590116201-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 518.261049] env[61923]: DEBUG oslo_concurrency.lockutils [None req-35a5fc5c-7e44-4806-b390-80f5a8a499d8 tempest-DeleteServersAdminTestJSON-590116201 tempest-DeleteServersAdminTestJSON-590116201-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 518.262908] env[61923]: INFO nova.compute.claims [None req-35a5fc5c-7e44-4806-b390-80f5a8a499d8 tempest-DeleteServersAdminTestJSON-590116201 tempest-DeleteServersAdminTestJSON-590116201-project-member] [instance: 9ea2600a-d439-4fc3-a131-9bb93250d8cd] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 518.267496] env[61923]: DEBUG oslo_concurrency.lockutils [None req-464a16f8-1845-4d4d-b3ea-fd8ba00de8bb tempest-ServersAdminNegativeTestJSON-1828705899 tempest-ServersAdminNegativeTestJSON-1828705899-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 518.377516] env[61923]: DEBUG nova.compute.manager [None req-8a6b6bc2-54c9-4389-ab30-057158d86227 tempest-ServerDiagnosticsNegativeTest-664316905 tempest-ServerDiagnosticsNegativeTest-664316905-project-member] [instance: b0a45314-eed8-4d0a-bea2-d4b47dc0bc80] Starting instance... {{(pid=61923) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 518.699021] env[61923]: DEBUG nova.compute.manager [None req-a2070053-8e36-4abe-a723-0b93480b3283 tempest-ServerDiagnosticsTest-1095156200 tempest-ServerDiagnosticsTest-1095156200-project-member] [instance: a134ab6d-b261-4df6-b440-d32a3e578645] Start spawning the instance on the hypervisor. {{(pid=61923) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 518.736639] env[61923]: DEBUG nova.virt.hardware [None req-a2070053-8e36-4abe-a723-0b93480b3283 tempest-ServerDiagnosticsTest-1095156200 tempest-ServerDiagnosticsTest-1095156200-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-29T20:07:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-29T20:07:35Z,direct_url=,disk_format='vmdk',id=0f153f63-ae0a-45b1-b7f5-7f9b673c947f,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='4e7b5e3b3c3443c8bd5c70f8a15739f5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-29T20:07:36Z,virtual_size=,visibility=), allow threads: False {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 518.736921] env[61923]: DEBUG nova.virt.hardware [None req-a2070053-8e36-4abe-a723-0b93480b3283 tempest-ServerDiagnosticsTest-1095156200 tempest-ServerDiagnosticsTest-1095156200-project-member] Flavor limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 518.739390] env[61923]: DEBUG nova.virt.hardware [None req-a2070053-8e36-4abe-a723-0b93480b3283 tempest-ServerDiagnosticsTest-1095156200 tempest-ServerDiagnosticsTest-1095156200-project-member] Image limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 518.739792] env[61923]: DEBUG nova.virt.hardware [None req-a2070053-8e36-4abe-a723-0b93480b3283 tempest-ServerDiagnosticsTest-1095156200 tempest-ServerDiagnosticsTest-1095156200-project-member] Flavor pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 518.739792] env[61923]: DEBUG nova.virt.hardware [None req-a2070053-8e36-4abe-a723-0b93480b3283 tempest-ServerDiagnosticsTest-1095156200 tempest-ServerDiagnosticsTest-1095156200-project-member] Image pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 518.739953] env[61923]: DEBUG nova.virt.hardware [None req-a2070053-8e36-4abe-a723-0b93480b3283 tempest-ServerDiagnosticsTest-1095156200 tempest-ServerDiagnosticsTest-1095156200-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 518.740219] env[61923]: DEBUG nova.virt.hardware [None req-a2070053-8e36-4abe-a723-0b93480b3283 tempest-ServerDiagnosticsTest-1095156200 tempest-ServerDiagnosticsTest-1095156200-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 518.740408] env[61923]: DEBUG nova.virt.hardware [None req-a2070053-8e36-4abe-a723-0b93480b3283 tempest-ServerDiagnosticsTest-1095156200 tempest-ServerDiagnosticsTest-1095156200-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 518.740583] env[61923]: DEBUG nova.virt.hardware [None req-a2070053-8e36-4abe-a723-0b93480b3283 tempest-ServerDiagnosticsTest-1095156200 tempest-ServerDiagnosticsTest-1095156200-project-member] Got 1 possible topologies {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 518.740745] env[61923]: DEBUG nova.virt.hardware [None req-a2070053-8e36-4abe-a723-0b93480b3283 tempest-ServerDiagnosticsTest-1095156200 tempest-ServerDiagnosticsTest-1095156200-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 518.740915] env[61923]: DEBUG nova.virt.hardware [None req-a2070053-8e36-4abe-a723-0b93480b3283 tempest-ServerDiagnosticsTest-1095156200 tempest-ServerDiagnosticsTest-1095156200-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 518.743291] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f058e92-5871-4452-84b5-1316b89aefab {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 518.753754] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3013a8c-7ffe-487e-80ac-0ce8ead21764 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 518.907481] env[61923]: DEBUG oslo_concurrency.lockutils [None req-8a6b6bc2-54c9-4389-ab30-057158d86227 tempest-ServerDiagnosticsNegativeTest-664316905 tempest-ServerDiagnosticsNegativeTest-664316905-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 519.381525] env[61923]: ERROR nova.compute.manager [None req-dc345cdf-ff58-4514-9baa-d2d8d13601af tempest-MigrationsAdminTest-1311988845 tempest-MigrationsAdminTest-1311988845-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 88c39c48-cf39-4b87-abda-5dbe55ea29c8, please check neutron logs for more information. [ 519.381525] env[61923]: ERROR nova.compute.manager Traceback (most recent call last): [ 519.381525] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 519.381525] env[61923]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 519.381525] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 519.381525] env[61923]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 519.381525] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 519.381525] env[61923]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 519.381525] env[61923]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 519.381525] env[61923]: ERROR nova.compute.manager self.force_reraise() [ 519.381525] env[61923]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 519.381525] env[61923]: ERROR nova.compute.manager raise self.value [ 519.381525] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 519.381525] env[61923]: ERROR nova.compute.manager updated_port = self._update_port( [ 519.381525] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 519.381525] env[61923]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 519.382215] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 519.382215] env[61923]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 519.382215] env[61923]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 88c39c48-cf39-4b87-abda-5dbe55ea29c8, please check neutron logs for more information. [ 519.382215] env[61923]: ERROR nova.compute.manager [ 519.382215] env[61923]: Traceback (most recent call last): [ 519.382215] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 519.382215] env[61923]: listener.cb(fileno) [ 519.382215] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 519.382215] env[61923]: result = function(*args, **kwargs) [ 519.382215] env[61923]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 519.382215] env[61923]: return func(*args, **kwargs) [ 519.382215] env[61923]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 519.382215] env[61923]: raise e [ 519.382215] env[61923]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 519.382215] env[61923]: nwinfo = self.network_api.allocate_for_instance( [ 519.382215] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 519.382215] env[61923]: created_port_ids = self._update_ports_for_instance( [ 519.382215] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 519.382215] env[61923]: with excutils.save_and_reraise_exception(): [ 519.382215] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 519.382215] env[61923]: self.force_reraise() [ 519.382215] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 519.382215] env[61923]: raise self.value [ 519.382215] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 519.382215] env[61923]: updated_port = self._update_port( [ 519.382215] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 519.382215] env[61923]: _ensure_no_port_binding_failure(port) [ 519.382215] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 519.382215] env[61923]: raise exception.PortBindingFailed(port_id=port['id']) [ 519.382964] env[61923]: nova.exception.PortBindingFailed: Binding failed for port 88c39c48-cf39-4b87-abda-5dbe55ea29c8, please check neutron logs for more information. [ 519.382964] env[61923]: Removing descriptor: 14 [ 519.384855] env[61923]: ERROR nova.compute.manager [None req-dc345cdf-ff58-4514-9baa-d2d8d13601af tempest-MigrationsAdminTest-1311988845 tempest-MigrationsAdminTest-1311988845-project-member] [instance: 242f5cab-3c0c-4f8d-82a8-0cae8d3d419e] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 88c39c48-cf39-4b87-abda-5dbe55ea29c8, please check neutron logs for more information. [ 519.384855] env[61923]: ERROR nova.compute.manager [instance: 242f5cab-3c0c-4f8d-82a8-0cae8d3d419e] Traceback (most recent call last): [ 519.384855] env[61923]: ERROR nova.compute.manager [instance: 242f5cab-3c0c-4f8d-82a8-0cae8d3d419e] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 519.384855] env[61923]: ERROR nova.compute.manager [instance: 242f5cab-3c0c-4f8d-82a8-0cae8d3d419e] yield resources [ 519.384855] env[61923]: ERROR nova.compute.manager [instance: 242f5cab-3c0c-4f8d-82a8-0cae8d3d419e] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 519.384855] env[61923]: ERROR nova.compute.manager [instance: 242f5cab-3c0c-4f8d-82a8-0cae8d3d419e] self.driver.spawn(context, instance, image_meta, [ 519.384855] env[61923]: ERROR nova.compute.manager [instance: 242f5cab-3c0c-4f8d-82a8-0cae8d3d419e] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 519.384855] env[61923]: ERROR nova.compute.manager [instance: 242f5cab-3c0c-4f8d-82a8-0cae8d3d419e] self._vmops.spawn(context, instance, image_meta, injected_files, [ 519.384855] env[61923]: ERROR nova.compute.manager [instance: 242f5cab-3c0c-4f8d-82a8-0cae8d3d419e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 519.384855] env[61923]: ERROR nova.compute.manager [instance: 242f5cab-3c0c-4f8d-82a8-0cae8d3d419e] vm_ref = self.build_virtual_machine(instance, [ 519.384855] env[61923]: ERROR nova.compute.manager [instance: 242f5cab-3c0c-4f8d-82a8-0cae8d3d419e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 519.385223] env[61923]: ERROR nova.compute.manager [instance: 242f5cab-3c0c-4f8d-82a8-0cae8d3d419e] vif_infos = vmwarevif.get_vif_info(self._session, [ 519.385223] env[61923]: ERROR nova.compute.manager [instance: 242f5cab-3c0c-4f8d-82a8-0cae8d3d419e] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 519.385223] env[61923]: ERROR nova.compute.manager [instance: 242f5cab-3c0c-4f8d-82a8-0cae8d3d419e] for vif in network_info: [ 519.385223] env[61923]: ERROR nova.compute.manager [instance: 242f5cab-3c0c-4f8d-82a8-0cae8d3d419e] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 519.385223] env[61923]: ERROR nova.compute.manager [instance: 242f5cab-3c0c-4f8d-82a8-0cae8d3d419e] return self._sync_wrapper(fn, *args, **kwargs) [ 519.385223] env[61923]: ERROR nova.compute.manager [instance: 242f5cab-3c0c-4f8d-82a8-0cae8d3d419e] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 519.385223] env[61923]: ERROR nova.compute.manager [instance: 242f5cab-3c0c-4f8d-82a8-0cae8d3d419e] self.wait() [ 519.385223] env[61923]: ERROR nova.compute.manager [instance: 242f5cab-3c0c-4f8d-82a8-0cae8d3d419e] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 519.385223] env[61923]: ERROR nova.compute.manager [instance: 242f5cab-3c0c-4f8d-82a8-0cae8d3d419e] self[:] = self._gt.wait() [ 519.385223] env[61923]: ERROR nova.compute.manager [instance: 242f5cab-3c0c-4f8d-82a8-0cae8d3d419e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 519.385223] env[61923]: ERROR nova.compute.manager [instance: 242f5cab-3c0c-4f8d-82a8-0cae8d3d419e] return self._exit_event.wait() [ 519.385223] env[61923]: ERROR nova.compute.manager [instance: 242f5cab-3c0c-4f8d-82a8-0cae8d3d419e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 519.385223] env[61923]: ERROR nova.compute.manager [instance: 242f5cab-3c0c-4f8d-82a8-0cae8d3d419e] result = hub.switch() [ 519.385559] env[61923]: ERROR nova.compute.manager [instance: 242f5cab-3c0c-4f8d-82a8-0cae8d3d419e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 519.385559] env[61923]: ERROR nova.compute.manager [instance: 242f5cab-3c0c-4f8d-82a8-0cae8d3d419e] return self.greenlet.switch() [ 519.385559] env[61923]: ERROR nova.compute.manager [instance: 242f5cab-3c0c-4f8d-82a8-0cae8d3d419e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 519.385559] env[61923]: ERROR nova.compute.manager [instance: 242f5cab-3c0c-4f8d-82a8-0cae8d3d419e] result = function(*args, **kwargs) [ 519.385559] env[61923]: ERROR nova.compute.manager [instance: 242f5cab-3c0c-4f8d-82a8-0cae8d3d419e] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 519.385559] env[61923]: ERROR nova.compute.manager [instance: 242f5cab-3c0c-4f8d-82a8-0cae8d3d419e] return func(*args, **kwargs) [ 519.385559] env[61923]: ERROR nova.compute.manager [instance: 242f5cab-3c0c-4f8d-82a8-0cae8d3d419e] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 519.385559] env[61923]: ERROR nova.compute.manager [instance: 242f5cab-3c0c-4f8d-82a8-0cae8d3d419e] raise e [ 519.385559] env[61923]: ERROR nova.compute.manager [instance: 242f5cab-3c0c-4f8d-82a8-0cae8d3d419e] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 519.385559] env[61923]: ERROR nova.compute.manager [instance: 242f5cab-3c0c-4f8d-82a8-0cae8d3d419e] nwinfo = self.network_api.allocate_for_instance( [ 519.385559] env[61923]: ERROR nova.compute.manager [instance: 242f5cab-3c0c-4f8d-82a8-0cae8d3d419e] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 519.385559] env[61923]: ERROR nova.compute.manager [instance: 242f5cab-3c0c-4f8d-82a8-0cae8d3d419e] created_port_ids = self._update_ports_for_instance( [ 519.385559] env[61923]: ERROR nova.compute.manager [instance: 242f5cab-3c0c-4f8d-82a8-0cae8d3d419e] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 519.385918] env[61923]: ERROR nova.compute.manager [instance: 242f5cab-3c0c-4f8d-82a8-0cae8d3d419e] with excutils.save_and_reraise_exception(): [ 519.385918] env[61923]: ERROR nova.compute.manager [instance: 242f5cab-3c0c-4f8d-82a8-0cae8d3d419e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 519.385918] env[61923]: ERROR nova.compute.manager [instance: 242f5cab-3c0c-4f8d-82a8-0cae8d3d419e] self.force_reraise() [ 519.385918] env[61923]: ERROR nova.compute.manager [instance: 242f5cab-3c0c-4f8d-82a8-0cae8d3d419e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 519.385918] env[61923]: ERROR nova.compute.manager [instance: 242f5cab-3c0c-4f8d-82a8-0cae8d3d419e] raise self.value [ 519.385918] env[61923]: ERROR nova.compute.manager [instance: 242f5cab-3c0c-4f8d-82a8-0cae8d3d419e] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 519.385918] env[61923]: ERROR nova.compute.manager [instance: 242f5cab-3c0c-4f8d-82a8-0cae8d3d419e] updated_port = self._update_port( [ 519.385918] env[61923]: ERROR nova.compute.manager [instance: 242f5cab-3c0c-4f8d-82a8-0cae8d3d419e] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 519.385918] env[61923]: ERROR nova.compute.manager [instance: 242f5cab-3c0c-4f8d-82a8-0cae8d3d419e] _ensure_no_port_binding_failure(port) [ 519.385918] env[61923]: ERROR nova.compute.manager [instance: 242f5cab-3c0c-4f8d-82a8-0cae8d3d419e] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 519.385918] env[61923]: ERROR nova.compute.manager [instance: 242f5cab-3c0c-4f8d-82a8-0cae8d3d419e] raise exception.PortBindingFailed(port_id=port['id']) [ 519.385918] env[61923]: ERROR nova.compute.manager [instance: 242f5cab-3c0c-4f8d-82a8-0cae8d3d419e] nova.exception.PortBindingFailed: Binding failed for port 88c39c48-cf39-4b87-abda-5dbe55ea29c8, please check neutron logs for more information. [ 519.385918] env[61923]: ERROR nova.compute.manager [instance: 242f5cab-3c0c-4f8d-82a8-0cae8d3d419e] [ 519.386303] env[61923]: INFO nova.compute.manager [None req-dc345cdf-ff58-4514-9baa-d2d8d13601af tempest-MigrationsAdminTest-1311988845 tempest-MigrationsAdminTest-1311988845-project-member] [instance: 242f5cab-3c0c-4f8d-82a8-0cae8d3d419e] Terminating instance [ 519.388079] env[61923]: DEBUG oslo_concurrency.lockutils [None req-dc345cdf-ff58-4514-9baa-d2d8d13601af tempest-MigrationsAdminTest-1311988845 tempest-MigrationsAdminTest-1311988845-project-member] Acquiring lock "refresh_cache-242f5cab-3c0c-4f8d-82a8-0cae8d3d419e" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 519.388253] env[61923]: DEBUG oslo_concurrency.lockutils [None req-dc345cdf-ff58-4514-9baa-d2d8d13601af tempest-MigrationsAdminTest-1311988845 tempest-MigrationsAdminTest-1311988845-project-member] Acquired lock "refresh_cache-242f5cab-3c0c-4f8d-82a8-0cae8d3d419e" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 519.388910] env[61923]: DEBUG nova.network.neutron [None req-dc345cdf-ff58-4514-9baa-d2d8d13601af tempest-MigrationsAdminTest-1311988845 tempest-MigrationsAdminTest-1311988845-project-member] [instance: 242f5cab-3c0c-4f8d-82a8-0cae8d3d419e] Building network info cache for instance {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 519.419817] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e7ac655-57b9-4100-9d23-89057a344c93 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 519.430239] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91585caa-0854-4f5d-8aa4-0620709fa2be {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 519.475613] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1dafa41-b6ed-4542-b325-2c4ac40a27f4 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 519.488568] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7de70ed5-b975-47a2-acea-9b5ed7ed5420 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 519.502300] env[61923]: DEBUG nova.compute.provider_tree [None req-35a5fc5c-7e44-4806-b390-80f5a8a499d8 tempest-DeleteServersAdminTestJSON-590116201 tempest-DeleteServersAdminTestJSON-590116201-project-member] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 519.741285] env[61923]: ERROR nova.compute.manager [None req-a2070053-8e36-4abe-a723-0b93480b3283 tempest-ServerDiagnosticsTest-1095156200 tempest-ServerDiagnosticsTest-1095156200-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 6aa16849-cb00-49ad-9abe-b5ca6484171d, please check neutron logs for more information. [ 519.741285] env[61923]: ERROR nova.compute.manager Traceback (most recent call last): [ 519.741285] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 519.741285] env[61923]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 519.741285] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 519.741285] env[61923]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 519.741285] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 519.741285] env[61923]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 519.741285] env[61923]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 519.741285] env[61923]: ERROR nova.compute.manager self.force_reraise() [ 519.741285] env[61923]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 519.741285] env[61923]: ERROR nova.compute.manager raise self.value [ 519.741285] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 519.741285] env[61923]: ERROR nova.compute.manager updated_port = self._update_port( [ 519.741285] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 519.741285] env[61923]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 519.741749] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 519.741749] env[61923]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 519.741749] env[61923]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 6aa16849-cb00-49ad-9abe-b5ca6484171d, please check neutron logs for more information. [ 519.741749] env[61923]: ERROR nova.compute.manager [ 519.741749] env[61923]: Traceback (most recent call last): [ 519.741749] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 519.741749] env[61923]: listener.cb(fileno) [ 519.741749] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 519.741749] env[61923]: result = function(*args, **kwargs) [ 519.741749] env[61923]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 519.741749] env[61923]: return func(*args, **kwargs) [ 519.741749] env[61923]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 519.741749] env[61923]: raise e [ 519.741749] env[61923]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 519.741749] env[61923]: nwinfo = self.network_api.allocate_for_instance( [ 519.741749] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 519.741749] env[61923]: created_port_ids = self._update_ports_for_instance( [ 519.741749] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 519.741749] env[61923]: with excutils.save_and_reraise_exception(): [ 519.741749] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 519.741749] env[61923]: self.force_reraise() [ 519.741749] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 519.741749] env[61923]: raise self.value [ 519.741749] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 519.741749] env[61923]: updated_port = self._update_port( [ 519.741749] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 519.741749] env[61923]: _ensure_no_port_binding_failure(port) [ 519.741749] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 519.741749] env[61923]: raise exception.PortBindingFailed(port_id=port['id']) [ 519.742503] env[61923]: nova.exception.PortBindingFailed: Binding failed for port 6aa16849-cb00-49ad-9abe-b5ca6484171d, please check neutron logs for more information. [ 519.742503] env[61923]: Removing descriptor: 15 [ 519.742503] env[61923]: ERROR nova.compute.manager [None req-a2070053-8e36-4abe-a723-0b93480b3283 tempest-ServerDiagnosticsTest-1095156200 tempest-ServerDiagnosticsTest-1095156200-project-member] [instance: a134ab6d-b261-4df6-b440-d32a3e578645] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 6aa16849-cb00-49ad-9abe-b5ca6484171d, please check neutron logs for more information. [ 519.742503] env[61923]: ERROR nova.compute.manager [instance: a134ab6d-b261-4df6-b440-d32a3e578645] Traceback (most recent call last): [ 519.742503] env[61923]: ERROR nova.compute.manager [instance: a134ab6d-b261-4df6-b440-d32a3e578645] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 519.742503] env[61923]: ERROR nova.compute.manager [instance: a134ab6d-b261-4df6-b440-d32a3e578645] yield resources [ 519.742503] env[61923]: ERROR nova.compute.manager [instance: a134ab6d-b261-4df6-b440-d32a3e578645] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 519.742503] env[61923]: ERROR nova.compute.manager [instance: a134ab6d-b261-4df6-b440-d32a3e578645] self.driver.spawn(context, instance, image_meta, [ 519.742503] env[61923]: ERROR nova.compute.manager [instance: a134ab6d-b261-4df6-b440-d32a3e578645] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 519.742503] env[61923]: ERROR nova.compute.manager [instance: a134ab6d-b261-4df6-b440-d32a3e578645] self._vmops.spawn(context, instance, image_meta, injected_files, [ 519.742503] env[61923]: ERROR nova.compute.manager [instance: a134ab6d-b261-4df6-b440-d32a3e578645] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 519.742503] env[61923]: ERROR nova.compute.manager [instance: a134ab6d-b261-4df6-b440-d32a3e578645] vm_ref = self.build_virtual_machine(instance, [ 519.742833] env[61923]: ERROR nova.compute.manager [instance: a134ab6d-b261-4df6-b440-d32a3e578645] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 519.742833] env[61923]: ERROR nova.compute.manager [instance: a134ab6d-b261-4df6-b440-d32a3e578645] vif_infos = vmwarevif.get_vif_info(self._session, [ 519.742833] env[61923]: ERROR nova.compute.manager [instance: a134ab6d-b261-4df6-b440-d32a3e578645] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 519.742833] env[61923]: ERROR nova.compute.manager [instance: a134ab6d-b261-4df6-b440-d32a3e578645] for vif in network_info: [ 519.742833] env[61923]: ERROR nova.compute.manager [instance: a134ab6d-b261-4df6-b440-d32a3e578645] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 519.742833] env[61923]: ERROR nova.compute.manager [instance: a134ab6d-b261-4df6-b440-d32a3e578645] return self._sync_wrapper(fn, *args, **kwargs) [ 519.742833] env[61923]: ERROR nova.compute.manager [instance: a134ab6d-b261-4df6-b440-d32a3e578645] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 519.742833] env[61923]: ERROR nova.compute.manager [instance: a134ab6d-b261-4df6-b440-d32a3e578645] self.wait() [ 519.742833] env[61923]: ERROR nova.compute.manager [instance: a134ab6d-b261-4df6-b440-d32a3e578645] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 519.742833] env[61923]: ERROR nova.compute.manager [instance: a134ab6d-b261-4df6-b440-d32a3e578645] self[:] = self._gt.wait() [ 519.742833] env[61923]: ERROR nova.compute.manager [instance: a134ab6d-b261-4df6-b440-d32a3e578645] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 519.742833] env[61923]: ERROR nova.compute.manager [instance: a134ab6d-b261-4df6-b440-d32a3e578645] return self._exit_event.wait() [ 519.742833] env[61923]: ERROR nova.compute.manager [instance: a134ab6d-b261-4df6-b440-d32a3e578645] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 519.743256] env[61923]: ERROR nova.compute.manager [instance: a134ab6d-b261-4df6-b440-d32a3e578645] result = hub.switch() [ 519.743256] env[61923]: ERROR nova.compute.manager [instance: a134ab6d-b261-4df6-b440-d32a3e578645] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 519.743256] env[61923]: ERROR nova.compute.manager [instance: a134ab6d-b261-4df6-b440-d32a3e578645] return self.greenlet.switch() [ 519.743256] env[61923]: ERROR nova.compute.manager [instance: a134ab6d-b261-4df6-b440-d32a3e578645] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 519.743256] env[61923]: ERROR nova.compute.manager [instance: a134ab6d-b261-4df6-b440-d32a3e578645] result = function(*args, **kwargs) [ 519.743256] env[61923]: ERROR nova.compute.manager [instance: a134ab6d-b261-4df6-b440-d32a3e578645] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 519.743256] env[61923]: ERROR nova.compute.manager [instance: a134ab6d-b261-4df6-b440-d32a3e578645] return func(*args, **kwargs) [ 519.743256] env[61923]: ERROR nova.compute.manager [instance: a134ab6d-b261-4df6-b440-d32a3e578645] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 519.743256] env[61923]: ERROR nova.compute.manager [instance: a134ab6d-b261-4df6-b440-d32a3e578645] raise e [ 519.743256] env[61923]: ERROR nova.compute.manager [instance: a134ab6d-b261-4df6-b440-d32a3e578645] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 519.743256] env[61923]: ERROR nova.compute.manager [instance: a134ab6d-b261-4df6-b440-d32a3e578645] nwinfo = self.network_api.allocate_for_instance( [ 519.743256] env[61923]: ERROR nova.compute.manager [instance: a134ab6d-b261-4df6-b440-d32a3e578645] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 519.743256] env[61923]: ERROR nova.compute.manager [instance: a134ab6d-b261-4df6-b440-d32a3e578645] created_port_ids = self._update_ports_for_instance( [ 519.743564] env[61923]: ERROR nova.compute.manager [instance: a134ab6d-b261-4df6-b440-d32a3e578645] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 519.743564] env[61923]: ERROR nova.compute.manager [instance: a134ab6d-b261-4df6-b440-d32a3e578645] with excutils.save_and_reraise_exception(): [ 519.743564] env[61923]: ERROR nova.compute.manager [instance: a134ab6d-b261-4df6-b440-d32a3e578645] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 519.743564] env[61923]: ERROR nova.compute.manager [instance: a134ab6d-b261-4df6-b440-d32a3e578645] self.force_reraise() [ 519.743564] env[61923]: ERROR nova.compute.manager [instance: a134ab6d-b261-4df6-b440-d32a3e578645] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 519.743564] env[61923]: ERROR nova.compute.manager [instance: a134ab6d-b261-4df6-b440-d32a3e578645] raise self.value [ 519.743564] env[61923]: ERROR nova.compute.manager [instance: a134ab6d-b261-4df6-b440-d32a3e578645] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 519.743564] env[61923]: ERROR nova.compute.manager [instance: a134ab6d-b261-4df6-b440-d32a3e578645] updated_port = self._update_port( [ 519.743564] env[61923]: ERROR nova.compute.manager [instance: a134ab6d-b261-4df6-b440-d32a3e578645] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 519.743564] env[61923]: ERROR nova.compute.manager [instance: a134ab6d-b261-4df6-b440-d32a3e578645] _ensure_no_port_binding_failure(port) [ 519.743564] env[61923]: ERROR nova.compute.manager [instance: a134ab6d-b261-4df6-b440-d32a3e578645] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 519.743564] env[61923]: ERROR nova.compute.manager [instance: a134ab6d-b261-4df6-b440-d32a3e578645] raise exception.PortBindingFailed(port_id=port['id']) [ 519.743867] env[61923]: ERROR nova.compute.manager [instance: a134ab6d-b261-4df6-b440-d32a3e578645] nova.exception.PortBindingFailed: Binding failed for port 6aa16849-cb00-49ad-9abe-b5ca6484171d, please check neutron logs for more information. [ 519.743867] env[61923]: ERROR nova.compute.manager [instance: a134ab6d-b261-4df6-b440-d32a3e578645] [ 519.743867] env[61923]: INFO nova.compute.manager [None req-a2070053-8e36-4abe-a723-0b93480b3283 tempest-ServerDiagnosticsTest-1095156200 tempest-ServerDiagnosticsTest-1095156200-project-member] [instance: a134ab6d-b261-4df6-b440-d32a3e578645] Terminating instance [ 519.743943] env[61923]: DEBUG oslo_concurrency.lockutils [None req-a2070053-8e36-4abe-a723-0b93480b3283 tempest-ServerDiagnosticsTest-1095156200 tempest-ServerDiagnosticsTest-1095156200-project-member] Acquiring lock "refresh_cache-a134ab6d-b261-4df6-b440-d32a3e578645" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 519.744056] env[61923]: DEBUG oslo_concurrency.lockutils [None req-a2070053-8e36-4abe-a723-0b93480b3283 tempest-ServerDiagnosticsTest-1095156200 tempest-ServerDiagnosticsTest-1095156200-project-member] Acquired lock "refresh_cache-a134ab6d-b261-4df6-b440-d32a3e578645" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 519.744217] env[61923]: DEBUG nova.network.neutron [None req-a2070053-8e36-4abe-a723-0b93480b3283 tempest-ServerDiagnosticsTest-1095156200 tempest-ServerDiagnosticsTest-1095156200-project-member] [instance: a134ab6d-b261-4df6-b440-d32a3e578645] Building network info cache for instance {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 519.926262] env[61923]: DEBUG nova.network.neutron [None req-dc345cdf-ff58-4514-9baa-d2d8d13601af tempest-MigrationsAdminTest-1311988845 tempest-MigrationsAdminTest-1311988845-project-member] [instance: 242f5cab-3c0c-4f8d-82a8-0cae8d3d419e] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 520.005250] env[61923]: DEBUG nova.network.neutron [None req-dc345cdf-ff58-4514-9baa-d2d8d13601af tempest-MigrationsAdminTest-1311988845 tempest-MigrationsAdminTest-1311988845-project-member] [instance: 242f5cab-3c0c-4f8d-82a8-0cae8d3d419e] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 520.007466] env[61923]: DEBUG nova.scheduler.client.report [None req-35a5fc5c-7e44-4806-b390-80f5a8a499d8 tempest-DeleteServersAdminTestJSON-590116201 tempest-DeleteServersAdminTestJSON-590116201-project-member] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 520.063614] env[61923]: DEBUG oslo_concurrency.lockutils [None req-7781f812-6fae-4bd6-9d91-ec6098c06372 tempest-InstanceActionsTestJSON-2015629344 tempest-InstanceActionsTestJSON-2015629344-project-member] Acquiring lock "97148ea5-398a-4974-b8f6-62290e519794" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 520.063839] env[61923]: DEBUG oslo_concurrency.lockutils [None req-7781f812-6fae-4bd6-9d91-ec6098c06372 tempest-InstanceActionsTestJSON-2015629344 tempest-InstanceActionsTestJSON-2015629344-project-member] Lock "97148ea5-398a-4974-b8f6-62290e519794" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 520.273837] env[61923]: DEBUG nova.network.neutron [None req-a2070053-8e36-4abe-a723-0b93480b3283 tempest-ServerDiagnosticsTest-1095156200 tempest-ServerDiagnosticsTest-1095156200-project-member] [instance: a134ab6d-b261-4df6-b440-d32a3e578645] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 520.383147] env[61923]: DEBUG nova.network.neutron [None req-a2070053-8e36-4abe-a723-0b93480b3283 tempest-ServerDiagnosticsTest-1095156200 tempest-ServerDiagnosticsTest-1095156200-project-member] [instance: a134ab6d-b261-4df6-b440-d32a3e578645] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 520.513584] env[61923]: DEBUG oslo_concurrency.lockutils [None req-35a5fc5c-7e44-4806-b390-80f5a8a499d8 tempest-DeleteServersAdminTestJSON-590116201 tempest-DeleteServersAdminTestJSON-590116201-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.253s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 520.514696] env[61923]: DEBUG nova.compute.manager [None req-35a5fc5c-7e44-4806-b390-80f5a8a499d8 tempest-DeleteServersAdminTestJSON-590116201 tempest-DeleteServersAdminTestJSON-590116201-project-member] [instance: 9ea2600a-d439-4fc3-a131-9bb93250d8cd] Start building networks asynchronously for instance. {{(pid=61923) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 520.517081] env[61923]: DEBUG oslo_concurrency.lockutils [None req-dc345cdf-ff58-4514-9baa-d2d8d13601af tempest-MigrationsAdminTest-1311988845 tempest-MigrationsAdminTest-1311988845-project-member] Releasing lock "refresh_cache-242f5cab-3c0c-4f8d-82a8-0cae8d3d419e" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 520.517469] env[61923]: DEBUG nova.compute.manager [None req-dc345cdf-ff58-4514-9baa-d2d8d13601af tempest-MigrationsAdminTest-1311988845 tempest-MigrationsAdminTest-1311988845-project-member] [instance: 242f5cab-3c0c-4f8d-82a8-0cae8d3d419e] Start destroying the instance on the hypervisor. {{(pid=61923) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 520.517654] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-dc345cdf-ff58-4514-9baa-d2d8d13601af tempest-MigrationsAdminTest-1311988845 tempest-MigrationsAdminTest-1311988845-project-member] [instance: 242f5cab-3c0c-4f8d-82a8-0cae8d3d419e] Destroying instance {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 520.517946] env[61923]: DEBUG oslo_concurrency.lockutils [None req-464a16f8-1845-4d4d-b3ea-fd8ba00de8bb tempest-ServersAdminNegativeTestJSON-1828705899 tempest-ServersAdminNegativeTestJSON-1828705899-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.251s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 520.522356] env[61923]: INFO nova.compute.claims [None req-464a16f8-1845-4d4d-b3ea-fd8ba00de8bb tempest-ServersAdminNegativeTestJSON-1828705899 tempest-ServersAdminNegativeTestJSON-1828705899-project-member] [instance: 1ade1b37-8319-4d74-9544-6b83e8738498] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 520.522356] env[61923]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-deab3ca9-69ae-4355-a606-6c953bb17e77 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 520.531501] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c1f2b9c-26e0-4140-9750-cf9247921132 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 520.559640] env[61923]: WARNING nova.virt.vmwareapi.vmops [None req-dc345cdf-ff58-4514-9baa-d2d8d13601af tempest-MigrationsAdminTest-1311988845 tempest-MigrationsAdminTest-1311988845-project-member] [instance: 242f5cab-3c0c-4f8d-82a8-0cae8d3d419e] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 242f5cab-3c0c-4f8d-82a8-0cae8d3d419e could not be found. [ 520.559762] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-dc345cdf-ff58-4514-9baa-d2d8d13601af tempest-MigrationsAdminTest-1311988845 tempest-MigrationsAdminTest-1311988845-project-member] [instance: 242f5cab-3c0c-4f8d-82a8-0cae8d3d419e] Instance destroyed {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 520.560156] env[61923]: INFO nova.compute.manager [None req-dc345cdf-ff58-4514-9baa-d2d8d13601af tempest-MigrationsAdminTest-1311988845 tempest-MigrationsAdminTest-1311988845-project-member] [instance: 242f5cab-3c0c-4f8d-82a8-0cae8d3d419e] Took 0.04 seconds to destroy the instance on the hypervisor. [ 520.560446] env[61923]: DEBUG oslo.service.loopingcall [None req-dc345cdf-ff58-4514-9baa-d2d8d13601af tempest-MigrationsAdminTest-1311988845 tempest-MigrationsAdminTest-1311988845-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61923) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 520.560658] env[61923]: DEBUG nova.compute.manager [-] [instance: 242f5cab-3c0c-4f8d-82a8-0cae8d3d419e] Deallocating network for instance {{(pid=61923) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 520.560749] env[61923]: DEBUG nova.network.neutron [-] [instance: 242f5cab-3c0c-4f8d-82a8-0cae8d3d419e] deallocate_for_instance() {{(pid=61923) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 520.566069] env[61923]: DEBUG nova.compute.manager [None req-7781f812-6fae-4bd6-9d91-ec6098c06372 tempest-InstanceActionsTestJSON-2015629344 tempest-InstanceActionsTestJSON-2015629344-project-member] [instance: 97148ea5-398a-4974-b8f6-62290e519794] Starting instance... {{(pid=61923) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 520.638893] env[61923]: DEBUG nova.network.neutron [-] [instance: 242f5cab-3c0c-4f8d-82a8-0cae8d3d419e] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 520.890924] env[61923]: DEBUG oslo_concurrency.lockutils [None req-a2070053-8e36-4abe-a723-0b93480b3283 tempest-ServerDiagnosticsTest-1095156200 tempest-ServerDiagnosticsTest-1095156200-project-member] Releasing lock "refresh_cache-a134ab6d-b261-4df6-b440-d32a3e578645" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 520.891731] env[61923]: DEBUG nova.compute.manager [None req-a2070053-8e36-4abe-a723-0b93480b3283 tempest-ServerDiagnosticsTest-1095156200 tempest-ServerDiagnosticsTest-1095156200-project-member] [instance: a134ab6d-b261-4df6-b440-d32a3e578645] Start destroying the instance on the hypervisor. {{(pid=61923) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 520.892882] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-a2070053-8e36-4abe-a723-0b93480b3283 tempest-ServerDiagnosticsTest-1095156200 tempest-ServerDiagnosticsTest-1095156200-project-member] [instance: a134ab6d-b261-4df6-b440-d32a3e578645] Destroying instance {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 520.893352] env[61923]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e9aed40b-8378-4192-bde3-5615a873e6c5 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 520.904979] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-202e6ae1-1b96-4049-9515-39bd19bfb132 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 520.932886] env[61923]: DEBUG oslo_concurrency.lockutils [None req-844a2215-1038-4a30-8a79-40c1c3117bde tempest-FloatingIPsAssociationNegativeTestJSON-1301210771 tempest-FloatingIPsAssociationNegativeTestJSON-1301210771-project-member] Acquiring lock "5d57303a-ae17-4b92-8f12-303677eb0d7e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 520.933146] env[61923]: DEBUG oslo_concurrency.lockutils [None req-844a2215-1038-4a30-8a79-40c1c3117bde tempest-FloatingIPsAssociationNegativeTestJSON-1301210771 tempest-FloatingIPsAssociationNegativeTestJSON-1301210771-project-member] Lock "5d57303a-ae17-4b92-8f12-303677eb0d7e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 520.934552] env[61923]: WARNING nova.virt.vmwareapi.vmops [None req-a2070053-8e36-4abe-a723-0b93480b3283 tempest-ServerDiagnosticsTest-1095156200 tempest-ServerDiagnosticsTest-1095156200-project-member] [instance: a134ab6d-b261-4df6-b440-d32a3e578645] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance a134ab6d-b261-4df6-b440-d32a3e578645 could not be found. [ 520.934751] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-a2070053-8e36-4abe-a723-0b93480b3283 tempest-ServerDiagnosticsTest-1095156200 tempest-ServerDiagnosticsTest-1095156200-project-member] [instance: a134ab6d-b261-4df6-b440-d32a3e578645] Instance destroyed {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 520.934923] env[61923]: INFO nova.compute.manager [None req-a2070053-8e36-4abe-a723-0b93480b3283 tempest-ServerDiagnosticsTest-1095156200 tempest-ServerDiagnosticsTest-1095156200-project-member] [instance: a134ab6d-b261-4df6-b440-d32a3e578645] Took 0.04 seconds to destroy the instance on the hypervisor. [ 520.935199] env[61923]: DEBUG oslo.service.loopingcall [None req-a2070053-8e36-4abe-a723-0b93480b3283 tempest-ServerDiagnosticsTest-1095156200 tempest-ServerDiagnosticsTest-1095156200-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61923) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 520.936924] env[61923]: DEBUG nova.compute.manager [-] [instance: a134ab6d-b261-4df6-b440-d32a3e578645] Deallocating network for instance {{(pid=61923) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 520.936924] env[61923]: DEBUG nova.network.neutron [-] [instance: a134ab6d-b261-4df6-b440-d32a3e578645] deallocate_for_instance() {{(pid=61923) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 520.979142] env[61923]: DEBUG nova.network.neutron [-] [instance: a134ab6d-b261-4df6-b440-d32a3e578645] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 521.024258] env[61923]: DEBUG nova.compute.utils [None req-35a5fc5c-7e44-4806-b390-80f5a8a499d8 tempest-DeleteServersAdminTestJSON-590116201 tempest-DeleteServersAdminTestJSON-590116201-project-member] Using /dev/sd instead of None {{(pid=61923) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 521.026244] env[61923]: DEBUG nova.compute.manager [None req-35a5fc5c-7e44-4806-b390-80f5a8a499d8 tempest-DeleteServersAdminTestJSON-590116201 tempest-DeleteServersAdminTestJSON-590116201-project-member] [instance: 9ea2600a-d439-4fc3-a131-9bb93250d8cd] Allocating IP information in the background. {{(pid=61923) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 521.026404] env[61923]: DEBUG nova.network.neutron [None req-35a5fc5c-7e44-4806-b390-80f5a8a499d8 tempest-DeleteServersAdminTestJSON-590116201 tempest-DeleteServersAdminTestJSON-590116201-project-member] [instance: 9ea2600a-d439-4fc3-a131-9bb93250d8cd] allocate_for_instance() {{(pid=61923) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 521.093101] env[61923]: DEBUG oslo_concurrency.lockutils [None req-7781f812-6fae-4bd6-9d91-ec6098c06372 tempest-InstanceActionsTestJSON-2015629344 tempest-InstanceActionsTestJSON-2015629344-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 521.131505] env[61923]: DEBUG nova.policy [None req-35a5fc5c-7e44-4806-b390-80f5a8a499d8 tempest-DeleteServersAdminTestJSON-590116201 tempest-DeleteServersAdminTestJSON-590116201-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '379d4ef2827847eeadcb2728bd5859f4', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8786415c53184755b01f324d83b72b9a', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61923) authorize /opt/stack/nova/nova/policy.py:201}} [ 521.141604] env[61923]: DEBUG nova.network.neutron [-] [instance: 242f5cab-3c0c-4f8d-82a8-0cae8d3d419e] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 521.436993] env[61923]: DEBUG nova.compute.manager [None req-844a2215-1038-4a30-8a79-40c1c3117bde tempest-FloatingIPsAssociationNegativeTestJSON-1301210771 tempest-FloatingIPsAssociationNegativeTestJSON-1301210771-project-member] [instance: 5d57303a-ae17-4b92-8f12-303677eb0d7e] Starting instance... {{(pid=61923) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 521.482620] env[61923]: DEBUG nova.network.neutron [-] [instance: a134ab6d-b261-4df6-b440-d32a3e578645] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 521.531711] env[61923]: DEBUG nova.compute.manager [None req-35a5fc5c-7e44-4806-b390-80f5a8a499d8 tempest-DeleteServersAdminTestJSON-590116201 tempest-DeleteServersAdminTestJSON-590116201-project-member] [instance: 9ea2600a-d439-4fc3-a131-9bb93250d8cd] Start building block device mappings for instance. {{(pid=61923) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 521.549236] env[61923]: DEBUG nova.compute.manager [req-6e480063-d329-4352-92bd-7b062f222935 req-94b1703a-4ffc-4e2b-a56a-a4b1dd2385e0 service nova] [instance: 242f5cab-3c0c-4f8d-82a8-0cae8d3d419e] Received event network-changed-88c39c48-cf39-4b87-abda-5dbe55ea29c8 {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 521.549465] env[61923]: DEBUG nova.compute.manager [req-6e480063-d329-4352-92bd-7b062f222935 req-94b1703a-4ffc-4e2b-a56a-a4b1dd2385e0 service nova] [instance: 242f5cab-3c0c-4f8d-82a8-0cae8d3d419e] Refreshing instance network info cache due to event network-changed-88c39c48-cf39-4b87-abda-5dbe55ea29c8. {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 521.549690] env[61923]: DEBUG oslo_concurrency.lockutils [req-6e480063-d329-4352-92bd-7b062f222935 req-94b1703a-4ffc-4e2b-a56a-a4b1dd2385e0 service nova] Acquiring lock "refresh_cache-242f5cab-3c0c-4f8d-82a8-0cae8d3d419e" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 521.549866] env[61923]: DEBUG oslo_concurrency.lockutils [req-6e480063-d329-4352-92bd-7b062f222935 req-94b1703a-4ffc-4e2b-a56a-a4b1dd2385e0 service nova] Acquired lock "refresh_cache-242f5cab-3c0c-4f8d-82a8-0cae8d3d419e" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 521.550046] env[61923]: DEBUG nova.network.neutron [req-6e480063-d329-4352-92bd-7b062f222935 req-94b1703a-4ffc-4e2b-a56a-a4b1dd2385e0 service nova] [instance: 242f5cab-3c0c-4f8d-82a8-0cae8d3d419e] Refreshing network info cache for port 88c39c48-cf39-4b87-abda-5dbe55ea29c8 {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 521.645450] env[61923]: INFO nova.compute.manager [-] [instance: 242f5cab-3c0c-4f8d-82a8-0cae8d3d419e] Took 1.08 seconds to deallocate network for instance. [ 521.648893] env[61923]: DEBUG nova.compute.claims [None req-dc345cdf-ff58-4514-9baa-d2d8d13601af tempest-MigrationsAdminTest-1311988845 tempest-MigrationsAdminTest-1311988845-project-member] [instance: 242f5cab-3c0c-4f8d-82a8-0cae8d3d419e] Aborting claim: {{(pid=61923) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 521.649583] env[61923]: DEBUG oslo_concurrency.lockutils [None req-dc345cdf-ff58-4514-9baa-d2d8d13601af tempest-MigrationsAdminTest-1311988845 tempest-MigrationsAdminTest-1311988845-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 521.692102] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5f7a569-8478-4f77-8918-d9a93626134c {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 521.701585] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1275ba9-20bd-4d44-84b9-ad728596f64e {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 521.740610] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca92009b-8d84-48e8-aa9e-45c561e2e498 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 521.748139] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b9ffbd0-4231-4df3-90af-a816bab62128 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 521.762612] env[61923]: DEBUG nova.compute.provider_tree [None req-464a16f8-1845-4d4d-b3ea-fd8ba00de8bb tempest-ServersAdminNegativeTestJSON-1828705899 tempest-ServersAdminNegativeTestJSON-1828705899-project-member] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 521.846754] env[61923]: DEBUG nova.network.neutron [None req-35a5fc5c-7e44-4806-b390-80f5a8a499d8 tempest-DeleteServersAdminTestJSON-590116201 tempest-DeleteServersAdminTestJSON-590116201-project-member] [instance: 9ea2600a-d439-4fc3-a131-9bb93250d8cd] Successfully created port: a1ecfb50-9098-4408-b9bd-8db498ae622e {{(pid=61923) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 521.955663] env[61923]: DEBUG oslo_concurrency.lockutils [None req-f30f4038-8fd0-4bbd-8b35-7fb8e0dea3a4 tempest-SecurityGroupsTestJSON-509490211 tempest-SecurityGroupsTestJSON-509490211-project-member] Acquiring lock "59fc11e8-27be-4553-9de9-3eb5ea612ec4" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 521.955772] env[61923]: DEBUG oslo_concurrency.lockutils [None req-f30f4038-8fd0-4bbd-8b35-7fb8e0dea3a4 tempest-SecurityGroupsTestJSON-509490211 tempest-SecurityGroupsTestJSON-509490211-project-member] Lock "59fc11e8-27be-4553-9de9-3eb5ea612ec4" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 521.982856] env[61923]: DEBUG nova.compute.manager [req-383fe115-b7e1-4dfc-9c0a-8a6616c3d6a2 req-b134f071-e222-42a4-9f4c-efa5274736a4 service nova] [instance: a134ab6d-b261-4df6-b440-d32a3e578645] Received event network-changed-6aa16849-cb00-49ad-9abe-b5ca6484171d {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 521.984472] env[61923]: DEBUG nova.compute.manager [req-383fe115-b7e1-4dfc-9c0a-8a6616c3d6a2 req-b134f071-e222-42a4-9f4c-efa5274736a4 service nova] [instance: a134ab6d-b261-4df6-b440-d32a3e578645] Refreshing instance network info cache due to event network-changed-6aa16849-cb00-49ad-9abe-b5ca6484171d. {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 521.984706] env[61923]: DEBUG oslo_concurrency.lockutils [req-383fe115-b7e1-4dfc-9c0a-8a6616c3d6a2 req-b134f071-e222-42a4-9f4c-efa5274736a4 service nova] Acquiring lock "refresh_cache-a134ab6d-b261-4df6-b440-d32a3e578645" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 521.984818] env[61923]: DEBUG oslo_concurrency.lockutils [req-383fe115-b7e1-4dfc-9c0a-8a6616c3d6a2 req-b134f071-e222-42a4-9f4c-efa5274736a4 service nova] Acquired lock "refresh_cache-a134ab6d-b261-4df6-b440-d32a3e578645" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 521.984959] env[61923]: DEBUG nova.network.neutron [req-383fe115-b7e1-4dfc-9c0a-8a6616c3d6a2 req-b134f071-e222-42a4-9f4c-efa5274736a4 service nova] [instance: a134ab6d-b261-4df6-b440-d32a3e578645] Refreshing network info cache for port 6aa16849-cb00-49ad-9abe-b5ca6484171d {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 521.988779] env[61923]: DEBUG oslo_concurrency.lockutils [None req-844a2215-1038-4a30-8a79-40c1c3117bde tempest-FloatingIPsAssociationNegativeTestJSON-1301210771 tempest-FloatingIPsAssociationNegativeTestJSON-1301210771-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 521.989601] env[61923]: INFO nova.compute.manager [-] [instance: a134ab6d-b261-4df6-b440-d32a3e578645] Took 1.05 seconds to deallocate network for instance. [ 521.992125] env[61923]: DEBUG nova.compute.claims [None req-a2070053-8e36-4abe-a723-0b93480b3283 tempest-ServerDiagnosticsTest-1095156200 tempest-ServerDiagnosticsTest-1095156200-project-member] [instance: a134ab6d-b261-4df6-b440-d32a3e578645] Aborting claim: {{(pid=61923) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 521.992305] env[61923]: DEBUG oslo_concurrency.lockutils [None req-a2070053-8e36-4abe-a723-0b93480b3283 tempest-ServerDiagnosticsTest-1095156200 tempest-ServerDiagnosticsTest-1095156200-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 522.111490] env[61923]: DEBUG nova.network.neutron [req-6e480063-d329-4352-92bd-7b062f222935 req-94b1703a-4ffc-4e2b-a56a-a4b1dd2385e0 service nova] [instance: 242f5cab-3c0c-4f8d-82a8-0cae8d3d419e] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 522.232044] env[61923]: DEBUG nova.network.neutron [req-6e480063-d329-4352-92bd-7b062f222935 req-94b1703a-4ffc-4e2b-a56a-a4b1dd2385e0 service nova] [instance: 242f5cab-3c0c-4f8d-82a8-0cae8d3d419e] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 522.269357] env[61923]: DEBUG nova.scheduler.client.report [None req-464a16f8-1845-4d4d-b3ea-fd8ba00de8bb tempest-ServersAdminNegativeTestJSON-1828705899 tempest-ServersAdminNegativeTestJSON-1828705899-project-member] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 522.458459] env[61923]: DEBUG nova.compute.manager [None req-f30f4038-8fd0-4bbd-8b35-7fb8e0dea3a4 tempest-SecurityGroupsTestJSON-509490211 tempest-SecurityGroupsTestJSON-509490211-project-member] [instance: 59fc11e8-27be-4553-9de9-3eb5ea612ec4] Starting instance... {{(pid=61923) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 522.513203] env[61923]: DEBUG nova.network.neutron [req-383fe115-b7e1-4dfc-9c0a-8a6616c3d6a2 req-b134f071-e222-42a4-9f4c-efa5274736a4 service nova] [instance: a134ab6d-b261-4df6-b440-d32a3e578645] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 522.546300] env[61923]: DEBUG nova.compute.manager [None req-35a5fc5c-7e44-4806-b390-80f5a8a499d8 tempest-DeleteServersAdminTestJSON-590116201 tempest-DeleteServersAdminTestJSON-590116201-project-member] [instance: 9ea2600a-d439-4fc3-a131-9bb93250d8cd] Start spawning the instance on the hypervisor. {{(pid=61923) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 522.586185] env[61923]: DEBUG nova.virt.hardware [None req-35a5fc5c-7e44-4806-b390-80f5a8a499d8 tempest-DeleteServersAdminTestJSON-590116201 tempest-DeleteServersAdminTestJSON-590116201-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-29T20:07:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-29T20:07:35Z,direct_url=,disk_format='vmdk',id=0f153f63-ae0a-45b1-b7f5-7f9b673c947f,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='4e7b5e3b3c3443c8bd5c70f8a15739f5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-29T20:07:36Z,virtual_size=,visibility=), allow threads: False {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 522.586363] env[61923]: DEBUG nova.virt.hardware [None req-35a5fc5c-7e44-4806-b390-80f5a8a499d8 tempest-DeleteServersAdminTestJSON-590116201 tempest-DeleteServersAdminTestJSON-590116201-project-member] Flavor limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 522.586619] env[61923]: DEBUG nova.virt.hardware [None req-35a5fc5c-7e44-4806-b390-80f5a8a499d8 tempest-DeleteServersAdminTestJSON-590116201 tempest-DeleteServersAdminTestJSON-590116201-project-member] Image limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 522.586793] env[61923]: DEBUG nova.virt.hardware [None req-35a5fc5c-7e44-4806-b390-80f5a8a499d8 tempest-DeleteServersAdminTestJSON-590116201 tempest-DeleteServersAdminTestJSON-590116201-project-member] Flavor pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 522.587028] env[61923]: DEBUG nova.virt.hardware [None req-35a5fc5c-7e44-4806-b390-80f5a8a499d8 tempest-DeleteServersAdminTestJSON-590116201 tempest-DeleteServersAdminTestJSON-590116201-project-member] Image pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 522.587211] env[61923]: DEBUG nova.virt.hardware [None req-35a5fc5c-7e44-4806-b390-80f5a8a499d8 tempest-DeleteServersAdminTestJSON-590116201 tempest-DeleteServersAdminTestJSON-590116201-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 522.587426] env[61923]: DEBUG nova.virt.hardware [None req-35a5fc5c-7e44-4806-b390-80f5a8a499d8 tempest-DeleteServersAdminTestJSON-590116201 tempest-DeleteServersAdminTestJSON-590116201-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 522.587579] env[61923]: DEBUG nova.virt.hardware [None req-35a5fc5c-7e44-4806-b390-80f5a8a499d8 tempest-DeleteServersAdminTestJSON-590116201 tempest-DeleteServersAdminTestJSON-590116201-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 522.587737] env[61923]: DEBUG nova.virt.hardware [None req-35a5fc5c-7e44-4806-b390-80f5a8a499d8 tempest-DeleteServersAdminTestJSON-590116201 tempest-DeleteServersAdminTestJSON-590116201-project-member] Got 1 possible topologies {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 522.587908] env[61923]: DEBUG nova.virt.hardware [None req-35a5fc5c-7e44-4806-b390-80f5a8a499d8 tempest-DeleteServersAdminTestJSON-590116201 tempest-DeleteServersAdminTestJSON-590116201-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 522.588072] env[61923]: DEBUG nova.virt.hardware [None req-35a5fc5c-7e44-4806-b390-80f5a8a499d8 tempest-DeleteServersAdminTestJSON-590116201 tempest-DeleteServersAdminTestJSON-590116201-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 522.589528] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab92d171-96ac-427e-b091-8dc6aee41ae8 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 522.594571] env[61923]: DEBUG nova.network.neutron [req-383fe115-b7e1-4dfc-9c0a-8a6616c3d6a2 req-b134f071-e222-42a4-9f4c-efa5274736a4 service nova] [instance: a134ab6d-b261-4df6-b440-d32a3e578645] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 522.601549] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39369881-331f-40d2-aeaa-dc97094d04b3 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 522.735472] env[61923]: DEBUG oslo_concurrency.lockutils [req-6e480063-d329-4352-92bd-7b062f222935 req-94b1703a-4ffc-4e2b-a56a-a4b1dd2385e0 service nova] Releasing lock "refresh_cache-242f5cab-3c0c-4f8d-82a8-0cae8d3d419e" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 522.773315] env[61923]: DEBUG oslo_concurrency.lockutils [None req-464a16f8-1845-4d4d-b3ea-fd8ba00de8bb tempest-ServersAdminNegativeTestJSON-1828705899 tempest-ServersAdminNegativeTestJSON-1828705899-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.255s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 522.773826] env[61923]: DEBUG nova.compute.manager [None req-464a16f8-1845-4d4d-b3ea-fd8ba00de8bb tempest-ServersAdminNegativeTestJSON-1828705899 tempest-ServersAdminNegativeTestJSON-1828705899-project-member] [instance: 1ade1b37-8319-4d74-9544-6b83e8738498] Start building networks asynchronously for instance. {{(pid=61923) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 522.777558] env[61923]: DEBUG oslo_concurrency.lockutils [None req-8a6b6bc2-54c9-4389-ab30-057158d86227 tempest-ServerDiagnosticsNegativeTest-664316905 tempest-ServerDiagnosticsNegativeTest-664316905-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 3.870s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 522.779349] env[61923]: INFO nova.compute.claims [None req-8a6b6bc2-54c9-4389-ab30-057158d86227 tempest-ServerDiagnosticsNegativeTest-664316905 tempest-ServerDiagnosticsNegativeTest-664316905-project-member] [instance: b0a45314-eed8-4d0a-bea2-d4b47dc0bc80] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 523.033645] env[61923]: DEBUG oslo_concurrency.lockutils [None req-f30f4038-8fd0-4bbd-8b35-7fb8e0dea3a4 tempest-SecurityGroupsTestJSON-509490211 tempest-SecurityGroupsTestJSON-509490211-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 523.097476] env[61923]: DEBUG oslo_concurrency.lockutils [req-383fe115-b7e1-4dfc-9c0a-8a6616c3d6a2 req-b134f071-e222-42a4-9f4c-efa5274736a4 service nova] Releasing lock "refresh_cache-a134ab6d-b261-4df6-b440-d32a3e578645" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 523.286820] env[61923]: DEBUG nova.compute.utils [None req-464a16f8-1845-4d4d-b3ea-fd8ba00de8bb tempest-ServersAdminNegativeTestJSON-1828705899 tempest-ServersAdminNegativeTestJSON-1828705899-project-member] Using /dev/sd instead of None {{(pid=61923) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 523.288218] env[61923]: DEBUG nova.compute.manager [None req-464a16f8-1845-4d4d-b3ea-fd8ba00de8bb tempest-ServersAdminNegativeTestJSON-1828705899 tempest-ServersAdminNegativeTestJSON-1828705899-project-member] [instance: 1ade1b37-8319-4d74-9544-6b83e8738498] Allocating IP information in the background. {{(pid=61923) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 523.288382] env[61923]: DEBUG nova.network.neutron [None req-464a16f8-1845-4d4d-b3ea-fd8ba00de8bb tempest-ServersAdminNegativeTestJSON-1828705899 tempest-ServersAdminNegativeTestJSON-1828705899-project-member] [instance: 1ade1b37-8319-4d74-9544-6b83e8738498] allocate_for_instance() {{(pid=61923) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 523.386990] env[61923]: DEBUG nova.policy [None req-464a16f8-1845-4d4d-b3ea-fd8ba00de8bb tempest-ServersAdminNegativeTestJSON-1828705899 tempest-ServersAdminNegativeTestJSON-1828705899-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0a002d5cc4ff48b0a41dc0f6c4ede86d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4e3eb68dec6748cc96d2ada081ba4579', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61923) authorize /opt/stack/nova/nova/policy.py:201}} [ 523.563112] env[61923]: ERROR nova.compute.manager [None req-35a5fc5c-7e44-4806-b390-80f5a8a499d8 tempest-DeleteServersAdminTestJSON-590116201 tempest-DeleteServersAdminTestJSON-590116201-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port a1ecfb50-9098-4408-b9bd-8db498ae622e, please check neutron logs for more information. [ 523.563112] env[61923]: ERROR nova.compute.manager Traceback (most recent call last): [ 523.563112] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 523.563112] env[61923]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 523.563112] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 523.563112] env[61923]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 523.563112] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 523.563112] env[61923]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 523.563112] env[61923]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 523.563112] env[61923]: ERROR nova.compute.manager self.force_reraise() [ 523.563112] env[61923]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 523.563112] env[61923]: ERROR nova.compute.manager raise self.value [ 523.563112] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 523.563112] env[61923]: ERROR nova.compute.manager updated_port = self._update_port( [ 523.563112] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 523.563112] env[61923]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 523.563862] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 523.563862] env[61923]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 523.563862] env[61923]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port a1ecfb50-9098-4408-b9bd-8db498ae622e, please check neutron logs for more information. [ 523.563862] env[61923]: ERROR nova.compute.manager [ 523.563862] env[61923]: Traceback (most recent call last): [ 523.563862] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 523.563862] env[61923]: listener.cb(fileno) [ 523.563862] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 523.563862] env[61923]: result = function(*args, **kwargs) [ 523.563862] env[61923]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 523.563862] env[61923]: return func(*args, **kwargs) [ 523.563862] env[61923]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 523.563862] env[61923]: raise e [ 523.563862] env[61923]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 523.563862] env[61923]: nwinfo = self.network_api.allocate_for_instance( [ 523.563862] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 523.563862] env[61923]: created_port_ids = self._update_ports_for_instance( [ 523.563862] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 523.563862] env[61923]: with excutils.save_and_reraise_exception(): [ 523.563862] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 523.563862] env[61923]: self.force_reraise() [ 523.563862] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 523.563862] env[61923]: raise self.value [ 523.563862] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 523.563862] env[61923]: updated_port = self._update_port( [ 523.563862] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 523.563862] env[61923]: _ensure_no_port_binding_failure(port) [ 523.563862] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 523.563862] env[61923]: raise exception.PortBindingFailed(port_id=port['id']) [ 523.564827] env[61923]: nova.exception.PortBindingFailed: Binding failed for port a1ecfb50-9098-4408-b9bd-8db498ae622e, please check neutron logs for more information. [ 523.564827] env[61923]: Removing descriptor: 15 [ 523.564827] env[61923]: ERROR nova.compute.manager [None req-35a5fc5c-7e44-4806-b390-80f5a8a499d8 tempest-DeleteServersAdminTestJSON-590116201 tempest-DeleteServersAdminTestJSON-590116201-project-member] [instance: 9ea2600a-d439-4fc3-a131-9bb93250d8cd] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port a1ecfb50-9098-4408-b9bd-8db498ae622e, please check neutron logs for more information. [ 523.564827] env[61923]: ERROR nova.compute.manager [instance: 9ea2600a-d439-4fc3-a131-9bb93250d8cd] Traceback (most recent call last): [ 523.564827] env[61923]: ERROR nova.compute.manager [instance: 9ea2600a-d439-4fc3-a131-9bb93250d8cd] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 523.564827] env[61923]: ERROR nova.compute.manager [instance: 9ea2600a-d439-4fc3-a131-9bb93250d8cd] yield resources [ 523.564827] env[61923]: ERROR nova.compute.manager [instance: 9ea2600a-d439-4fc3-a131-9bb93250d8cd] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 523.564827] env[61923]: ERROR nova.compute.manager [instance: 9ea2600a-d439-4fc3-a131-9bb93250d8cd] self.driver.spawn(context, instance, image_meta, [ 523.564827] env[61923]: ERROR nova.compute.manager [instance: 9ea2600a-d439-4fc3-a131-9bb93250d8cd] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 523.564827] env[61923]: ERROR nova.compute.manager [instance: 9ea2600a-d439-4fc3-a131-9bb93250d8cd] self._vmops.spawn(context, instance, image_meta, injected_files, [ 523.564827] env[61923]: ERROR nova.compute.manager [instance: 9ea2600a-d439-4fc3-a131-9bb93250d8cd] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 523.564827] env[61923]: ERROR nova.compute.manager [instance: 9ea2600a-d439-4fc3-a131-9bb93250d8cd] vm_ref = self.build_virtual_machine(instance, [ 523.565236] env[61923]: ERROR nova.compute.manager [instance: 9ea2600a-d439-4fc3-a131-9bb93250d8cd] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 523.565236] env[61923]: ERROR nova.compute.manager [instance: 9ea2600a-d439-4fc3-a131-9bb93250d8cd] vif_infos = vmwarevif.get_vif_info(self._session, [ 523.565236] env[61923]: ERROR nova.compute.manager [instance: 9ea2600a-d439-4fc3-a131-9bb93250d8cd] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 523.565236] env[61923]: ERROR nova.compute.manager [instance: 9ea2600a-d439-4fc3-a131-9bb93250d8cd] for vif in network_info: [ 523.565236] env[61923]: ERROR nova.compute.manager [instance: 9ea2600a-d439-4fc3-a131-9bb93250d8cd] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 523.565236] env[61923]: ERROR nova.compute.manager [instance: 9ea2600a-d439-4fc3-a131-9bb93250d8cd] return self._sync_wrapper(fn, *args, **kwargs) [ 523.565236] env[61923]: ERROR nova.compute.manager [instance: 9ea2600a-d439-4fc3-a131-9bb93250d8cd] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 523.565236] env[61923]: ERROR nova.compute.manager [instance: 9ea2600a-d439-4fc3-a131-9bb93250d8cd] self.wait() [ 523.565236] env[61923]: ERROR nova.compute.manager [instance: 9ea2600a-d439-4fc3-a131-9bb93250d8cd] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 523.565236] env[61923]: ERROR nova.compute.manager [instance: 9ea2600a-d439-4fc3-a131-9bb93250d8cd] self[:] = self._gt.wait() [ 523.565236] env[61923]: ERROR nova.compute.manager [instance: 9ea2600a-d439-4fc3-a131-9bb93250d8cd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 523.565236] env[61923]: ERROR nova.compute.manager [instance: 9ea2600a-d439-4fc3-a131-9bb93250d8cd] return self._exit_event.wait() [ 523.565236] env[61923]: ERROR nova.compute.manager [instance: 9ea2600a-d439-4fc3-a131-9bb93250d8cd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 523.565653] env[61923]: ERROR nova.compute.manager [instance: 9ea2600a-d439-4fc3-a131-9bb93250d8cd] result = hub.switch() [ 523.565653] env[61923]: ERROR nova.compute.manager [instance: 9ea2600a-d439-4fc3-a131-9bb93250d8cd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 523.565653] env[61923]: ERROR nova.compute.manager [instance: 9ea2600a-d439-4fc3-a131-9bb93250d8cd] return self.greenlet.switch() [ 523.565653] env[61923]: ERROR nova.compute.manager [instance: 9ea2600a-d439-4fc3-a131-9bb93250d8cd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 523.565653] env[61923]: ERROR nova.compute.manager [instance: 9ea2600a-d439-4fc3-a131-9bb93250d8cd] result = function(*args, **kwargs) [ 523.565653] env[61923]: ERROR nova.compute.manager [instance: 9ea2600a-d439-4fc3-a131-9bb93250d8cd] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 523.565653] env[61923]: ERROR nova.compute.manager [instance: 9ea2600a-d439-4fc3-a131-9bb93250d8cd] return func(*args, **kwargs) [ 523.565653] env[61923]: ERROR nova.compute.manager [instance: 9ea2600a-d439-4fc3-a131-9bb93250d8cd] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 523.565653] env[61923]: ERROR nova.compute.manager [instance: 9ea2600a-d439-4fc3-a131-9bb93250d8cd] raise e [ 523.565653] env[61923]: ERROR nova.compute.manager [instance: 9ea2600a-d439-4fc3-a131-9bb93250d8cd] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 523.565653] env[61923]: ERROR nova.compute.manager [instance: 9ea2600a-d439-4fc3-a131-9bb93250d8cd] nwinfo = self.network_api.allocate_for_instance( [ 523.565653] env[61923]: ERROR nova.compute.manager [instance: 9ea2600a-d439-4fc3-a131-9bb93250d8cd] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 523.565653] env[61923]: ERROR nova.compute.manager [instance: 9ea2600a-d439-4fc3-a131-9bb93250d8cd] created_port_ids = self._update_ports_for_instance( [ 523.569396] env[61923]: ERROR nova.compute.manager [instance: 9ea2600a-d439-4fc3-a131-9bb93250d8cd] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 523.569396] env[61923]: ERROR nova.compute.manager [instance: 9ea2600a-d439-4fc3-a131-9bb93250d8cd] with excutils.save_and_reraise_exception(): [ 523.569396] env[61923]: ERROR nova.compute.manager [instance: 9ea2600a-d439-4fc3-a131-9bb93250d8cd] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 523.569396] env[61923]: ERROR nova.compute.manager [instance: 9ea2600a-d439-4fc3-a131-9bb93250d8cd] self.force_reraise() [ 523.569396] env[61923]: ERROR nova.compute.manager [instance: 9ea2600a-d439-4fc3-a131-9bb93250d8cd] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 523.569396] env[61923]: ERROR nova.compute.manager [instance: 9ea2600a-d439-4fc3-a131-9bb93250d8cd] raise self.value [ 523.569396] env[61923]: ERROR nova.compute.manager [instance: 9ea2600a-d439-4fc3-a131-9bb93250d8cd] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 523.569396] env[61923]: ERROR nova.compute.manager [instance: 9ea2600a-d439-4fc3-a131-9bb93250d8cd] updated_port = self._update_port( [ 523.569396] env[61923]: ERROR nova.compute.manager [instance: 9ea2600a-d439-4fc3-a131-9bb93250d8cd] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 523.569396] env[61923]: ERROR nova.compute.manager [instance: 9ea2600a-d439-4fc3-a131-9bb93250d8cd] _ensure_no_port_binding_failure(port) [ 523.569396] env[61923]: ERROR nova.compute.manager [instance: 9ea2600a-d439-4fc3-a131-9bb93250d8cd] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 523.569396] env[61923]: ERROR nova.compute.manager [instance: 9ea2600a-d439-4fc3-a131-9bb93250d8cd] raise exception.PortBindingFailed(port_id=port['id']) [ 523.570082] env[61923]: ERROR nova.compute.manager [instance: 9ea2600a-d439-4fc3-a131-9bb93250d8cd] nova.exception.PortBindingFailed: Binding failed for port a1ecfb50-9098-4408-b9bd-8db498ae622e, please check neutron logs for more information. [ 523.570082] env[61923]: ERROR nova.compute.manager [instance: 9ea2600a-d439-4fc3-a131-9bb93250d8cd] [ 523.570082] env[61923]: INFO nova.compute.manager [None req-35a5fc5c-7e44-4806-b390-80f5a8a499d8 tempest-DeleteServersAdminTestJSON-590116201 tempest-DeleteServersAdminTestJSON-590116201-project-member] [instance: 9ea2600a-d439-4fc3-a131-9bb93250d8cd] Terminating instance [ 523.570082] env[61923]: DEBUG oslo_concurrency.lockutils [None req-35a5fc5c-7e44-4806-b390-80f5a8a499d8 tempest-DeleteServersAdminTestJSON-590116201 tempest-DeleteServersAdminTestJSON-590116201-project-member] Acquiring lock "refresh_cache-9ea2600a-d439-4fc3-a131-9bb93250d8cd" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 523.570082] env[61923]: DEBUG oslo_concurrency.lockutils [None req-35a5fc5c-7e44-4806-b390-80f5a8a499d8 tempest-DeleteServersAdminTestJSON-590116201 tempest-DeleteServersAdminTestJSON-590116201-project-member] Acquired lock "refresh_cache-9ea2600a-d439-4fc3-a131-9bb93250d8cd" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 523.570082] env[61923]: DEBUG nova.network.neutron [None req-35a5fc5c-7e44-4806-b390-80f5a8a499d8 tempest-DeleteServersAdminTestJSON-590116201 tempest-DeleteServersAdminTestJSON-590116201-project-member] [instance: 9ea2600a-d439-4fc3-a131-9bb93250d8cd] Building network info cache for instance {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 523.754445] env[61923]: DEBUG nova.network.neutron [None req-464a16f8-1845-4d4d-b3ea-fd8ba00de8bb tempest-ServersAdminNegativeTestJSON-1828705899 tempest-ServersAdminNegativeTestJSON-1828705899-project-member] [instance: 1ade1b37-8319-4d74-9544-6b83e8738498] Successfully created port: c188410e-a247-4617-a118-46ee79a2683c {{(pid=61923) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 523.794375] env[61923]: DEBUG nova.compute.manager [None req-464a16f8-1845-4d4d-b3ea-fd8ba00de8bb tempest-ServersAdminNegativeTestJSON-1828705899 tempest-ServersAdminNegativeTestJSON-1828705899-project-member] [instance: 1ade1b37-8319-4d74-9544-6b83e8738498] Start building block device mappings for instance. {{(pid=61923) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 523.948568] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b191936-13bb-4baf-a0c4-da9c947dcc3f {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 523.956745] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be48bb57-ee77-4735-8a0f-7659827827f5 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 523.993100] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05d69d6b-55b1-46a7-9e60-c72bb560bee7 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 524.000538] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eaf9e840-45b2-4cdb-a56a-453a9893529b {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 524.015398] env[61923]: DEBUG nova.compute.provider_tree [None req-8a6b6bc2-54c9-4389-ab30-057158d86227 tempest-ServerDiagnosticsNegativeTest-664316905 tempest-ServerDiagnosticsNegativeTest-664316905-project-member] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 524.094275] env[61923]: DEBUG nova.network.neutron [None req-35a5fc5c-7e44-4806-b390-80f5a8a499d8 tempest-DeleteServersAdminTestJSON-590116201 tempest-DeleteServersAdminTestJSON-590116201-project-member] [instance: 9ea2600a-d439-4fc3-a131-9bb93250d8cd] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 524.153215] env[61923]: DEBUG nova.network.neutron [None req-35a5fc5c-7e44-4806-b390-80f5a8a499d8 tempest-DeleteServersAdminTestJSON-590116201 tempest-DeleteServersAdminTestJSON-590116201-project-member] [instance: 9ea2600a-d439-4fc3-a131-9bb93250d8cd] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 524.519541] env[61923]: DEBUG nova.scheduler.client.report [None req-8a6b6bc2-54c9-4389-ab30-057158d86227 tempest-ServerDiagnosticsNegativeTest-664316905 tempest-ServerDiagnosticsNegativeTest-664316905-project-member] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 524.662182] env[61923]: DEBUG oslo_concurrency.lockutils [None req-35a5fc5c-7e44-4806-b390-80f5a8a499d8 tempest-DeleteServersAdminTestJSON-590116201 tempest-DeleteServersAdminTestJSON-590116201-project-member] Releasing lock "refresh_cache-9ea2600a-d439-4fc3-a131-9bb93250d8cd" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 524.662541] env[61923]: DEBUG nova.compute.manager [None req-35a5fc5c-7e44-4806-b390-80f5a8a499d8 tempest-DeleteServersAdminTestJSON-590116201 tempest-DeleteServersAdminTestJSON-590116201-project-member] [instance: 9ea2600a-d439-4fc3-a131-9bb93250d8cd] Start destroying the instance on the hypervisor. {{(pid=61923) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 524.662739] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-35a5fc5c-7e44-4806-b390-80f5a8a499d8 tempest-DeleteServersAdminTestJSON-590116201 tempest-DeleteServersAdminTestJSON-590116201-project-member] [instance: 9ea2600a-d439-4fc3-a131-9bb93250d8cd] Destroying instance {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 524.664063] env[61923]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d91c3ab9-afc5-4a59-b6e8-64345c8c465b {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 524.675146] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dcf27a15-871e-47dd-8f01-f3ef9a2831d7 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 524.691053] env[61923]: DEBUG nova.compute.manager [req-ed1710d2-9a07-48ee-9ffa-06420fef14af req-d83e5e71-2189-477a-8044-58a9af2a9eec service nova] [instance: 242f5cab-3c0c-4f8d-82a8-0cae8d3d419e] Received event network-vif-deleted-88c39c48-cf39-4b87-abda-5dbe55ea29c8 {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 524.704124] env[61923]: WARNING nova.virt.vmwareapi.vmops [None req-35a5fc5c-7e44-4806-b390-80f5a8a499d8 tempest-DeleteServersAdminTestJSON-590116201 tempest-DeleteServersAdminTestJSON-590116201-project-member] [instance: 9ea2600a-d439-4fc3-a131-9bb93250d8cd] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 9ea2600a-d439-4fc3-a131-9bb93250d8cd could not be found. [ 524.704350] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-35a5fc5c-7e44-4806-b390-80f5a8a499d8 tempest-DeleteServersAdminTestJSON-590116201 tempest-DeleteServersAdminTestJSON-590116201-project-member] [instance: 9ea2600a-d439-4fc3-a131-9bb93250d8cd] Instance destroyed {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 524.704529] env[61923]: INFO nova.compute.manager [None req-35a5fc5c-7e44-4806-b390-80f5a8a499d8 tempest-DeleteServersAdminTestJSON-590116201 tempest-DeleteServersAdminTestJSON-590116201-project-member] [instance: 9ea2600a-d439-4fc3-a131-9bb93250d8cd] Took 0.04 seconds to destroy the instance on the hypervisor. [ 524.704954] env[61923]: DEBUG oslo.service.loopingcall [None req-35a5fc5c-7e44-4806-b390-80f5a8a499d8 tempest-DeleteServersAdminTestJSON-590116201 tempest-DeleteServersAdminTestJSON-590116201-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61923) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 524.705170] env[61923]: DEBUG nova.compute.manager [-] [instance: 9ea2600a-d439-4fc3-a131-9bb93250d8cd] Deallocating network for instance {{(pid=61923) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 524.705278] env[61923]: DEBUG nova.network.neutron [-] [instance: 9ea2600a-d439-4fc3-a131-9bb93250d8cd] deallocate_for_instance() {{(pid=61923) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 524.752498] env[61923]: DEBUG nova.network.neutron [-] [instance: 9ea2600a-d439-4fc3-a131-9bb93250d8cd] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 524.804164] env[61923]: DEBUG nova.compute.manager [None req-464a16f8-1845-4d4d-b3ea-fd8ba00de8bb tempest-ServersAdminNegativeTestJSON-1828705899 tempest-ServersAdminNegativeTestJSON-1828705899-project-member] [instance: 1ade1b37-8319-4d74-9544-6b83e8738498] Start spawning the instance on the hypervisor. {{(pid=61923) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 524.838595] env[61923]: DEBUG nova.virt.hardware [None req-464a16f8-1845-4d4d-b3ea-fd8ba00de8bb tempest-ServersAdminNegativeTestJSON-1828705899 tempest-ServersAdminNegativeTestJSON-1828705899-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-29T20:07:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-29T20:07:35Z,direct_url=,disk_format='vmdk',id=0f153f63-ae0a-45b1-b7f5-7f9b673c947f,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='4e7b5e3b3c3443c8bd5c70f8a15739f5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-29T20:07:36Z,virtual_size=,visibility=), allow threads: False {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 524.839158] env[61923]: DEBUG nova.virt.hardware [None req-464a16f8-1845-4d4d-b3ea-fd8ba00de8bb tempest-ServersAdminNegativeTestJSON-1828705899 tempest-ServersAdminNegativeTestJSON-1828705899-project-member] Flavor limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 524.839378] env[61923]: DEBUG nova.virt.hardware [None req-464a16f8-1845-4d4d-b3ea-fd8ba00de8bb tempest-ServersAdminNegativeTestJSON-1828705899 tempest-ServersAdminNegativeTestJSON-1828705899-project-member] Image limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 524.839576] env[61923]: DEBUG nova.virt.hardware [None req-464a16f8-1845-4d4d-b3ea-fd8ba00de8bb tempest-ServersAdminNegativeTestJSON-1828705899 tempest-ServersAdminNegativeTestJSON-1828705899-project-member] Flavor pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 524.839720] env[61923]: DEBUG nova.virt.hardware [None req-464a16f8-1845-4d4d-b3ea-fd8ba00de8bb tempest-ServersAdminNegativeTestJSON-1828705899 tempest-ServersAdminNegativeTestJSON-1828705899-project-member] Image pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 524.840085] env[61923]: DEBUG nova.virt.hardware [None req-464a16f8-1845-4d4d-b3ea-fd8ba00de8bb tempest-ServersAdminNegativeTestJSON-1828705899 tempest-ServersAdminNegativeTestJSON-1828705899-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 524.840329] env[61923]: DEBUG nova.virt.hardware [None req-464a16f8-1845-4d4d-b3ea-fd8ba00de8bb tempest-ServersAdminNegativeTestJSON-1828705899 tempest-ServersAdminNegativeTestJSON-1828705899-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 524.840515] env[61923]: DEBUG nova.virt.hardware [None req-464a16f8-1845-4d4d-b3ea-fd8ba00de8bb tempest-ServersAdminNegativeTestJSON-1828705899 tempest-ServersAdminNegativeTestJSON-1828705899-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 524.840649] env[61923]: DEBUG nova.virt.hardware [None req-464a16f8-1845-4d4d-b3ea-fd8ba00de8bb tempest-ServersAdminNegativeTestJSON-1828705899 tempest-ServersAdminNegativeTestJSON-1828705899-project-member] Got 1 possible topologies {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 524.840977] env[61923]: DEBUG nova.virt.hardware [None req-464a16f8-1845-4d4d-b3ea-fd8ba00de8bb tempest-ServersAdminNegativeTestJSON-1828705899 tempest-ServersAdminNegativeTestJSON-1828705899-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 524.841193] env[61923]: DEBUG nova.virt.hardware [None req-464a16f8-1845-4d4d-b3ea-fd8ba00de8bb tempest-ServersAdminNegativeTestJSON-1828705899 tempest-ServersAdminNegativeTestJSON-1828705899-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 524.842225] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf41be99-fd0f-4ae9-943b-aaebd3a5277c {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 524.851148] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5527dc3c-4733-4ead-a2e0-4579faafca5d {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 525.029079] env[61923]: DEBUG oslo_concurrency.lockutils [None req-8a6b6bc2-54c9-4389-ab30-057158d86227 tempest-ServerDiagnosticsNegativeTest-664316905 tempest-ServerDiagnosticsNegativeTest-664316905-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.249s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 525.029079] env[61923]: DEBUG nova.compute.manager [None req-8a6b6bc2-54c9-4389-ab30-057158d86227 tempest-ServerDiagnosticsNegativeTest-664316905 tempest-ServerDiagnosticsNegativeTest-664316905-project-member] [instance: b0a45314-eed8-4d0a-bea2-d4b47dc0bc80] Start building networks asynchronously for instance. {{(pid=61923) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 525.031314] env[61923]: DEBUG oslo_concurrency.lockutils [None req-7781f812-6fae-4bd6-9d91-ec6098c06372 tempest-InstanceActionsTestJSON-2015629344 tempest-InstanceActionsTestJSON-2015629344-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 3.938s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 525.034207] env[61923]: INFO nova.compute.claims [None req-7781f812-6fae-4bd6-9d91-ec6098c06372 tempest-InstanceActionsTestJSON-2015629344 tempest-InstanceActionsTestJSON-2015629344-project-member] [instance: 97148ea5-398a-4974-b8f6-62290e519794] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 525.053304] env[61923]: DEBUG nova.compute.manager [req-855bd337-71e8-405d-8445-bff56e2db847 req-0f64ee1b-4409-4f5f-af15-20e897afe787 service nova] [instance: a134ab6d-b261-4df6-b440-d32a3e578645] Received event network-vif-deleted-6aa16849-cb00-49ad-9abe-b5ca6484171d {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 525.053604] env[61923]: DEBUG nova.compute.manager [req-855bd337-71e8-405d-8445-bff56e2db847 req-0f64ee1b-4409-4f5f-af15-20e897afe787 service nova] [instance: 9ea2600a-d439-4fc3-a131-9bb93250d8cd] Received event network-changed-a1ecfb50-9098-4408-b9bd-8db498ae622e {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 525.054160] env[61923]: DEBUG nova.compute.manager [req-855bd337-71e8-405d-8445-bff56e2db847 req-0f64ee1b-4409-4f5f-af15-20e897afe787 service nova] [instance: 9ea2600a-d439-4fc3-a131-9bb93250d8cd] Refreshing instance network info cache due to event network-changed-a1ecfb50-9098-4408-b9bd-8db498ae622e. {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 525.054160] env[61923]: DEBUG oslo_concurrency.lockutils [req-855bd337-71e8-405d-8445-bff56e2db847 req-0f64ee1b-4409-4f5f-af15-20e897afe787 service nova] Acquiring lock "refresh_cache-9ea2600a-d439-4fc3-a131-9bb93250d8cd" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 525.054231] env[61923]: DEBUG oslo_concurrency.lockutils [req-855bd337-71e8-405d-8445-bff56e2db847 req-0f64ee1b-4409-4f5f-af15-20e897afe787 service nova] Acquired lock "refresh_cache-9ea2600a-d439-4fc3-a131-9bb93250d8cd" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 525.054348] env[61923]: DEBUG nova.network.neutron [req-855bd337-71e8-405d-8445-bff56e2db847 req-0f64ee1b-4409-4f5f-af15-20e897afe787 service nova] [instance: 9ea2600a-d439-4fc3-a131-9bb93250d8cd] Refreshing network info cache for port a1ecfb50-9098-4408-b9bd-8db498ae622e {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 525.254599] env[61923]: DEBUG nova.network.neutron [-] [instance: 9ea2600a-d439-4fc3-a131-9bb93250d8cd] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 525.536017] env[61923]: DEBUG nova.compute.utils [None req-8a6b6bc2-54c9-4389-ab30-057158d86227 tempest-ServerDiagnosticsNegativeTest-664316905 tempest-ServerDiagnosticsNegativeTest-664316905-project-member] Using /dev/sd instead of None {{(pid=61923) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 525.538216] env[61923]: DEBUG nova.compute.manager [None req-8a6b6bc2-54c9-4389-ab30-057158d86227 tempest-ServerDiagnosticsNegativeTest-664316905 tempest-ServerDiagnosticsNegativeTest-664316905-project-member] [instance: b0a45314-eed8-4d0a-bea2-d4b47dc0bc80] Allocating IP information in the background. {{(pid=61923) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 525.538216] env[61923]: DEBUG nova.network.neutron [None req-8a6b6bc2-54c9-4389-ab30-057158d86227 tempest-ServerDiagnosticsNegativeTest-664316905 tempest-ServerDiagnosticsNegativeTest-664316905-project-member] [instance: b0a45314-eed8-4d0a-bea2-d4b47dc0bc80] allocate_for_instance() {{(pid=61923) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 525.610457] env[61923]: DEBUG nova.network.neutron [req-855bd337-71e8-405d-8445-bff56e2db847 req-0f64ee1b-4409-4f5f-af15-20e897afe787 service nova] [instance: 9ea2600a-d439-4fc3-a131-9bb93250d8cd] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 525.710088] env[61923]: DEBUG nova.policy [None req-8a6b6bc2-54c9-4389-ab30-057158d86227 tempest-ServerDiagnosticsNegativeTest-664316905 tempest-ServerDiagnosticsNegativeTest-664316905-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '91e0603ddfc04953b150ef6845e6a592', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '597e4f4519f54b1eb065ebf9b820a87b', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61923) authorize /opt/stack/nova/nova/policy.py:201}} [ 525.759995] env[61923]: INFO nova.compute.manager [-] [instance: 9ea2600a-d439-4fc3-a131-9bb93250d8cd] Took 1.05 seconds to deallocate network for instance. [ 525.761506] env[61923]: DEBUG nova.compute.claims [None req-35a5fc5c-7e44-4806-b390-80f5a8a499d8 tempest-DeleteServersAdminTestJSON-590116201 tempest-DeleteServersAdminTestJSON-590116201-project-member] [instance: 9ea2600a-d439-4fc3-a131-9bb93250d8cd] Aborting claim: {{(pid=61923) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 525.762061] env[61923]: DEBUG oslo_concurrency.lockutils [None req-35a5fc5c-7e44-4806-b390-80f5a8a499d8 tempest-DeleteServersAdminTestJSON-590116201 tempest-DeleteServersAdminTestJSON-590116201-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 525.770879] env[61923]: DEBUG nova.network.neutron [req-855bd337-71e8-405d-8445-bff56e2db847 req-0f64ee1b-4409-4f5f-af15-20e897afe787 service nova] [instance: 9ea2600a-d439-4fc3-a131-9bb93250d8cd] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 525.804606] env[61923]: DEBUG oslo_concurrency.lockutils [None req-dcb477a6-85d7-4838-be85-b9b2ecf96877 tempest-ImagesOneServerNegativeTestJSON-1560548781 tempest-ImagesOneServerNegativeTestJSON-1560548781-project-member] Acquiring lock "ca62f959-43b9-4838-8349-11b7d6a7f0a7" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 525.806161] env[61923]: DEBUG oslo_concurrency.lockutils [None req-dcb477a6-85d7-4838-be85-b9b2ecf96877 tempest-ImagesOneServerNegativeTestJSON-1560548781 tempest-ImagesOneServerNegativeTestJSON-1560548781-project-member] Lock "ca62f959-43b9-4838-8349-11b7d6a7f0a7" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 526.041747] env[61923]: DEBUG nova.compute.manager [None req-8a6b6bc2-54c9-4389-ab30-057158d86227 tempest-ServerDiagnosticsNegativeTest-664316905 tempest-ServerDiagnosticsNegativeTest-664316905-project-member] [instance: b0a45314-eed8-4d0a-bea2-d4b47dc0bc80] Start building block device mappings for instance. {{(pid=61923) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 526.277405] env[61923]: DEBUG oslo_concurrency.lockutils [req-855bd337-71e8-405d-8445-bff56e2db847 req-0f64ee1b-4409-4f5f-af15-20e897afe787 service nova] Releasing lock "refresh_cache-9ea2600a-d439-4fc3-a131-9bb93250d8cd" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 526.277680] env[61923]: DEBUG nova.compute.manager [req-855bd337-71e8-405d-8445-bff56e2db847 req-0f64ee1b-4409-4f5f-af15-20e897afe787 service nova] [instance: 9ea2600a-d439-4fc3-a131-9bb93250d8cd] Received event network-vif-deleted-a1ecfb50-9098-4408-b9bd-8db498ae622e {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 526.308567] env[61923]: DEBUG nova.compute.manager [None req-dcb477a6-85d7-4838-be85-b9b2ecf96877 tempest-ImagesOneServerNegativeTestJSON-1560548781 tempest-ImagesOneServerNegativeTestJSON-1560548781-project-member] [instance: ca62f959-43b9-4838-8349-11b7d6a7f0a7] Starting instance... {{(pid=61923) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 526.316164] env[61923]: ERROR nova.compute.manager [None req-464a16f8-1845-4d4d-b3ea-fd8ba00de8bb tempest-ServersAdminNegativeTestJSON-1828705899 tempest-ServersAdminNegativeTestJSON-1828705899-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port c188410e-a247-4617-a118-46ee79a2683c, please check neutron logs for more information. [ 526.316164] env[61923]: ERROR nova.compute.manager Traceback (most recent call last): [ 526.316164] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 526.316164] env[61923]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 526.316164] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 526.316164] env[61923]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 526.316164] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 526.316164] env[61923]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 526.316164] env[61923]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 526.316164] env[61923]: ERROR nova.compute.manager self.force_reraise() [ 526.316164] env[61923]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 526.316164] env[61923]: ERROR nova.compute.manager raise self.value [ 526.316164] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 526.316164] env[61923]: ERROR nova.compute.manager updated_port = self._update_port( [ 526.316164] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 526.316164] env[61923]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 526.316696] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 526.316696] env[61923]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 526.316696] env[61923]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port c188410e-a247-4617-a118-46ee79a2683c, please check neutron logs for more information. [ 526.316696] env[61923]: ERROR nova.compute.manager [ 526.316696] env[61923]: Traceback (most recent call last): [ 526.316696] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 526.316696] env[61923]: listener.cb(fileno) [ 526.316696] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 526.316696] env[61923]: result = function(*args, **kwargs) [ 526.316696] env[61923]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 526.316696] env[61923]: return func(*args, **kwargs) [ 526.316696] env[61923]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 526.316696] env[61923]: raise e [ 526.316696] env[61923]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 526.316696] env[61923]: nwinfo = self.network_api.allocate_for_instance( [ 526.316696] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 526.316696] env[61923]: created_port_ids = self._update_ports_for_instance( [ 526.316696] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 526.316696] env[61923]: with excutils.save_and_reraise_exception(): [ 526.316696] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 526.316696] env[61923]: self.force_reraise() [ 526.316696] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 526.316696] env[61923]: raise self.value [ 526.316696] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 526.316696] env[61923]: updated_port = self._update_port( [ 526.316696] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 526.316696] env[61923]: _ensure_no_port_binding_failure(port) [ 526.316696] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 526.316696] env[61923]: raise exception.PortBindingFailed(port_id=port['id']) [ 526.317462] env[61923]: nova.exception.PortBindingFailed: Binding failed for port c188410e-a247-4617-a118-46ee79a2683c, please check neutron logs for more information. [ 526.317462] env[61923]: Removing descriptor: 14 [ 526.317462] env[61923]: ERROR nova.compute.manager [None req-464a16f8-1845-4d4d-b3ea-fd8ba00de8bb tempest-ServersAdminNegativeTestJSON-1828705899 tempest-ServersAdminNegativeTestJSON-1828705899-project-member] [instance: 1ade1b37-8319-4d74-9544-6b83e8738498] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port c188410e-a247-4617-a118-46ee79a2683c, please check neutron logs for more information. [ 526.317462] env[61923]: ERROR nova.compute.manager [instance: 1ade1b37-8319-4d74-9544-6b83e8738498] Traceback (most recent call last): [ 526.317462] env[61923]: ERROR nova.compute.manager [instance: 1ade1b37-8319-4d74-9544-6b83e8738498] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 526.317462] env[61923]: ERROR nova.compute.manager [instance: 1ade1b37-8319-4d74-9544-6b83e8738498] yield resources [ 526.317462] env[61923]: ERROR nova.compute.manager [instance: 1ade1b37-8319-4d74-9544-6b83e8738498] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 526.317462] env[61923]: ERROR nova.compute.manager [instance: 1ade1b37-8319-4d74-9544-6b83e8738498] self.driver.spawn(context, instance, image_meta, [ 526.317462] env[61923]: ERROR nova.compute.manager [instance: 1ade1b37-8319-4d74-9544-6b83e8738498] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 526.317462] env[61923]: ERROR nova.compute.manager [instance: 1ade1b37-8319-4d74-9544-6b83e8738498] self._vmops.spawn(context, instance, image_meta, injected_files, [ 526.317462] env[61923]: ERROR nova.compute.manager [instance: 1ade1b37-8319-4d74-9544-6b83e8738498] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 526.317462] env[61923]: ERROR nova.compute.manager [instance: 1ade1b37-8319-4d74-9544-6b83e8738498] vm_ref = self.build_virtual_machine(instance, [ 526.317766] env[61923]: ERROR nova.compute.manager [instance: 1ade1b37-8319-4d74-9544-6b83e8738498] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 526.317766] env[61923]: ERROR nova.compute.manager [instance: 1ade1b37-8319-4d74-9544-6b83e8738498] vif_infos = vmwarevif.get_vif_info(self._session, [ 526.317766] env[61923]: ERROR nova.compute.manager [instance: 1ade1b37-8319-4d74-9544-6b83e8738498] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 526.317766] env[61923]: ERROR nova.compute.manager [instance: 1ade1b37-8319-4d74-9544-6b83e8738498] for vif in network_info: [ 526.317766] env[61923]: ERROR nova.compute.manager [instance: 1ade1b37-8319-4d74-9544-6b83e8738498] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 526.317766] env[61923]: ERROR nova.compute.manager [instance: 1ade1b37-8319-4d74-9544-6b83e8738498] return self._sync_wrapper(fn, *args, **kwargs) [ 526.317766] env[61923]: ERROR nova.compute.manager [instance: 1ade1b37-8319-4d74-9544-6b83e8738498] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 526.317766] env[61923]: ERROR nova.compute.manager [instance: 1ade1b37-8319-4d74-9544-6b83e8738498] self.wait() [ 526.317766] env[61923]: ERROR nova.compute.manager [instance: 1ade1b37-8319-4d74-9544-6b83e8738498] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 526.317766] env[61923]: ERROR nova.compute.manager [instance: 1ade1b37-8319-4d74-9544-6b83e8738498] self[:] = self._gt.wait() [ 526.317766] env[61923]: ERROR nova.compute.manager [instance: 1ade1b37-8319-4d74-9544-6b83e8738498] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 526.317766] env[61923]: ERROR nova.compute.manager [instance: 1ade1b37-8319-4d74-9544-6b83e8738498] return self._exit_event.wait() [ 526.317766] env[61923]: ERROR nova.compute.manager [instance: 1ade1b37-8319-4d74-9544-6b83e8738498] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 526.318085] env[61923]: ERROR nova.compute.manager [instance: 1ade1b37-8319-4d74-9544-6b83e8738498] result = hub.switch() [ 526.318085] env[61923]: ERROR nova.compute.manager [instance: 1ade1b37-8319-4d74-9544-6b83e8738498] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 526.318085] env[61923]: ERROR nova.compute.manager [instance: 1ade1b37-8319-4d74-9544-6b83e8738498] return self.greenlet.switch() [ 526.318085] env[61923]: ERROR nova.compute.manager [instance: 1ade1b37-8319-4d74-9544-6b83e8738498] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 526.318085] env[61923]: ERROR nova.compute.manager [instance: 1ade1b37-8319-4d74-9544-6b83e8738498] result = function(*args, **kwargs) [ 526.318085] env[61923]: ERROR nova.compute.manager [instance: 1ade1b37-8319-4d74-9544-6b83e8738498] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 526.318085] env[61923]: ERROR nova.compute.manager [instance: 1ade1b37-8319-4d74-9544-6b83e8738498] return func(*args, **kwargs) [ 526.318085] env[61923]: ERROR nova.compute.manager [instance: 1ade1b37-8319-4d74-9544-6b83e8738498] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 526.318085] env[61923]: ERROR nova.compute.manager [instance: 1ade1b37-8319-4d74-9544-6b83e8738498] raise e [ 526.318085] env[61923]: ERROR nova.compute.manager [instance: 1ade1b37-8319-4d74-9544-6b83e8738498] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 526.318085] env[61923]: ERROR nova.compute.manager [instance: 1ade1b37-8319-4d74-9544-6b83e8738498] nwinfo = self.network_api.allocate_for_instance( [ 526.318085] env[61923]: ERROR nova.compute.manager [instance: 1ade1b37-8319-4d74-9544-6b83e8738498] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 526.318085] env[61923]: ERROR nova.compute.manager [instance: 1ade1b37-8319-4d74-9544-6b83e8738498] created_port_ids = self._update_ports_for_instance( [ 526.318433] env[61923]: ERROR nova.compute.manager [instance: 1ade1b37-8319-4d74-9544-6b83e8738498] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 526.318433] env[61923]: ERROR nova.compute.manager [instance: 1ade1b37-8319-4d74-9544-6b83e8738498] with excutils.save_and_reraise_exception(): [ 526.318433] env[61923]: ERROR nova.compute.manager [instance: 1ade1b37-8319-4d74-9544-6b83e8738498] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 526.318433] env[61923]: ERROR nova.compute.manager [instance: 1ade1b37-8319-4d74-9544-6b83e8738498] self.force_reraise() [ 526.318433] env[61923]: ERROR nova.compute.manager [instance: 1ade1b37-8319-4d74-9544-6b83e8738498] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 526.318433] env[61923]: ERROR nova.compute.manager [instance: 1ade1b37-8319-4d74-9544-6b83e8738498] raise self.value [ 526.318433] env[61923]: ERROR nova.compute.manager [instance: 1ade1b37-8319-4d74-9544-6b83e8738498] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 526.318433] env[61923]: ERROR nova.compute.manager [instance: 1ade1b37-8319-4d74-9544-6b83e8738498] updated_port = self._update_port( [ 526.318433] env[61923]: ERROR nova.compute.manager [instance: 1ade1b37-8319-4d74-9544-6b83e8738498] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 526.318433] env[61923]: ERROR nova.compute.manager [instance: 1ade1b37-8319-4d74-9544-6b83e8738498] _ensure_no_port_binding_failure(port) [ 526.318433] env[61923]: ERROR nova.compute.manager [instance: 1ade1b37-8319-4d74-9544-6b83e8738498] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 526.318433] env[61923]: ERROR nova.compute.manager [instance: 1ade1b37-8319-4d74-9544-6b83e8738498] raise exception.PortBindingFailed(port_id=port['id']) [ 526.318801] env[61923]: ERROR nova.compute.manager [instance: 1ade1b37-8319-4d74-9544-6b83e8738498] nova.exception.PortBindingFailed: Binding failed for port c188410e-a247-4617-a118-46ee79a2683c, please check neutron logs for more information. [ 526.318801] env[61923]: ERROR nova.compute.manager [instance: 1ade1b37-8319-4d74-9544-6b83e8738498] [ 526.318801] env[61923]: INFO nova.compute.manager [None req-464a16f8-1845-4d4d-b3ea-fd8ba00de8bb tempest-ServersAdminNegativeTestJSON-1828705899 tempest-ServersAdminNegativeTestJSON-1828705899-project-member] [instance: 1ade1b37-8319-4d74-9544-6b83e8738498] Terminating instance [ 526.330514] env[61923]: DEBUG oslo_concurrency.lockutils [None req-464a16f8-1845-4d4d-b3ea-fd8ba00de8bb tempest-ServersAdminNegativeTestJSON-1828705899 tempest-ServersAdminNegativeTestJSON-1828705899-project-member] Acquiring lock "refresh_cache-1ade1b37-8319-4d74-9544-6b83e8738498" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 526.330514] env[61923]: DEBUG oslo_concurrency.lockutils [None req-464a16f8-1845-4d4d-b3ea-fd8ba00de8bb tempest-ServersAdminNegativeTestJSON-1828705899 tempest-ServersAdminNegativeTestJSON-1828705899-project-member] Acquired lock "refresh_cache-1ade1b37-8319-4d74-9544-6b83e8738498" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 526.330514] env[61923]: DEBUG nova.network.neutron [None req-464a16f8-1845-4d4d-b3ea-fd8ba00de8bb tempest-ServersAdminNegativeTestJSON-1828705899 tempest-ServersAdminNegativeTestJSON-1828705899-project-member] [instance: 1ade1b37-8319-4d74-9544-6b83e8738498] Building network info cache for instance {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 526.343873] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7549c75a-46a6-48d9-a634-e74273a59d24 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 526.354693] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e270b1f-8c83-45a9-a3d7-7b5ad28bdb69 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 526.365265] env[61923]: DEBUG oslo_service.periodic_task [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61923) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 526.394793] env[61923]: DEBUG oslo_service.periodic_task [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61923) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 526.394793] env[61923]: DEBUG nova.compute.manager [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Starting heal instance info cache {{(pid=61923) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 526.394793] env[61923]: DEBUG nova.compute.manager [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Rebuilding the list of instances to heal {{(pid=61923) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 526.400102] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea2f353b-3599-49cb-bf76-46ba8983a203 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 526.407910] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4131608-716f-44d9-ae1b-3aa20ba338e0 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 526.423597] env[61923]: DEBUG nova.compute.provider_tree [None req-7781f812-6fae-4bd6-9d91-ec6098c06372 tempest-InstanceActionsTestJSON-2015629344 tempest-InstanceActionsTestJSON-2015629344-project-member] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 526.837511] env[61923]: DEBUG oslo_concurrency.lockutils [None req-dcb477a6-85d7-4838-be85-b9b2ecf96877 tempest-ImagesOneServerNegativeTestJSON-1560548781 tempest-ImagesOneServerNegativeTestJSON-1560548781-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 526.866160] env[61923]: DEBUG nova.network.neutron [None req-464a16f8-1845-4d4d-b3ea-fd8ba00de8bb tempest-ServersAdminNegativeTestJSON-1828705899 tempest-ServersAdminNegativeTestJSON-1828705899-project-member] [instance: 1ade1b37-8319-4d74-9544-6b83e8738498] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 526.906724] env[61923]: DEBUG nova.compute.manager [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] [instance: 242f5cab-3c0c-4f8d-82a8-0cae8d3d419e] Skipping network cache update for instance because it is Building. {{(pid=61923) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 526.906724] env[61923]: DEBUG nova.compute.manager [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] [instance: a134ab6d-b261-4df6-b440-d32a3e578645] Skipping network cache update for instance because it is Building. {{(pid=61923) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 526.906724] env[61923]: DEBUG nova.compute.manager [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] [instance: 9ea2600a-d439-4fc3-a131-9bb93250d8cd] Skipping network cache update for instance because it is Building. {{(pid=61923) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 526.906724] env[61923]: DEBUG nova.compute.manager [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] [instance: 1ade1b37-8319-4d74-9544-6b83e8738498] Skipping network cache update for instance because it is Building. {{(pid=61923) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 526.909203] env[61923]: DEBUG nova.compute.manager [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] [instance: b0a45314-eed8-4d0a-bea2-d4b47dc0bc80] Skipping network cache update for instance because it is Building. {{(pid=61923) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 526.909203] env[61923]: DEBUG nova.compute.manager [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] [instance: 97148ea5-398a-4974-b8f6-62290e519794] Skipping network cache update for instance because it is Building. {{(pid=61923) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 526.909203] env[61923]: DEBUG nova.compute.manager [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Didn't find any instances for network info cache update. {{(pid=61923) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10014}} [ 526.910463] env[61923]: DEBUG oslo_service.periodic_task [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61923) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 526.910463] env[61923]: DEBUG oslo_service.periodic_task [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61923) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 526.910463] env[61923]: DEBUG oslo_service.periodic_task [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61923) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 526.910463] env[61923]: DEBUG oslo_service.periodic_task [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61923) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 526.910852] env[61923]: DEBUG oslo_service.periodic_task [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61923) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 526.910852] env[61923]: DEBUG oslo_service.periodic_task [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61923) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 526.910969] env[61923]: DEBUG nova.compute.manager [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61923) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 526.911055] env[61923]: DEBUG oslo_service.periodic_task [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Running periodic task ComputeManager.update_available_resource {{(pid=61923) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 526.927406] env[61923]: DEBUG nova.scheduler.client.report [None req-7781f812-6fae-4bd6-9d91-ec6098c06372 tempest-InstanceActionsTestJSON-2015629344 tempest-InstanceActionsTestJSON-2015629344-project-member] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 527.030526] env[61923]: DEBUG nova.network.neutron [None req-464a16f8-1845-4d4d-b3ea-fd8ba00de8bb tempest-ServersAdminNegativeTestJSON-1828705899 tempest-ServersAdminNegativeTestJSON-1828705899-project-member] [instance: 1ade1b37-8319-4d74-9544-6b83e8738498] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 527.064876] env[61923]: DEBUG nova.compute.manager [None req-8a6b6bc2-54c9-4389-ab30-057158d86227 tempest-ServerDiagnosticsNegativeTest-664316905 tempest-ServerDiagnosticsNegativeTest-664316905-project-member] [instance: b0a45314-eed8-4d0a-bea2-d4b47dc0bc80] Start spawning the instance on the hypervisor. {{(pid=61923) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 527.094857] env[61923]: DEBUG nova.virt.hardware [None req-8a6b6bc2-54c9-4389-ab30-057158d86227 tempest-ServerDiagnosticsNegativeTest-664316905 tempest-ServerDiagnosticsNegativeTest-664316905-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-29T20:07:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-29T20:07:35Z,direct_url=,disk_format='vmdk',id=0f153f63-ae0a-45b1-b7f5-7f9b673c947f,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='4e7b5e3b3c3443c8bd5c70f8a15739f5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-29T20:07:36Z,virtual_size=,visibility=), allow threads: False {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 527.094857] env[61923]: DEBUG nova.virt.hardware [None req-8a6b6bc2-54c9-4389-ab30-057158d86227 tempest-ServerDiagnosticsNegativeTest-664316905 tempest-ServerDiagnosticsNegativeTest-664316905-project-member] Flavor limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 527.094857] env[61923]: DEBUG nova.virt.hardware [None req-8a6b6bc2-54c9-4389-ab30-057158d86227 tempest-ServerDiagnosticsNegativeTest-664316905 tempest-ServerDiagnosticsNegativeTest-664316905-project-member] Image limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 527.095074] env[61923]: DEBUG nova.virt.hardware [None req-8a6b6bc2-54c9-4389-ab30-057158d86227 tempest-ServerDiagnosticsNegativeTest-664316905 tempest-ServerDiagnosticsNegativeTest-664316905-project-member] Flavor pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 527.095074] env[61923]: DEBUG nova.virt.hardware [None req-8a6b6bc2-54c9-4389-ab30-057158d86227 tempest-ServerDiagnosticsNegativeTest-664316905 tempest-ServerDiagnosticsNegativeTest-664316905-project-member] Image pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 527.095074] env[61923]: DEBUG nova.virt.hardware [None req-8a6b6bc2-54c9-4389-ab30-057158d86227 tempest-ServerDiagnosticsNegativeTest-664316905 tempest-ServerDiagnosticsNegativeTest-664316905-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 527.095074] env[61923]: DEBUG nova.virt.hardware [None req-8a6b6bc2-54c9-4389-ab30-057158d86227 tempest-ServerDiagnosticsNegativeTest-664316905 tempest-ServerDiagnosticsNegativeTest-664316905-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 527.095074] env[61923]: DEBUG nova.virt.hardware [None req-8a6b6bc2-54c9-4389-ab30-057158d86227 tempest-ServerDiagnosticsNegativeTest-664316905 tempest-ServerDiagnosticsNegativeTest-664316905-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 527.095244] env[61923]: DEBUG nova.virt.hardware [None req-8a6b6bc2-54c9-4389-ab30-057158d86227 tempest-ServerDiagnosticsNegativeTest-664316905 tempest-ServerDiagnosticsNegativeTest-664316905-project-member] Got 1 possible topologies {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 527.095244] env[61923]: DEBUG nova.virt.hardware [None req-8a6b6bc2-54c9-4389-ab30-057158d86227 tempest-ServerDiagnosticsNegativeTest-664316905 tempest-ServerDiagnosticsNegativeTest-664316905-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 527.095244] env[61923]: DEBUG nova.virt.hardware [None req-8a6b6bc2-54c9-4389-ab30-057158d86227 tempest-ServerDiagnosticsNegativeTest-664316905 tempest-ServerDiagnosticsNegativeTest-664316905-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 527.096492] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ccc1c62a-4e2d-4543-9f07-7cd302c39a3e {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 527.107827] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f71ea1d2-d037-4557-8f2e-4d751292dcc8 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 527.355831] env[61923]: DEBUG nova.network.neutron [None req-8a6b6bc2-54c9-4389-ab30-057158d86227 tempest-ServerDiagnosticsNegativeTest-664316905 tempest-ServerDiagnosticsNegativeTest-664316905-project-member] [instance: b0a45314-eed8-4d0a-bea2-d4b47dc0bc80] Successfully created port: 6bf512ed-8713-43d4-bc7e-ce2b9a514f52 {{(pid=61923) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 527.379035] env[61923]: DEBUG nova.compute.manager [req-7ecc5c8b-658a-47b8-8f7f-85d8ef7b3872 req-0fa4bb21-7f34-4d3c-a5bd-5a4a4c0d8c8b service nova] [instance: 1ade1b37-8319-4d74-9544-6b83e8738498] Received event network-changed-c188410e-a247-4617-a118-46ee79a2683c {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 527.379217] env[61923]: DEBUG nova.compute.manager [req-7ecc5c8b-658a-47b8-8f7f-85d8ef7b3872 req-0fa4bb21-7f34-4d3c-a5bd-5a4a4c0d8c8b service nova] [instance: 1ade1b37-8319-4d74-9544-6b83e8738498] Refreshing instance network info cache due to event network-changed-c188410e-a247-4617-a118-46ee79a2683c. {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 527.379422] env[61923]: DEBUG oslo_concurrency.lockutils [req-7ecc5c8b-658a-47b8-8f7f-85d8ef7b3872 req-0fa4bb21-7f34-4d3c-a5bd-5a4a4c0d8c8b service nova] Acquiring lock "refresh_cache-1ade1b37-8319-4d74-9544-6b83e8738498" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 527.416326] env[61923]: DEBUG oslo_concurrency.lockutils [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 527.434466] env[61923]: DEBUG oslo_concurrency.lockutils [None req-7781f812-6fae-4bd6-9d91-ec6098c06372 tempest-InstanceActionsTestJSON-2015629344 tempest-InstanceActionsTestJSON-2015629344-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.403s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 527.434863] env[61923]: DEBUG nova.compute.manager [None req-7781f812-6fae-4bd6-9d91-ec6098c06372 tempest-InstanceActionsTestJSON-2015629344 tempest-InstanceActionsTestJSON-2015629344-project-member] [instance: 97148ea5-398a-4974-b8f6-62290e519794] Start building networks asynchronously for instance. {{(pid=61923) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 527.442702] env[61923]: DEBUG oslo_concurrency.lockutils [None req-dc345cdf-ff58-4514-9baa-d2d8d13601af tempest-MigrationsAdminTest-1311988845 tempest-MigrationsAdminTest-1311988845-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 5.793s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 527.534567] env[61923]: DEBUG oslo_concurrency.lockutils [None req-464a16f8-1845-4d4d-b3ea-fd8ba00de8bb tempest-ServersAdminNegativeTestJSON-1828705899 tempest-ServersAdminNegativeTestJSON-1828705899-project-member] Releasing lock "refresh_cache-1ade1b37-8319-4d74-9544-6b83e8738498" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 527.536629] env[61923]: DEBUG nova.compute.manager [None req-464a16f8-1845-4d4d-b3ea-fd8ba00de8bb tempest-ServersAdminNegativeTestJSON-1828705899 tempest-ServersAdminNegativeTestJSON-1828705899-project-member] [instance: 1ade1b37-8319-4d74-9544-6b83e8738498] Start destroying the instance on the hypervisor. {{(pid=61923) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 527.536629] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-464a16f8-1845-4d4d-b3ea-fd8ba00de8bb tempest-ServersAdminNegativeTestJSON-1828705899 tempest-ServersAdminNegativeTestJSON-1828705899-project-member] [instance: 1ade1b37-8319-4d74-9544-6b83e8738498] Destroying instance {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 527.536629] env[61923]: DEBUG oslo_concurrency.lockutils [req-7ecc5c8b-658a-47b8-8f7f-85d8ef7b3872 req-0fa4bb21-7f34-4d3c-a5bd-5a4a4c0d8c8b service nova] Acquired lock "refresh_cache-1ade1b37-8319-4d74-9544-6b83e8738498" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 527.536629] env[61923]: DEBUG nova.network.neutron [req-7ecc5c8b-658a-47b8-8f7f-85d8ef7b3872 req-0fa4bb21-7f34-4d3c-a5bd-5a4a4c0d8c8b service nova] [instance: 1ade1b37-8319-4d74-9544-6b83e8738498] Refreshing network info cache for port c188410e-a247-4617-a118-46ee79a2683c {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 527.537532] env[61923]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ae1cbb7c-3e0f-4a55-9385-e48f1526409f {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 527.551525] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91ea15a0-2c72-4539-9985-eb251a39715b {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 527.588229] env[61923]: WARNING nova.virt.vmwareapi.vmops [None req-464a16f8-1845-4d4d-b3ea-fd8ba00de8bb tempest-ServersAdminNegativeTestJSON-1828705899 tempest-ServersAdminNegativeTestJSON-1828705899-project-member] [instance: 1ade1b37-8319-4d74-9544-6b83e8738498] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 1ade1b37-8319-4d74-9544-6b83e8738498 could not be found. [ 527.588229] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-464a16f8-1845-4d4d-b3ea-fd8ba00de8bb tempest-ServersAdminNegativeTestJSON-1828705899 tempest-ServersAdminNegativeTestJSON-1828705899-project-member] [instance: 1ade1b37-8319-4d74-9544-6b83e8738498] Instance destroyed {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 527.588229] env[61923]: INFO nova.compute.manager [None req-464a16f8-1845-4d4d-b3ea-fd8ba00de8bb tempest-ServersAdminNegativeTestJSON-1828705899 tempest-ServersAdminNegativeTestJSON-1828705899-project-member] [instance: 1ade1b37-8319-4d74-9544-6b83e8738498] Took 0.05 seconds to destroy the instance on the hypervisor. [ 527.588994] env[61923]: DEBUG oslo.service.loopingcall [None req-464a16f8-1845-4d4d-b3ea-fd8ba00de8bb tempest-ServersAdminNegativeTestJSON-1828705899 tempest-ServersAdminNegativeTestJSON-1828705899-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61923) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 527.588994] env[61923]: DEBUG nova.compute.manager [-] [instance: 1ade1b37-8319-4d74-9544-6b83e8738498] Deallocating network for instance {{(pid=61923) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 527.588994] env[61923]: DEBUG nova.network.neutron [-] [instance: 1ade1b37-8319-4d74-9544-6b83e8738498] deallocate_for_instance() {{(pid=61923) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 527.628660] env[61923]: DEBUG nova.network.neutron [-] [instance: 1ade1b37-8319-4d74-9544-6b83e8738498] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 527.948371] env[61923]: DEBUG nova.compute.utils [None req-7781f812-6fae-4bd6-9d91-ec6098c06372 tempest-InstanceActionsTestJSON-2015629344 tempest-InstanceActionsTestJSON-2015629344-project-member] Using /dev/sd instead of None {{(pid=61923) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 527.954446] env[61923]: DEBUG nova.compute.manager [None req-7781f812-6fae-4bd6-9d91-ec6098c06372 tempest-InstanceActionsTestJSON-2015629344 tempest-InstanceActionsTestJSON-2015629344-project-member] [instance: 97148ea5-398a-4974-b8f6-62290e519794] Allocating IP information in the background. {{(pid=61923) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 527.956060] env[61923]: DEBUG nova.network.neutron [None req-7781f812-6fae-4bd6-9d91-ec6098c06372 tempest-InstanceActionsTestJSON-2015629344 tempest-InstanceActionsTestJSON-2015629344-project-member] [instance: 97148ea5-398a-4974-b8f6-62290e519794] allocate_for_instance() {{(pid=61923) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 528.099677] env[61923]: DEBUG nova.network.neutron [req-7ecc5c8b-658a-47b8-8f7f-85d8ef7b3872 req-0fa4bb21-7f34-4d3c-a5bd-5a4a4c0d8c8b service nova] [instance: 1ade1b37-8319-4d74-9544-6b83e8738498] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 528.131707] env[61923]: DEBUG nova.network.neutron [-] [instance: 1ade1b37-8319-4d74-9544-6b83e8738498] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 528.139980] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39e8f001-69b3-4518-9ffc-6bb316b25a09 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 528.152180] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78a27266-6753-4f07-91f0-dbc89f4b62a7 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 528.159880] env[61923]: DEBUG nova.policy [None req-7781f812-6fae-4bd6-9d91-ec6098c06372 tempest-InstanceActionsTestJSON-2015629344 tempest-InstanceActionsTestJSON-2015629344-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '928aee5023774c4ca74cd9d6d661c01c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1fb06cc31ab14532a79c4c7fd467d4fd', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61923) authorize /opt/stack/nova/nova/policy.py:201}} [ 528.219141] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b6a9c43-4561-45d0-9094-b97a896905e9 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 528.227140] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e225e63-3866-43b0-bcbf-be5746bd7549 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 528.242167] env[61923]: DEBUG nova.compute.provider_tree [None req-dc345cdf-ff58-4514-9baa-d2d8d13601af tempest-MigrationsAdminTest-1311988845 tempest-MigrationsAdminTest-1311988845-project-member] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 528.456784] env[61923]: DEBUG nova.compute.manager [None req-7781f812-6fae-4bd6-9d91-ec6098c06372 tempest-InstanceActionsTestJSON-2015629344 tempest-InstanceActionsTestJSON-2015629344-project-member] [instance: 97148ea5-398a-4974-b8f6-62290e519794] Start building block device mappings for instance. {{(pid=61923) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 528.472534] env[61923]: DEBUG nova.network.neutron [req-7ecc5c8b-658a-47b8-8f7f-85d8ef7b3872 req-0fa4bb21-7f34-4d3c-a5bd-5a4a4c0d8c8b service nova] [instance: 1ade1b37-8319-4d74-9544-6b83e8738498] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 528.635420] env[61923]: INFO nova.compute.manager [-] [instance: 1ade1b37-8319-4d74-9544-6b83e8738498] Took 1.05 seconds to deallocate network for instance. [ 528.644222] env[61923]: DEBUG nova.compute.claims [None req-464a16f8-1845-4d4d-b3ea-fd8ba00de8bb tempest-ServersAdminNegativeTestJSON-1828705899 tempest-ServersAdminNegativeTestJSON-1828705899-project-member] [instance: 1ade1b37-8319-4d74-9544-6b83e8738498] Aborting claim: {{(pid=61923) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 528.644222] env[61923]: DEBUG oslo_concurrency.lockutils [None req-464a16f8-1845-4d4d-b3ea-fd8ba00de8bb tempest-ServersAdminNegativeTestJSON-1828705899 tempest-ServersAdminNegativeTestJSON-1828705899-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 528.746600] env[61923]: DEBUG nova.scheduler.client.report [None req-dc345cdf-ff58-4514-9baa-d2d8d13601af tempest-MigrationsAdminTest-1311988845 tempest-MigrationsAdminTest-1311988845-project-member] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 528.975660] env[61923]: DEBUG oslo_concurrency.lockutils [req-7ecc5c8b-658a-47b8-8f7f-85d8ef7b3872 req-0fa4bb21-7f34-4d3c-a5bd-5a4a4c0d8c8b service nova] Releasing lock "refresh_cache-1ade1b37-8319-4d74-9544-6b83e8738498" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 528.975660] env[61923]: DEBUG nova.compute.manager [req-7ecc5c8b-658a-47b8-8f7f-85d8ef7b3872 req-0fa4bb21-7f34-4d3c-a5bd-5a4a4c0d8c8b service nova] [instance: 1ade1b37-8319-4d74-9544-6b83e8738498] Received event network-vif-deleted-c188410e-a247-4617-a118-46ee79a2683c {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 529.251980] env[61923]: DEBUG oslo_concurrency.lockutils [None req-dc345cdf-ff58-4514-9baa-d2d8d13601af tempest-MigrationsAdminTest-1311988845 tempest-MigrationsAdminTest-1311988845-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.809s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 529.252650] env[61923]: ERROR nova.compute.manager [None req-dc345cdf-ff58-4514-9baa-d2d8d13601af tempest-MigrationsAdminTest-1311988845 tempest-MigrationsAdminTest-1311988845-project-member] [instance: 242f5cab-3c0c-4f8d-82a8-0cae8d3d419e] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 88c39c48-cf39-4b87-abda-5dbe55ea29c8, please check neutron logs for more information. [ 529.252650] env[61923]: ERROR nova.compute.manager [instance: 242f5cab-3c0c-4f8d-82a8-0cae8d3d419e] Traceback (most recent call last): [ 529.252650] env[61923]: ERROR nova.compute.manager [instance: 242f5cab-3c0c-4f8d-82a8-0cae8d3d419e] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 529.252650] env[61923]: ERROR nova.compute.manager [instance: 242f5cab-3c0c-4f8d-82a8-0cae8d3d419e] self.driver.spawn(context, instance, image_meta, [ 529.252650] env[61923]: ERROR nova.compute.manager [instance: 242f5cab-3c0c-4f8d-82a8-0cae8d3d419e] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 529.252650] env[61923]: ERROR nova.compute.manager [instance: 242f5cab-3c0c-4f8d-82a8-0cae8d3d419e] self._vmops.spawn(context, instance, image_meta, injected_files, [ 529.252650] env[61923]: ERROR nova.compute.manager [instance: 242f5cab-3c0c-4f8d-82a8-0cae8d3d419e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 529.252650] env[61923]: ERROR nova.compute.manager [instance: 242f5cab-3c0c-4f8d-82a8-0cae8d3d419e] vm_ref = self.build_virtual_machine(instance, [ 529.252650] env[61923]: ERROR nova.compute.manager [instance: 242f5cab-3c0c-4f8d-82a8-0cae8d3d419e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 529.252650] env[61923]: ERROR nova.compute.manager [instance: 242f5cab-3c0c-4f8d-82a8-0cae8d3d419e] vif_infos = vmwarevif.get_vif_info(self._session, [ 529.252650] env[61923]: ERROR nova.compute.manager [instance: 242f5cab-3c0c-4f8d-82a8-0cae8d3d419e] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 529.253217] env[61923]: ERROR nova.compute.manager [instance: 242f5cab-3c0c-4f8d-82a8-0cae8d3d419e] for vif in network_info: [ 529.253217] env[61923]: ERROR nova.compute.manager [instance: 242f5cab-3c0c-4f8d-82a8-0cae8d3d419e] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 529.253217] env[61923]: ERROR nova.compute.manager [instance: 242f5cab-3c0c-4f8d-82a8-0cae8d3d419e] return self._sync_wrapper(fn, *args, **kwargs) [ 529.253217] env[61923]: ERROR nova.compute.manager [instance: 242f5cab-3c0c-4f8d-82a8-0cae8d3d419e] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 529.253217] env[61923]: ERROR nova.compute.manager [instance: 242f5cab-3c0c-4f8d-82a8-0cae8d3d419e] self.wait() [ 529.253217] env[61923]: ERROR nova.compute.manager [instance: 242f5cab-3c0c-4f8d-82a8-0cae8d3d419e] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 529.253217] env[61923]: ERROR nova.compute.manager [instance: 242f5cab-3c0c-4f8d-82a8-0cae8d3d419e] self[:] = self._gt.wait() [ 529.253217] env[61923]: ERROR nova.compute.manager [instance: 242f5cab-3c0c-4f8d-82a8-0cae8d3d419e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 529.253217] env[61923]: ERROR nova.compute.manager [instance: 242f5cab-3c0c-4f8d-82a8-0cae8d3d419e] return self._exit_event.wait() [ 529.253217] env[61923]: ERROR nova.compute.manager [instance: 242f5cab-3c0c-4f8d-82a8-0cae8d3d419e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 529.253217] env[61923]: ERROR nova.compute.manager [instance: 242f5cab-3c0c-4f8d-82a8-0cae8d3d419e] result = hub.switch() [ 529.253217] env[61923]: ERROR nova.compute.manager [instance: 242f5cab-3c0c-4f8d-82a8-0cae8d3d419e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 529.253217] env[61923]: ERROR nova.compute.manager [instance: 242f5cab-3c0c-4f8d-82a8-0cae8d3d419e] return self.greenlet.switch() [ 529.253532] env[61923]: ERROR nova.compute.manager [instance: 242f5cab-3c0c-4f8d-82a8-0cae8d3d419e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 529.253532] env[61923]: ERROR nova.compute.manager [instance: 242f5cab-3c0c-4f8d-82a8-0cae8d3d419e] result = function(*args, **kwargs) [ 529.253532] env[61923]: ERROR nova.compute.manager [instance: 242f5cab-3c0c-4f8d-82a8-0cae8d3d419e] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 529.253532] env[61923]: ERROR nova.compute.manager [instance: 242f5cab-3c0c-4f8d-82a8-0cae8d3d419e] return func(*args, **kwargs) [ 529.253532] env[61923]: ERROR nova.compute.manager [instance: 242f5cab-3c0c-4f8d-82a8-0cae8d3d419e] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 529.253532] env[61923]: ERROR nova.compute.manager [instance: 242f5cab-3c0c-4f8d-82a8-0cae8d3d419e] raise e [ 529.253532] env[61923]: ERROR nova.compute.manager [instance: 242f5cab-3c0c-4f8d-82a8-0cae8d3d419e] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 529.253532] env[61923]: ERROR nova.compute.manager [instance: 242f5cab-3c0c-4f8d-82a8-0cae8d3d419e] nwinfo = self.network_api.allocate_for_instance( [ 529.253532] env[61923]: ERROR nova.compute.manager [instance: 242f5cab-3c0c-4f8d-82a8-0cae8d3d419e] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 529.253532] env[61923]: ERROR nova.compute.manager [instance: 242f5cab-3c0c-4f8d-82a8-0cae8d3d419e] created_port_ids = self._update_ports_for_instance( [ 529.253532] env[61923]: ERROR nova.compute.manager [instance: 242f5cab-3c0c-4f8d-82a8-0cae8d3d419e] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 529.253532] env[61923]: ERROR nova.compute.manager [instance: 242f5cab-3c0c-4f8d-82a8-0cae8d3d419e] with excutils.save_and_reraise_exception(): [ 529.253532] env[61923]: ERROR nova.compute.manager [instance: 242f5cab-3c0c-4f8d-82a8-0cae8d3d419e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 529.254400] env[61923]: ERROR nova.compute.manager [instance: 242f5cab-3c0c-4f8d-82a8-0cae8d3d419e] self.force_reraise() [ 529.254400] env[61923]: ERROR nova.compute.manager [instance: 242f5cab-3c0c-4f8d-82a8-0cae8d3d419e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 529.254400] env[61923]: ERROR nova.compute.manager [instance: 242f5cab-3c0c-4f8d-82a8-0cae8d3d419e] raise self.value [ 529.254400] env[61923]: ERROR nova.compute.manager [instance: 242f5cab-3c0c-4f8d-82a8-0cae8d3d419e] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 529.254400] env[61923]: ERROR nova.compute.manager [instance: 242f5cab-3c0c-4f8d-82a8-0cae8d3d419e] updated_port = self._update_port( [ 529.254400] env[61923]: ERROR nova.compute.manager [instance: 242f5cab-3c0c-4f8d-82a8-0cae8d3d419e] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 529.254400] env[61923]: ERROR nova.compute.manager [instance: 242f5cab-3c0c-4f8d-82a8-0cae8d3d419e] _ensure_no_port_binding_failure(port) [ 529.254400] env[61923]: ERROR nova.compute.manager [instance: 242f5cab-3c0c-4f8d-82a8-0cae8d3d419e] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 529.254400] env[61923]: ERROR nova.compute.manager [instance: 242f5cab-3c0c-4f8d-82a8-0cae8d3d419e] raise exception.PortBindingFailed(port_id=port['id']) [ 529.254400] env[61923]: ERROR nova.compute.manager [instance: 242f5cab-3c0c-4f8d-82a8-0cae8d3d419e] nova.exception.PortBindingFailed: Binding failed for port 88c39c48-cf39-4b87-abda-5dbe55ea29c8, please check neutron logs for more information. [ 529.254400] env[61923]: ERROR nova.compute.manager [instance: 242f5cab-3c0c-4f8d-82a8-0cae8d3d419e] [ 529.254822] env[61923]: DEBUG nova.compute.utils [None req-dc345cdf-ff58-4514-9baa-d2d8d13601af tempest-MigrationsAdminTest-1311988845 tempest-MigrationsAdminTest-1311988845-project-member] [instance: 242f5cab-3c0c-4f8d-82a8-0cae8d3d419e] Binding failed for port 88c39c48-cf39-4b87-abda-5dbe55ea29c8, please check neutron logs for more information. {{(pid=61923) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 529.254822] env[61923]: DEBUG oslo_concurrency.lockutils [None req-844a2215-1038-4a30-8a79-40c1c3117bde tempest-FloatingIPsAssociationNegativeTestJSON-1301210771 tempest-FloatingIPsAssociationNegativeTestJSON-1301210771-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 7.267s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 529.256288] env[61923]: INFO nova.compute.claims [None req-844a2215-1038-4a30-8a79-40c1c3117bde tempest-FloatingIPsAssociationNegativeTestJSON-1301210771 tempest-FloatingIPsAssociationNegativeTestJSON-1301210771-project-member] [instance: 5d57303a-ae17-4b92-8f12-303677eb0d7e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 529.266805] env[61923]: DEBUG nova.compute.manager [None req-dc345cdf-ff58-4514-9baa-d2d8d13601af tempest-MigrationsAdminTest-1311988845 tempest-MigrationsAdminTest-1311988845-project-member] [instance: 242f5cab-3c0c-4f8d-82a8-0cae8d3d419e] Build of instance 242f5cab-3c0c-4f8d-82a8-0cae8d3d419e was re-scheduled: Binding failed for port 88c39c48-cf39-4b87-abda-5dbe55ea29c8, please check neutron logs for more information. {{(pid=61923) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 529.267328] env[61923]: DEBUG nova.compute.manager [None req-dc345cdf-ff58-4514-9baa-d2d8d13601af tempest-MigrationsAdminTest-1311988845 tempest-MigrationsAdminTest-1311988845-project-member] [instance: 242f5cab-3c0c-4f8d-82a8-0cae8d3d419e] Unplugging VIFs for instance {{(pid=61923) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 529.267570] env[61923]: DEBUG oslo_concurrency.lockutils [None req-dc345cdf-ff58-4514-9baa-d2d8d13601af tempest-MigrationsAdminTest-1311988845 tempest-MigrationsAdminTest-1311988845-project-member] Acquiring lock "refresh_cache-242f5cab-3c0c-4f8d-82a8-0cae8d3d419e" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 529.267742] env[61923]: DEBUG oslo_concurrency.lockutils [None req-dc345cdf-ff58-4514-9baa-d2d8d13601af tempest-MigrationsAdminTest-1311988845 tempest-MigrationsAdminTest-1311988845-project-member] Acquired lock "refresh_cache-242f5cab-3c0c-4f8d-82a8-0cae8d3d419e" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 529.267875] env[61923]: DEBUG nova.network.neutron [None req-dc345cdf-ff58-4514-9baa-d2d8d13601af tempest-MigrationsAdminTest-1311988845 tempest-MigrationsAdminTest-1311988845-project-member] [instance: 242f5cab-3c0c-4f8d-82a8-0cae8d3d419e] Building network info cache for instance {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 529.476280] env[61923]: DEBUG nova.compute.manager [None req-7781f812-6fae-4bd6-9d91-ec6098c06372 tempest-InstanceActionsTestJSON-2015629344 tempest-InstanceActionsTestJSON-2015629344-project-member] [instance: 97148ea5-398a-4974-b8f6-62290e519794] Start spawning the instance on the hypervisor. {{(pid=61923) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 529.508319] env[61923]: DEBUG nova.virt.hardware [None req-7781f812-6fae-4bd6-9d91-ec6098c06372 tempest-InstanceActionsTestJSON-2015629344 tempest-InstanceActionsTestJSON-2015629344-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-29T20:07:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-29T20:07:35Z,direct_url=,disk_format='vmdk',id=0f153f63-ae0a-45b1-b7f5-7f9b673c947f,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='4e7b5e3b3c3443c8bd5c70f8a15739f5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-29T20:07:36Z,virtual_size=,visibility=), allow threads: False {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 529.508638] env[61923]: DEBUG nova.virt.hardware [None req-7781f812-6fae-4bd6-9d91-ec6098c06372 tempest-InstanceActionsTestJSON-2015629344 tempest-InstanceActionsTestJSON-2015629344-project-member] Flavor limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 529.508638] env[61923]: DEBUG nova.virt.hardware [None req-7781f812-6fae-4bd6-9d91-ec6098c06372 tempest-InstanceActionsTestJSON-2015629344 tempest-InstanceActionsTestJSON-2015629344-project-member] Image limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 529.508872] env[61923]: DEBUG nova.virt.hardware [None req-7781f812-6fae-4bd6-9d91-ec6098c06372 tempest-InstanceActionsTestJSON-2015629344 tempest-InstanceActionsTestJSON-2015629344-project-member] Flavor pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 529.509042] env[61923]: DEBUG nova.virt.hardware [None req-7781f812-6fae-4bd6-9d91-ec6098c06372 tempest-InstanceActionsTestJSON-2015629344 tempest-InstanceActionsTestJSON-2015629344-project-member] Image pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 529.510076] env[61923]: DEBUG nova.virt.hardware [None req-7781f812-6fae-4bd6-9d91-ec6098c06372 tempest-InstanceActionsTestJSON-2015629344 tempest-InstanceActionsTestJSON-2015629344-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 529.510247] env[61923]: DEBUG nova.virt.hardware [None req-7781f812-6fae-4bd6-9d91-ec6098c06372 tempest-InstanceActionsTestJSON-2015629344 tempest-InstanceActionsTestJSON-2015629344-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 529.510312] env[61923]: DEBUG nova.virt.hardware [None req-7781f812-6fae-4bd6-9d91-ec6098c06372 tempest-InstanceActionsTestJSON-2015629344 tempest-InstanceActionsTestJSON-2015629344-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 529.510485] env[61923]: DEBUG nova.virt.hardware [None req-7781f812-6fae-4bd6-9d91-ec6098c06372 tempest-InstanceActionsTestJSON-2015629344 tempest-InstanceActionsTestJSON-2015629344-project-member] Got 1 possible topologies {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 529.511027] env[61923]: DEBUG nova.virt.hardware [None req-7781f812-6fae-4bd6-9d91-ec6098c06372 tempest-InstanceActionsTestJSON-2015629344 tempest-InstanceActionsTestJSON-2015629344-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 529.511027] env[61923]: DEBUG nova.virt.hardware [None req-7781f812-6fae-4bd6-9d91-ec6098c06372 tempest-InstanceActionsTestJSON-2015629344 tempest-InstanceActionsTestJSON-2015629344-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 529.511800] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dae33c99-a9a6-45be-be19-c76cd352ccf3 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 529.521398] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0434f99c-b9b6-42df-a600-ed27ffb87a1b {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 529.796071] env[61923]: DEBUG nova.network.neutron [None req-7781f812-6fae-4bd6-9d91-ec6098c06372 tempest-InstanceActionsTestJSON-2015629344 tempest-InstanceActionsTestJSON-2015629344-project-member] [instance: 97148ea5-398a-4974-b8f6-62290e519794] Successfully created port: 7ec53f9e-f055-4a57-8228-7efbcdb7360f {{(pid=61923) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 529.835696] env[61923]: DEBUG nova.network.neutron [None req-dc345cdf-ff58-4514-9baa-d2d8d13601af tempest-MigrationsAdminTest-1311988845 tempest-MigrationsAdminTest-1311988845-project-member] [instance: 242f5cab-3c0c-4f8d-82a8-0cae8d3d419e] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 530.077937] env[61923]: DEBUG nova.network.neutron [None req-dc345cdf-ff58-4514-9baa-d2d8d13601af tempest-MigrationsAdminTest-1311988845 tempest-MigrationsAdminTest-1311988845-project-member] [instance: 242f5cab-3c0c-4f8d-82a8-0cae8d3d419e] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 530.425764] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58c22936-1bec-4ec6-945a-db99bd2dab42 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 530.436077] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc9e0571-952e-4027-ad95-18199160a580 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 530.473988] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17194bdc-08ab-4016-996c-5e4ee5c885fb {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 530.486068] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64217743-9491-42cd-87ed-c4d9934204d6 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 530.500177] env[61923]: DEBUG nova.compute.provider_tree [None req-844a2215-1038-4a30-8a79-40c1c3117bde tempest-FloatingIPsAssociationNegativeTestJSON-1301210771 tempest-FloatingIPsAssociationNegativeTestJSON-1301210771-project-member] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 530.581881] env[61923]: DEBUG oslo_concurrency.lockutils [None req-dc345cdf-ff58-4514-9baa-d2d8d13601af tempest-MigrationsAdminTest-1311988845 tempest-MigrationsAdminTest-1311988845-project-member] Releasing lock "refresh_cache-242f5cab-3c0c-4f8d-82a8-0cae8d3d419e" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 530.582222] env[61923]: DEBUG nova.compute.manager [None req-dc345cdf-ff58-4514-9baa-d2d8d13601af tempest-MigrationsAdminTest-1311988845 tempest-MigrationsAdminTest-1311988845-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61923) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 530.582391] env[61923]: DEBUG nova.compute.manager [None req-dc345cdf-ff58-4514-9baa-d2d8d13601af tempest-MigrationsAdminTest-1311988845 tempest-MigrationsAdminTest-1311988845-project-member] [instance: 242f5cab-3c0c-4f8d-82a8-0cae8d3d419e] Deallocating network for instance {{(pid=61923) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 530.582555] env[61923]: DEBUG nova.network.neutron [None req-dc345cdf-ff58-4514-9baa-d2d8d13601af tempest-MigrationsAdminTest-1311988845 tempest-MigrationsAdminTest-1311988845-project-member] [instance: 242f5cab-3c0c-4f8d-82a8-0cae8d3d419e] deallocate_for_instance() {{(pid=61923) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 530.625537] env[61923]: DEBUG nova.network.neutron [None req-dc345cdf-ff58-4514-9baa-d2d8d13601af tempest-MigrationsAdminTest-1311988845 tempest-MigrationsAdminTest-1311988845-project-member] [instance: 242f5cab-3c0c-4f8d-82a8-0cae8d3d419e] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 531.008340] env[61923]: DEBUG nova.scheduler.client.report [None req-844a2215-1038-4a30-8a79-40c1c3117bde tempest-FloatingIPsAssociationNegativeTestJSON-1301210771 tempest-FloatingIPsAssociationNegativeTestJSON-1301210771-project-member] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 531.129143] env[61923]: DEBUG nova.network.neutron [None req-dc345cdf-ff58-4514-9baa-d2d8d13601af tempest-MigrationsAdminTest-1311988845 tempest-MigrationsAdminTest-1311988845-project-member] [instance: 242f5cab-3c0c-4f8d-82a8-0cae8d3d419e] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 531.515866] env[61923]: DEBUG oslo_concurrency.lockutils [None req-844a2215-1038-4a30-8a79-40c1c3117bde tempest-FloatingIPsAssociationNegativeTestJSON-1301210771 tempest-FloatingIPsAssociationNegativeTestJSON-1301210771-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.261s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 531.517076] env[61923]: DEBUG nova.compute.manager [None req-844a2215-1038-4a30-8a79-40c1c3117bde tempest-FloatingIPsAssociationNegativeTestJSON-1301210771 tempest-FloatingIPsAssociationNegativeTestJSON-1301210771-project-member] [instance: 5d57303a-ae17-4b92-8f12-303677eb0d7e] Start building networks asynchronously for instance. {{(pid=61923) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 531.520800] env[61923]: DEBUG oslo_concurrency.lockutils [None req-a2070053-8e36-4abe-a723-0b93480b3283 tempest-ServerDiagnosticsTest-1095156200 tempest-ServerDiagnosticsTest-1095156200-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 9.528s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 531.632793] env[61923]: INFO nova.compute.manager [None req-dc345cdf-ff58-4514-9baa-d2d8d13601af tempest-MigrationsAdminTest-1311988845 tempest-MigrationsAdminTest-1311988845-project-member] [instance: 242f5cab-3c0c-4f8d-82a8-0cae8d3d419e] Took 1.05 seconds to deallocate network for instance. [ 531.903770] env[61923]: DEBUG oslo_concurrency.lockutils [None req-3b10f02e-4025-4174-94c7-8515a7e12175 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] Acquiring lock "f7cc960c-e06a-4c58-9367-ec8771fe09d5" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 531.904383] env[61923]: DEBUG oslo_concurrency.lockutils [None req-3b10f02e-4025-4174-94c7-8515a7e12175 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] Lock "f7cc960c-e06a-4c58-9367-ec8771fe09d5" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 532.026474] env[61923]: DEBUG nova.compute.utils [None req-844a2215-1038-4a30-8a79-40c1c3117bde tempest-FloatingIPsAssociationNegativeTestJSON-1301210771 tempest-FloatingIPsAssociationNegativeTestJSON-1301210771-project-member] Using /dev/sd instead of None {{(pid=61923) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 532.032195] env[61923]: DEBUG nova.compute.manager [None req-844a2215-1038-4a30-8a79-40c1c3117bde tempest-FloatingIPsAssociationNegativeTestJSON-1301210771 tempest-FloatingIPsAssociationNegativeTestJSON-1301210771-project-member] [instance: 5d57303a-ae17-4b92-8f12-303677eb0d7e] Allocating IP information in the background. {{(pid=61923) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 532.032465] env[61923]: DEBUG nova.network.neutron [None req-844a2215-1038-4a30-8a79-40c1c3117bde tempest-FloatingIPsAssociationNegativeTestJSON-1301210771 tempest-FloatingIPsAssociationNegativeTestJSON-1301210771-project-member] [instance: 5d57303a-ae17-4b92-8f12-303677eb0d7e] allocate_for_instance() {{(pid=61923) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 532.246898] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a11d636-201e-4e54-ba9f-4ac1a5d77112 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 532.255687] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33edd92e-0af7-4a87-819c-4999a5cf59e3 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 532.294510] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9737002-7a31-4242-806c-067728000ebe {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 532.303452] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e583bc29-b7a0-445d-8184-80f6b1fcbba1 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 532.320814] env[61923]: DEBUG nova.compute.provider_tree [None req-a2070053-8e36-4abe-a723-0b93480b3283 tempest-ServerDiagnosticsTest-1095156200 tempest-ServerDiagnosticsTest-1095156200-project-member] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 532.323161] env[61923]: DEBUG nova.policy [None req-844a2215-1038-4a30-8a79-40c1c3117bde tempest-FloatingIPsAssociationNegativeTestJSON-1301210771 tempest-FloatingIPsAssociationNegativeTestJSON-1301210771-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ae74bb78ee094384a234a11389986e74', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'cd7f9a9ec68b47ddb7ea21b5af26992c', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61923) authorize /opt/stack/nova/nova/policy.py:201}} [ 532.375806] env[61923]: ERROR nova.compute.manager [None req-8a6b6bc2-54c9-4389-ab30-057158d86227 tempest-ServerDiagnosticsNegativeTest-664316905 tempest-ServerDiagnosticsNegativeTest-664316905-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 6bf512ed-8713-43d4-bc7e-ce2b9a514f52, please check neutron logs for more information. [ 532.375806] env[61923]: ERROR nova.compute.manager Traceback (most recent call last): [ 532.375806] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 532.375806] env[61923]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 532.375806] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 532.375806] env[61923]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 532.375806] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 532.375806] env[61923]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 532.375806] env[61923]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 532.375806] env[61923]: ERROR nova.compute.manager self.force_reraise() [ 532.375806] env[61923]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 532.375806] env[61923]: ERROR nova.compute.manager raise self.value [ 532.375806] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 532.375806] env[61923]: ERROR nova.compute.manager updated_port = self._update_port( [ 532.375806] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 532.375806] env[61923]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 532.376347] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 532.376347] env[61923]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 532.376347] env[61923]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 6bf512ed-8713-43d4-bc7e-ce2b9a514f52, please check neutron logs for more information. [ 532.376347] env[61923]: ERROR nova.compute.manager [ 532.376347] env[61923]: Traceback (most recent call last): [ 532.376347] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 532.376347] env[61923]: listener.cb(fileno) [ 532.376347] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 532.376347] env[61923]: result = function(*args, **kwargs) [ 532.376347] env[61923]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 532.376347] env[61923]: return func(*args, **kwargs) [ 532.376347] env[61923]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 532.376347] env[61923]: raise e [ 532.376347] env[61923]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 532.376347] env[61923]: nwinfo = self.network_api.allocate_for_instance( [ 532.376347] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 532.376347] env[61923]: created_port_ids = self._update_ports_for_instance( [ 532.376347] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 532.376347] env[61923]: with excutils.save_and_reraise_exception(): [ 532.376347] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 532.376347] env[61923]: self.force_reraise() [ 532.376347] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 532.376347] env[61923]: raise self.value [ 532.376347] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 532.376347] env[61923]: updated_port = self._update_port( [ 532.376347] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 532.376347] env[61923]: _ensure_no_port_binding_failure(port) [ 532.376347] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 532.376347] env[61923]: raise exception.PortBindingFailed(port_id=port['id']) [ 532.377087] env[61923]: nova.exception.PortBindingFailed: Binding failed for port 6bf512ed-8713-43d4-bc7e-ce2b9a514f52, please check neutron logs for more information. [ 532.377087] env[61923]: Removing descriptor: 15 [ 532.377087] env[61923]: ERROR nova.compute.manager [None req-8a6b6bc2-54c9-4389-ab30-057158d86227 tempest-ServerDiagnosticsNegativeTest-664316905 tempest-ServerDiagnosticsNegativeTest-664316905-project-member] [instance: b0a45314-eed8-4d0a-bea2-d4b47dc0bc80] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 6bf512ed-8713-43d4-bc7e-ce2b9a514f52, please check neutron logs for more information. [ 532.377087] env[61923]: ERROR nova.compute.manager [instance: b0a45314-eed8-4d0a-bea2-d4b47dc0bc80] Traceback (most recent call last): [ 532.377087] env[61923]: ERROR nova.compute.manager [instance: b0a45314-eed8-4d0a-bea2-d4b47dc0bc80] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 532.377087] env[61923]: ERROR nova.compute.manager [instance: b0a45314-eed8-4d0a-bea2-d4b47dc0bc80] yield resources [ 532.377087] env[61923]: ERROR nova.compute.manager [instance: b0a45314-eed8-4d0a-bea2-d4b47dc0bc80] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 532.377087] env[61923]: ERROR nova.compute.manager [instance: b0a45314-eed8-4d0a-bea2-d4b47dc0bc80] self.driver.spawn(context, instance, image_meta, [ 532.377087] env[61923]: ERROR nova.compute.manager [instance: b0a45314-eed8-4d0a-bea2-d4b47dc0bc80] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 532.377087] env[61923]: ERROR nova.compute.manager [instance: b0a45314-eed8-4d0a-bea2-d4b47dc0bc80] self._vmops.spawn(context, instance, image_meta, injected_files, [ 532.377087] env[61923]: ERROR nova.compute.manager [instance: b0a45314-eed8-4d0a-bea2-d4b47dc0bc80] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 532.377087] env[61923]: ERROR nova.compute.manager [instance: b0a45314-eed8-4d0a-bea2-d4b47dc0bc80] vm_ref = self.build_virtual_machine(instance, [ 532.377549] env[61923]: ERROR nova.compute.manager [instance: b0a45314-eed8-4d0a-bea2-d4b47dc0bc80] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 532.377549] env[61923]: ERROR nova.compute.manager [instance: b0a45314-eed8-4d0a-bea2-d4b47dc0bc80] vif_infos = vmwarevif.get_vif_info(self._session, [ 532.377549] env[61923]: ERROR nova.compute.manager [instance: b0a45314-eed8-4d0a-bea2-d4b47dc0bc80] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 532.377549] env[61923]: ERROR nova.compute.manager [instance: b0a45314-eed8-4d0a-bea2-d4b47dc0bc80] for vif in network_info: [ 532.377549] env[61923]: ERROR nova.compute.manager [instance: b0a45314-eed8-4d0a-bea2-d4b47dc0bc80] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 532.377549] env[61923]: ERROR nova.compute.manager [instance: b0a45314-eed8-4d0a-bea2-d4b47dc0bc80] return self._sync_wrapper(fn, *args, **kwargs) [ 532.377549] env[61923]: ERROR nova.compute.manager [instance: b0a45314-eed8-4d0a-bea2-d4b47dc0bc80] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 532.377549] env[61923]: ERROR nova.compute.manager [instance: b0a45314-eed8-4d0a-bea2-d4b47dc0bc80] self.wait() [ 532.377549] env[61923]: ERROR nova.compute.manager [instance: b0a45314-eed8-4d0a-bea2-d4b47dc0bc80] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 532.377549] env[61923]: ERROR nova.compute.manager [instance: b0a45314-eed8-4d0a-bea2-d4b47dc0bc80] self[:] = self._gt.wait() [ 532.377549] env[61923]: ERROR nova.compute.manager [instance: b0a45314-eed8-4d0a-bea2-d4b47dc0bc80] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 532.377549] env[61923]: ERROR nova.compute.manager [instance: b0a45314-eed8-4d0a-bea2-d4b47dc0bc80] return self._exit_event.wait() [ 532.377549] env[61923]: ERROR nova.compute.manager [instance: b0a45314-eed8-4d0a-bea2-d4b47dc0bc80] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 532.377873] env[61923]: ERROR nova.compute.manager [instance: b0a45314-eed8-4d0a-bea2-d4b47dc0bc80] result = hub.switch() [ 532.377873] env[61923]: ERROR nova.compute.manager [instance: b0a45314-eed8-4d0a-bea2-d4b47dc0bc80] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 532.377873] env[61923]: ERROR nova.compute.manager [instance: b0a45314-eed8-4d0a-bea2-d4b47dc0bc80] return self.greenlet.switch() [ 532.377873] env[61923]: ERROR nova.compute.manager [instance: b0a45314-eed8-4d0a-bea2-d4b47dc0bc80] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 532.377873] env[61923]: ERROR nova.compute.manager [instance: b0a45314-eed8-4d0a-bea2-d4b47dc0bc80] result = function(*args, **kwargs) [ 532.377873] env[61923]: ERROR nova.compute.manager [instance: b0a45314-eed8-4d0a-bea2-d4b47dc0bc80] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 532.377873] env[61923]: ERROR nova.compute.manager [instance: b0a45314-eed8-4d0a-bea2-d4b47dc0bc80] return func(*args, **kwargs) [ 532.377873] env[61923]: ERROR nova.compute.manager [instance: b0a45314-eed8-4d0a-bea2-d4b47dc0bc80] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 532.377873] env[61923]: ERROR nova.compute.manager [instance: b0a45314-eed8-4d0a-bea2-d4b47dc0bc80] raise e [ 532.377873] env[61923]: ERROR nova.compute.manager [instance: b0a45314-eed8-4d0a-bea2-d4b47dc0bc80] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 532.377873] env[61923]: ERROR nova.compute.manager [instance: b0a45314-eed8-4d0a-bea2-d4b47dc0bc80] nwinfo = self.network_api.allocate_for_instance( [ 532.377873] env[61923]: ERROR nova.compute.manager [instance: b0a45314-eed8-4d0a-bea2-d4b47dc0bc80] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 532.377873] env[61923]: ERROR nova.compute.manager [instance: b0a45314-eed8-4d0a-bea2-d4b47dc0bc80] created_port_ids = self._update_ports_for_instance( [ 532.378311] env[61923]: ERROR nova.compute.manager [instance: b0a45314-eed8-4d0a-bea2-d4b47dc0bc80] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 532.378311] env[61923]: ERROR nova.compute.manager [instance: b0a45314-eed8-4d0a-bea2-d4b47dc0bc80] with excutils.save_and_reraise_exception(): [ 532.378311] env[61923]: ERROR nova.compute.manager [instance: b0a45314-eed8-4d0a-bea2-d4b47dc0bc80] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 532.378311] env[61923]: ERROR nova.compute.manager [instance: b0a45314-eed8-4d0a-bea2-d4b47dc0bc80] self.force_reraise() [ 532.378311] env[61923]: ERROR nova.compute.manager [instance: b0a45314-eed8-4d0a-bea2-d4b47dc0bc80] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 532.378311] env[61923]: ERROR nova.compute.manager [instance: b0a45314-eed8-4d0a-bea2-d4b47dc0bc80] raise self.value [ 532.378311] env[61923]: ERROR nova.compute.manager [instance: b0a45314-eed8-4d0a-bea2-d4b47dc0bc80] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 532.378311] env[61923]: ERROR nova.compute.manager [instance: b0a45314-eed8-4d0a-bea2-d4b47dc0bc80] updated_port = self._update_port( [ 532.378311] env[61923]: ERROR nova.compute.manager [instance: b0a45314-eed8-4d0a-bea2-d4b47dc0bc80] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 532.378311] env[61923]: ERROR nova.compute.manager [instance: b0a45314-eed8-4d0a-bea2-d4b47dc0bc80] _ensure_no_port_binding_failure(port) [ 532.378311] env[61923]: ERROR nova.compute.manager [instance: b0a45314-eed8-4d0a-bea2-d4b47dc0bc80] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 532.378311] env[61923]: ERROR nova.compute.manager [instance: b0a45314-eed8-4d0a-bea2-d4b47dc0bc80] raise exception.PortBindingFailed(port_id=port['id']) [ 532.378657] env[61923]: ERROR nova.compute.manager [instance: b0a45314-eed8-4d0a-bea2-d4b47dc0bc80] nova.exception.PortBindingFailed: Binding failed for port 6bf512ed-8713-43d4-bc7e-ce2b9a514f52, please check neutron logs for more information. [ 532.378657] env[61923]: ERROR nova.compute.manager [instance: b0a45314-eed8-4d0a-bea2-d4b47dc0bc80] [ 532.378657] env[61923]: INFO nova.compute.manager [None req-8a6b6bc2-54c9-4389-ab30-057158d86227 tempest-ServerDiagnosticsNegativeTest-664316905 tempest-ServerDiagnosticsNegativeTest-664316905-project-member] [instance: b0a45314-eed8-4d0a-bea2-d4b47dc0bc80] Terminating instance [ 532.382292] env[61923]: DEBUG oslo_concurrency.lockutils [None req-8a6b6bc2-54c9-4389-ab30-057158d86227 tempest-ServerDiagnosticsNegativeTest-664316905 tempest-ServerDiagnosticsNegativeTest-664316905-project-member] Acquiring lock "refresh_cache-b0a45314-eed8-4d0a-bea2-d4b47dc0bc80" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 532.382292] env[61923]: DEBUG oslo_concurrency.lockutils [None req-8a6b6bc2-54c9-4389-ab30-057158d86227 tempest-ServerDiagnosticsNegativeTest-664316905 tempest-ServerDiagnosticsNegativeTest-664316905-project-member] Acquired lock "refresh_cache-b0a45314-eed8-4d0a-bea2-d4b47dc0bc80" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 532.382292] env[61923]: DEBUG nova.network.neutron [None req-8a6b6bc2-54c9-4389-ab30-057158d86227 tempest-ServerDiagnosticsNegativeTest-664316905 tempest-ServerDiagnosticsNegativeTest-664316905-project-member] [instance: b0a45314-eed8-4d0a-bea2-d4b47dc0bc80] Building network info cache for instance {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 532.410285] env[61923]: DEBUG nova.compute.manager [None req-3b10f02e-4025-4174-94c7-8515a7e12175 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] [instance: f7cc960c-e06a-4c58-9367-ec8771fe09d5] Starting instance... {{(pid=61923) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 532.529911] env[61923]: DEBUG nova.compute.manager [None req-844a2215-1038-4a30-8a79-40c1c3117bde tempest-FloatingIPsAssociationNegativeTestJSON-1301210771 tempest-FloatingIPsAssociationNegativeTestJSON-1301210771-project-member] [instance: 5d57303a-ae17-4b92-8f12-303677eb0d7e] Start building block device mappings for instance. {{(pid=61923) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 532.550610] env[61923]: DEBUG oslo_concurrency.lockutils [None req-d9e2f878-40e6-4b41-8b07-da611f2f81f1 tempest-VolumesAssistedSnapshotsTest-2043366068 tempest-VolumesAssistedSnapshotsTest-2043366068-project-member] Acquiring lock "06bda5ee-c7a8-47d1-9e0c-a5af67d40006" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 532.550897] env[61923]: DEBUG oslo_concurrency.lockutils [None req-d9e2f878-40e6-4b41-8b07-da611f2f81f1 tempest-VolumesAssistedSnapshotsTest-2043366068 tempest-VolumesAssistedSnapshotsTest-2043366068-project-member] Lock "06bda5ee-c7a8-47d1-9e0c-a5af67d40006" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 532.686153] env[61923]: INFO nova.scheduler.client.report [None req-dc345cdf-ff58-4514-9baa-d2d8d13601af tempest-MigrationsAdminTest-1311988845 tempest-MigrationsAdminTest-1311988845-project-member] Deleted allocations for instance 242f5cab-3c0c-4f8d-82a8-0cae8d3d419e [ 532.743206] env[61923]: DEBUG nova.compute.manager [req-f4863623-308d-474c-9e99-25719cffeb5c req-d8001bb8-760a-4b78-918d-8fd6cbdd3480 service nova] [instance: b0a45314-eed8-4d0a-bea2-d4b47dc0bc80] Received event network-changed-6bf512ed-8713-43d4-bc7e-ce2b9a514f52 {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 532.743427] env[61923]: DEBUG nova.compute.manager [req-f4863623-308d-474c-9e99-25719cffeb5c req-d8001bb8-760a-4b78-918d-8fd6cbdd3480 service nova] [instance: b0a45314-eed8-4d0a-bea2-d4b47dc0bc80] Refreshing instance network info cache due to event network-changed-6bf512ed-8713-43d4-bc7e-ce2b9a514f52. {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 532.743623] env[61923]: DEBUG oslo_concurrency.lockutils [req-f4863623-308d-474c-9e99-25719cffeb5c req-d8001bb8-760a-4b78-918d-8fd6cbdd3480 service nova] Acquiring lock "refresh_cache-b0a45314-eed8-4d0a-bea2-d4b47dc0bc80" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 532.831020] env[61923]: DEBUG nova.scheduler.client.report [None req-a2070053-8e36-4abe-a723-0b93480b3283 tempest-ServerDiagnosticsTest-1095156200 tempest-ServerDiagnosticsTest-1095156200-project-member] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 532.908929] env[61923]: DEBUG nova.network.neutron [None req-8a6b6bc2-54c9-4389-ab30-057158d86227 tempest-ServerDiagnosticsNegativeTest-664316905 tempest-ServerDiagnosticsNegativeTest-664316905-project-member] [instance: b0a45314-eed8-4d0a-bea2-d4b47dc0bc80] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 532.948766] env[61923]: DEBUG oslo_concurrency.lockutils [None req-3b10f02e-4025-4174-94c7-8515a7e12175 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 533.203680] env[61923]: DEBUG oslo_concurrency.lockutils [None req-dc345cdf-ff58-4514-9baa-d2d8d13601af tempest-MigrationsAdminTest-1311988845 tempest-MigrationsAdminTest-1311988845-project-member] Lock "242f5cab-3c0c-4f8d-82a8-0cae8d3d419e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 22.546s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 533.214684] env[61923]: DEBUG nova.network.neutron [None req-844a2215-1038-4a30-8a79-40c1c3117bde tempest-FloatingIPsAssociationNegativeTestJSON-1301210771 tempest-FloatingIPsAssociationNegativeTestJSON-1301210771-project-member] [instance: 5d57303a-ae17-4b92-8f12-303677eb0d7e] Successfully created port: 4f3fea2e-b3d9-4c89-b291-071aa6707d27 {{(pid=61923) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 533.315328] env[61923]: DEBUG nova.network.neutron [None req-8a6b6bc2-54c9-4389-ab30-057158d86227 tempest-ServerDiagnosticsNegativeTest-664316905 tempest-ServerDiagnosticsNegativeTest-664316905-project-member] [instance: b0a45314-eed8-4d0a-bea2-d4b47dc0bc80] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 533.334370] env[61923]: DEBUG oslo_concurrency.lockutils [None req-a2070053-8e36-4abe-a723-0b93480b3283 tempest-ServerDiagnosticsTest-1095156200 tempest-ServerDiagnosticsTest-1095156200-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.814s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 533.335013] env[61923]: ERROR nova.compute.manager [None req-a2070053-8e36-4abe-a723-0b93480b3283 tempest-ServerDiagnosticsTest-1095156200 tempest-ServerDiagnosticsTest-1095156200-project-member] [instance: a134ab6d-b261-4df6-b440-d32a3e578645] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 6aa16849-cb00-49ad-9abe-b5ca6484171d, please check neutron logs for more information. [ 533.335013] env[61923]: ERROR nova.compute.manager [instance: a134ab6d-b261-4df6-b440-d32a3e578645] Traceback (most recent call last): [ 533.335013] env[61923]: ERROR nova.compute.manager [instance: a134ab6d-b261-4df6-b440-d32a3e578645] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 533.335013] env[61923]: ERROR nova.compute.manager [instance: a134ab6d-b261-4df6-b440-d32a3e578645] self.driver.spawn(context, instance, image_meta, [ 533.335013] env[61923]: ERROR nova.compute.manager [instance: a134ab6d-b261-4df6-b440-d32a3e578645] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 533.335013] env[61923]: ERROR nova.compute.manager [instance: a134ab6d-b261-4df6-b440-d32a3e578645] self._vmops.spawn(context, instance, image_meta, injected_files, [ 533.335013] env[61923]: ERROR nova.compute.manager [instance: a134ab6d-b261-4df6-b440-d32a3e578645] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 533.335013] env[61923]: ERROR nova.compute.manager [instance: a134ab6d-b261-4df6-b440-d32a3e578645] vm_ref = self.build_virtual_machine(instance, [ 533.335013] env[61923]: ERROR nova.compute.manager [instance: a134ab6d-b261-4df6-b440-d32a3e578645] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 533.335013] env[61923]: ERROR nova.compute.manager [instance: a134ab6d-b261-4df6-b440-d32a3e578645] vif_infos = vmwarevif.get_vif_info(self._session, [ 533.335013] env[61923]: ERROR nova.compute.manager [instance: a134ab6d-b261-4df6-b440-d32a3e578645] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 533.335357] env[61923]: ERROR nova.compute.manager [instance: a134ab6d-b261-4df6-b440-d32a3e578645] for vif in network_info: [ 533.335357] env[61923]: ERROR nova.compute.manager [instance: a134ab6d-b261-4df6-b440-d32a3e578645] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 533.335357] env[61923]: ERROR nova.compute.manager [instance: a134ab6d-b261-4df6-b440-d32a3e578645] return self._sync_wrapper(fn, *args, **kwargs) [ 533.335357] env[61923]: ERROR nova.compute.manager [instance: a134ab6d-b261-4df6-b440-d32a3e578645] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 533.335357] env[61923]: ERROR nova.compute.manager [instance: a134ab6d-b261-4df6-b440-d32a3e578645] self.wait() [ 533.335357] env[61923]: ERROR nova.compute.manager [instance: a134ab6d-b261-4df6-b440-d32a3e578645] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 533.335357] env[61923]: ERROR nova.compute.manager [instance: a134ab6d-b261-4df6-b440-d32a3e578645] self[:] = self._gt.wait() [ 533.335357] env[61923]: ERROR nova.compute.manager [instance: a134ab6d-b261-4df6-b440-d32a3e578645] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 533.335357] env[61923]: ERROR nova.compute.manager [instance: a134ab6d-b261-4df6-b440-d32a3e578645] return self._exit_event.wait() [ 533.335357] env[61923]: ERROR nova.compute.manager [instance: a134ab6d-b261-4df6-b440-d32a3e578645] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 533.335357] env[61923]: ERROR nova.compute.manager [instance: a134ab6d-b261-4df6-b440-d32a3e578645] result = hub.switch() [ 533.335357] env[61923]: ERROR nova.compute.manager [instance: a134ab6d-b261-4df6-b440-d32a3e578645] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 533.335357] env[61923]: ERROR nova.compute.manager [instance: a134ab6d-b261-4df6-b440-d32a3e578645] return self.greenlet.switch() [ 533.335669] env[61923]: ERROR nova.compute.manager [instance: a134ab6d-b261-4df6-b440-d32a3e578645] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 533.335669] env[61923]: ERROR nova.compute.manager [instance: a134ab6d-b261-4df6-b440-d32a3e578645] result = function(*args, **kwargs) [ 533.335669] env[61923]: ERROR nova.compute.manager [instance: a134ab6d-b261-4df6-b440-d32a3e578645] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 533.335669] env[61923]: ERROR nova.compute.manager [instance: a134ab6d-b261-4df6-b440-d32a3e578645] return func(*args, **kwargs) [ 533.335669] env[61923]: ERROR nova.compute.manager [instance: a134ab6d-b261-4df6-b440-d32a3e578645] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 533.335669] env[61923]: ERROR nova.compute.manager [instance: a134ab6d-b261-4df6-b440-d32a3e578645] raise e [ 533.335669] env[61923]: ERROR nova.compute.manager [instance: a134ab6d-b261-4df6-b440-d32a3e578645] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 533.335669] env[61923]: ERROR nova.compute.manager [instance: a134ab6d-b261-4df6-b440-d32a3e578645] nwinfo = self.network_api.allocate_for_instance( [ 533.335669] env[61923]: ERROR nova.compute.manager [instance: a134ab6d-b261-4df6-b440-d32a3e578645] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 533.335669] env[61923]: ERROR nova.compute.manager [instance: a134ab6d-b261-4df6-b440-d32a3e578645] created_port_ids = self._update_ports_for_instance( [ 533.335669] env[61923]: ERROR nova.compute.manager [instance: a134ab6d-b261-4df6-b440-d32a3e578645] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 533.335669] env[61923]: ERROR nova.compute.manager [instance: a134ab6d-b261-4df6-b440-d32a3e578645] with excutils.save_and_reraise_exception(): [ 533.335669] env[61923]: ERROR nova.compute.manager [instance: a134ab6d-b261-4df6-b440-d32a3e578645] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 533.336019] env[61923]: ERROR nova.compute.manager [instance: a134ab6d-b261-4df6-b440-d32a3e578645] self.force_reraise() [ 533.336019] env[61923]: ERROR nova.compute.manager [instance: a134ab6d-b261-4df6-b440-d32a3e578645] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 533.336019] env[61923]: ERROR nova.compute.manager [instance: a134ab6d-b261-4df6-b440-d32a3e578645] raise self.value [ 533.336019] env[61923]: ERROR nova.compute.manager [instance: a134ab6d-b261-4df6-b440-d32a3e578645] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 533.336019] env[61923]: ERROR nova.compute.manager [instance: a134ab6d-b261-4df6-b440-d32a3e578645] updated_port = self._update_port( [ 533.336019] env[61923]: ERROR nova.compute.manager [instance: a134ab6d-b261-4df6-b440-d32a3e578645] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 533.336019] env[61923]: ERROR nova.compute.manager [instance: a134ab6d-b261-4df6-b440-d32a3e578645] _ensure_no_port_binding_failure(port) [ 533.336019] env[61923]: ERROR nova.compute.manager [instance: a134ab6d-b261-4df6-b440-d32a3e578645] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 533.336019] env[61923]: ERROR nova.compute.manager [instance: a134ab6d-b261-4df6-b440-d32a3e578645] raise exception.PortBindingFailed(port_id=port['id']) [ 533.336019] env[61923]: ERROR nova.compute.manager [instance: a134ab6d-b261-4df6-b440-d32a3e578645] nova.exception.PortBindingFailed: Binding failed for port 6aa16849-cb00-49ad-9abe-b5ca6484171d, please check neutron logs for more information. [ 533.336019] env[61923]: ERROR nova.compute.manager [instance: a134ab6d-b261-4df6-b440-d32a3e578645] [ 533.336295] env[61923]: DEBUG nova.compute.utils [None req-a2070053-8e36-4abe-a723-0b93480b3283 tempest-ServerDiagnosticsTest-1095156200 tempest-ServerDiagnosticsTest-1095156200-project-member] [instance: a134ab6d-b261-4df6-b440-d32a3e578645] Binding failed for port 6aa16849-cb00-49ad-9abe-b5ca6484171d, please check neutron logs for more information. {{(pid=61923) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 533.337722] env[61923]: DEBUG oslo_concurrency.lockutils [None req-f30f4038-8fd0-4bbd-8b35-7fb8e0dea3a4 tempest-SecurityGroupsTestJSON-509490211 tempest-SecurityGroupsTestJSON-509490211-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 10.304s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 533.343651] env[61923]: INFO nova.compute.claims [None req-f30f4038-8fd0-4bbd-8b35-7fb8e0dea3a4 tempest-SecurityGroupsTestJSON-509490211 tempest-SecurityGroupsTestJSON-509490211-project-member] [instance: 59fc11e8-27be-4553-9de9-3eb5ea612ec4] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 533.346730] env[61923]: DEBUG nova.compute.manager [None req-a2070053-8e36-4abe-a723-0b93480b3283 tempest-ServerDiagnosticsTest-1095156200 tempest-ServerDiagnosticsTest-1095156200-project-member] [instance: a134ab6d-b261-4df6-b440-d32a3e578645] Build of instance a134ab6d-b261-4df6-b440-d32a3e578645 was re-scheduled: Binding failed for port 6aa16849-cb00-49ad-9abe-b5ca6484171d, please check neutron logs for more information. {{(pid=61923) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 533.347175] env[61923]: DEBUG nova.compute.manager [None req-a2070053-8e36-4abe-a723-0b93480b3283 tempest-ServerDiagnosticsTest-1095156200 tempest-ServerDiagnosticsTest-1095156200-project-member] [instance: a134ab6d-b261-4df6-b440-d32a3e578645] Unplugging VIFs for instance {{(pid=61923) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 533.347395] env[61923]: DEBUG oslo_concurrency.lockutils [None req-a2070053-8e36-4abe-a723-0b93480b3283 tempest-ServerDiagnosticsTest-1095156200 tempest-ServerDiagnosticsTest-1095156200-project-member] Acquiring lock "refresh_cache-a134ab6d-b261-4df6-b440-d32a3e578645" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 533.347532] env[61923]: DEBUG oslo_concurrency.lockutils [None req-a2070053-8e36-4abe-a723-0b93480b3283 tempest-ServerDiagnosticsTest-1095156200 tempest-ServerDiagnosticsTest-1095156200-project-member] Acquired lock "refresh_cache-a134ab6d-b261-4df6-b440-d32a3e578645" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 533.347685] env[61923]: DEBUG nova.network.neutron [None req-a2070053-8e36-4abe-a723-0b93480b3283 tempest-ServerDiagnosticsTest-1095156200 tempest-ServerDiagnosticsTest-1095156200-project-member] [instance: a134ab6d-b261-4df6-b440-d32a3e578645] Building network info cache for instance {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 533.547168] env[61923]: DEBUG nova.compute.manager [None req-844a2215-1038-4a30-8a79-40c1c3117bde tempest-FloatingIPsAssociationNegativeTestJSON-1301210771 tempest-FloatingIPsAssociationNegativeTestJSON-1301210771-project-member] [instance: 5d57303a-ae17-4b92-8f12-303677eb0d7e] Start spawning the instance on the hypervisor. {{(pid=61923) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 533.709620] env[61923]: DEBUG nova.compute.manager [None req-d9e2f878-40e6-4b41-8b07-da611f2f81f1 tempest-VolumesAssistedSnapshotsTest-2043366068 tempest-VolumesAssistedSnapshotsTest-2043366068-project-member] [instance: 06bda5ee-c7a8-47d1-9e0c-a5af67d40006] Starting instance... {{(pid=61923) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 533.818890] env[61923]: DEBUG oslo_concurrency.lockutils [None req-8a6b6bc2-54c9-4389-ab30-057158d86227 tempest-ServerDiagnosticsNegativeTest-664316905 tempest-ServerDiagnosticsNegativeTest-664316905-project-member] Releasing lock "refresh_cache-b0a45314-eed8-4d0a-bea2-d4b47dc0bc80" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 533.819357] env[61923]: DEBUG nova.compute.manager [None req-8a6b6bc2-54c9-4389-ab30-057158d86227 tempest-ServerDiagnosticsNegativeTest-664316905 tempest-ServerDiagnosticsNegativeTest-664316905-project-member] [instance: b0a45314-eed8-4d0a-bea2-d4b47dc0bc80] Start destroying the instance on the hypervisor. {{(pid=61923) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 533.819580] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-8a6b6bc2-54c9-4389-ab30-057158d86227 tempest-ServerDiagnosticsNegativeTest-664316905 tempest-ServerDiagnosticsNegativeTest-664316905-project-member] [instance: b0a45314-eed8-4d0a-bea2-d4b47dc0bc80] Destroying instance {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 533.819842] env[61923]: DEBUG oslo_concurrency.lockutils [req-f4863623-308d-474c-9e99-25719cffeb5c req-d8001bb8-760a-4b78-918d-8fd6cbdd3480 service nova] Acquired lock "refresh_cache-b0a45314-eed8-4d0a-bea2-d4b47dc0bc80" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 533.820253] env[61923]: DEBUG nova.network.neutron [req-f4863623-308d-474c-9e99-25719cffeb5c req-d8001bb8-760a-4b78-918d-8fd6cbdd3480 service nova] [instance: b0a45314-eed8-4d0a-bea2-d4b47dc0bc80] Refreshing network info cache for port 6bf512ed-8713-43d4-bc7e-ce2b9a514f52 {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 533.822775] env[61923]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-3c93f8c3-b23e-4608-80fc-1b471f463adf {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 533.833767] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8b291ee-476d-42df-9ec5-ebb5913e6005 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 533.859069] env[61923]: WARNING nova.virt.vmwareapi.vmops [None req-8a6b6bc2-54c9-4389-ab30-057158d86227 tempest-ServerDiagnosticsNegativeTest-664316905 tempest-ServerDiagnosticsNegativeTest-664316905-project-member] [instance: b0a45314-eed8-4d0a-bea2-d4b47dc0bc80] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance b0a45314-eed8-4d0a-bea2-d4b47dc0bc80 could not be found. [ 533.860631] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-8a6b6bc2-54c9-4389-ab30-057158d86227 tempest-ServerDiagnosticsNegativeTest-664316905 tempest-ServerDiagnosticsNegativeTest-664316905-project-member] [instance: b0a45314-eed8-4d0a-bea2-d4b47dc0bc80] Instance destroyed {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 533.860631] env[61923]: INFO nova.compute.manager [None req-8a6b6bc2-54c9-4389-ab30-057158d86227 tempest-ServerDiagnosticsNegativeTest-664316905 tempest-ServerDiagnosticsNegativeTest-664316905-project-member] [instance: b0a45314-eed8-4d0a-bea2-d4b47dc0bc80] Took 0.04 seconds to destroy the instance on the hypervisor. [ 533.860631] env[61923]: DEBUG oslo.service.loopingcall [None req-8a6b6bc2-54c9-4389-ab30-057158d86227 tempest-ServerDiagnosticsNegativeTest-664316905 tempest-ServerDiagnosticsNegativeTest-664316905-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61923) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 533.860631] env[61923]: DEBUG nova.compute.manager [-] [instance: b0a45314-eed8-4d0a-bea2-d4b47dc0bc80] Deallocating network for instance {{(pid=61923) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 533.860631] env[61923]: DEBUG nova.network.neutron [-] [instance: b0a45314-eed8-4d0a-bea2-d4b47dc0bc80] deallocate_for_instance() {{(pid=61923) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 533.922239] env[61923]: DEBUG nova.network.neutron [None req-a2070053-8e36-4abe-a723-0b93480b3283 tempest-ServerDiagnosticsTest-1095156200 tempest-ServerDiagnosticsTest-1095156200-project-member] [instance: a134ab6d-b261-4df6-b440-d32a3e578645] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 533.933437] env[61923]: DEBUG nova.network.neutron [-] [instance: b0a45314-eed8-4d0a-bea2-d4b47dc0bc80] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 534.046162] env[61923]: DEBUG nova.virt.hardware [None req-844a2215-1038-4a30-8a79-40c1c3117bde tempest-FloatingIPsAssociationNegativeTestJSON-1301210771 tempest-FloatingIPsAssociationNegativeTestJSON-1301210771-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-29T20:07:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-29T20:07:35Z,direct_url=,disk_format='vmdk',id=0f153f63-ae0a-45b1-b7f5-7f9b673c947f,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='4e7b5e3b3c3443c8bd5c70f8a15739f5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-29T20:07:36Z,virtual_size=,visibility=), allow threads: False {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 534.046635] env[61923]: DEBUG nova.virt.hardware [None req-844a2215-1038-4a30-8a79-40c1c3117bde tempest-FloatingIPsAssociationNegativeTestJSON-1301210771 tempest-FloatingIPsAssociationNegativeTestJSON-1301210771-project-member] Flavor limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 534.046951] env[61923]: DEBUG nova.virt.hardware [None req-844a2215-1038-4a30-8a79-40c1c3117bde tempest-FloatingIPsAssociationNegativeTestJSON-1301210771 tempest-FloatingIPsAssociationNegativeTestJSON-1301210771-project-member] Image limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 534.047293] env[61923]: DEBUG nova.virt.hardware [None req-844a2215-1038-4a30-8a79-40c1c3117bde tempest-FloatingIPsAssociationNegativeTestJSON-1301210771 tempest-FloatingIPsAssociationNegativeTestJSON-1301210771-project-member] Flavor pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 534.047563] env[61923]: DEBUG nova.virt.hardware [None req-844a2215-1038-4a30-8a79-40c1c3117bde tempest-FloatingIPsAssociationNegativeTestJSON-1301210771 tempest-FloatingIPsAssociationNegativeTestJSON-1301210771-project-member] Image pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 534.047832] env[61923]: DEBUG nova.virt.hardware [None req-844a2215-1038-4a30-8a79-40c1c3117bde tempest-FloatingIPsAssociationNegativeTestJSON-1301210771 tempest-FloatingIPsAssociationNegativeTestJSON-1301210771-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 534.048174] env[61923]: DEBUG nova.virt.hardware [None req-844a2215-1038-4a30-8a79-40c1c3117bde tempest-FloatingIPsAssociationNegativeTestJSON-1301210771 tempest-FloatingIPsAssociationNegativeTestJSON-1301210771-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 534.048476] env[61923]: DEBUG nova.virt.hardware [None req-844a2215-1038-4a30-8a79-40c1c3117bde tempest-FloatingIPsAssociationNegativeTestJSON-1301210771 tempest-FloatingIPsAssociationNegativeTestJSON-1301210771-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 534.048787] env[61923]: DEBUG nova.virt.hardware [None req-844a2215-1038-4a30-8a79-40c1c3117bde tempest-FloatingIPsAssociationNegativeTestJSON-1301210771 tempest-FloatingIPsAssociationNegativeTestJSON-1301210771-project-member] Got 1 possible topologies {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 534.049099] env[61923]: DEBUG nova.virt.hardware [None req-844a2215-1038-4a30-8a79-40c1c3117bde tempest-FloatingIPsAssociationNegativeTestJSON-1301210771 tempest-FloatingIPsAssociationNegativeTestJSON-1301210771-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 534.049522] env[61923]: DEBUG nova.virt.hardware [None req-844a2215-1038-4a30-8a79-40c1c3117bde tempest-FloatingIPsAssociationNegativeTestJSON-1301210771 tempest-FloatingIPsAssociationNegativeTestJSON-1301210771-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 534.050789] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72c0ff2c-a6df-4781-adb5-cb03f7f6fc27 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 534.062238] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8058a8ff-089b-4507-992d-bcb2de1f4624 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 534.109340] env[61923]: DEBUG nova.network.neutron [None req-a2070053-8e36-4abe-a723-0b93480b3283 tempest-ServerDiagnosticsTest-1095156200 tempest-ServerDiagnosticsTest-1095156200-project-member] [instance: a134ab6d-b261-4df6-b440-d32a3e578645] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 534.244976] env[61923]: DEBUG oslo_concurrency.lockutils [None req-d9e2f878-40e6-4b41-8b07-da611f2f81f1 tempest-VolumesAssistedSnapshotsTest-2043366068 tempest-VolumesAssistedSnapshotsTest-2043366068-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 534.306767] env[61923]: ERROR nova.compute.manager [None req-7781f812-6fae-4bd6-9d91-ec6098c06372 tempest-InstanceActionsTestJSON-2015629344 tempest-InstanceActionsTestJSON-2015629344-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 7ec53f9e-f055-4a57-8228-7efbcdb7360f, please check neutron logs for more information. [ 534.306767] env[61923]: ERROR nova.compute.manager Traceback (most recent call last): [ 534.306767] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 534.306767] env[61923]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 534.306767] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 534.306767] env[61923]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 534.306767] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 534.306767] env[61923]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 534.306767] env[61923]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 534.306767] env[61923]: ERROR nova.compute.manager self.force_reraise() [ 534.306767] env[61923]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 534.306767] env[61923]: ERROR nova.compute.manager raise self.value [ 534.306767] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 534.306767] env[61923]: ERROR nova.compute.manager updated_port = self._update_port( [ 534.306767] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 534.306767] env[61923]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 534.307249] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 534.307249] env[61923]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 534.307249] env[61923]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 7ec53f9e-f055-4a57-8228-7efbcdb7360f, please check neutron logs for more information. [ 534.307249] env[61923]: ERROR nova.compute.manager [ 534.307249] env[61923]: Traceback (most recent call last): [ 534.307249] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 534.307249] env[61923]: listener.cb(fileno) [ 534.307249] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 534.307249] env[61923]: result = function(*args, **kwargs) [ 534.307249] env[61923]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 534.307249] env[61923]: return func(*args, **kwargs) [ 534.307249] env[61923]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 534.307249] env[61923]: raise e [ 534.307249] env[61923]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 534.307249] env[61923]: nwinfo = self.network_api.allocate_for_instance( [ 534.307249] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 534.307249] env[61923]: created_port_ids = self._update_ports_for_instance( [ 534.307249] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 534.307249] env[61923]: with excutils.save_and_reraise_exception(): [ 534.307249] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 534.307249] env[61923]: self.force_reraise() [ 534.307249] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 534.307249] env[61923]: raise self.value [ 534.307249] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 534.307249] env[61923]: updated_port = self._update_port( [ 534.307249] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 534.307249] env[61923]: _ensure_no_port_binding_failure(port) [ 534.307249] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 534.307249] env[61923]: raise exception.PortBindingFailed(port_id=port['id']) [ 534.307952] env[61923]: nova.exception.PortBindingFailed: Binding failed for port 7ec53f9e-f055-4a57-8228-7efbcdb7360f, please check neutron logs for more information. [ 534.307952] env[61923]: Removing descriptor: 14 [ 534.307952] env[61923]: ERROR nova.compute.manager [None req-7781f812-6fae-4bd6-9d91-ec6098c06372 tempest-InstanceActionsTestJSON-2015629344 tempest-InstanceActionsTestJSON-2015629344-project-member] [instance: 97148ea5-398a-4974-b8f6-62290e519794] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 7ec53f9e-f055-4a57-8228-7efbcdb7360f, please check neutron logs for more information. [ 534.307952] env[61923]: ERROR nova.compute.manager [instance: 97148ea5-398a-4974-b8f6-62290e519794] Traceback (most recent call last): [ 534.307952] env[61923]: ERROR nova.compute.manager [instance: 97148ea5-398a-4974-b8f6-62290e519794] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 534.307952] env[61923]: ERROR nova.compute.manager [instance: 97148ea5-398a-4974-b8f6-62290e519794] yield resources [ 534.307952] env[61923]: ERROR nova.compute.manager [instance: 97148ea5-398a-4974-b8f6-62290e519794] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 534.307952] env[61923]: ERROR nova.compute.manager [instance: 97148ea5-398a-4974-b8f6-62290e519794] self.driver.spawn(context, instance, image_meta, [ 534.307952] env[61923]: ERROR nova.compute.manager [instance: 97148ea5-398a-4974-b8f6-62290e519794] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 534.307952] env[61923]: ERROR nova.compute.manager [instance: 97148ea5-398a-4974-b8f6-62290e519794] self._vmops.spawn(context, instance, image_meta, injected_files, [ 534.307952] env[61923]: ERROR nova.compute.manager [instance: 97148ea5-398a-4974-b8f6-62290e519794] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 534.307952] env[61923]: ERROR nova.compute.manager [instance: 97148ea5-398a-4974-b8f6-62290e519794] vm_ref = self.build_virtual_machine(instance, [ 534.308294] env[61923]: ERROR nova.compute.manager [instance: 97148ea5-398a-4974-b8f6-62290e519794] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 534.308294] env[61923]: ERROR nova.compute.manager [instance: 97148ea5-398a-4974-b8f6-62290e519794] vif_infos = vmwarevif.get_vif_info(self._session, [ 534.308294] env[61923]: ERROR nova.compute.manager [instance: 97148ea5-398a-4974-b8f6-62290e519794] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 534.308294] env[61923]: ERROR nova.compute.manager [instance: 97148ea5-398a-4974-b8f6-62290e519794] for vif in network_info: [ 534.308294] env[61923]: ERROR nova.compute.manager [instance: 97148ea5-398a-4974-b8f6-62290e519794] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 534.308294] env[61923]: ERROR nova.compute.manager [instance: 97148ea5-398a-4974-b8f6-62290e519794] return self._sync_wrapper(fn, *args, **kwargs) [ 534.308294] env[61923]: ERROR nova.compute.manager [instance: 97148ea5-398a-4974-b8f6-62290e519794] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 534.308294] env[61923]: ERROR nova.compute.manager [instance: 97148ea5-398a-4974-b8f6-62290e519794] self.wait() [ 534.308294] env[61923]: ERROR nova.compute.manager [instance: 97148ea5-398a-4974-b8f6-62290e519794] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 534.308294] env[61923]: ERROR nova.compute.manager [instance: 97148ea5-398a-4974-b8f6-62290e519794] self[:] = self._gt.wait() [ 534.308294] env[61923]: ERROR nova.compute.manager [instance: 97148ea5-398a-4974-b8f6-62290e519794] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 534.308294] env[61923]: ERROR nova.compute.manager [instance: 97148ea5-398a-4974-b8f6-62290e519794] return self._exit_event.wait() [ 534.308294] env[61923]: ERROR nova.compute.manager [instance: 97148ea5-398a-4974-b8f6-62290e519794] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 534.308613] env[61923]: ERROR nova.compute.manager [instance: 97148ea5-398a-4974-b8f6-62290e519794] result = hub.switch() [ 534.308613] env[61923]: ERROR nova.compute.manager [instance: 97148ea5-398a-4974-b8f6-62290e519794] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 534.308613] env[61923]: ERROR nova.compute.manager [instance: 97148ea5-398a-4974-b8f6-62290e519794] return self.greenlet.switch() [ 534.308613] env[61923]: ERROR nova.compute.manager [instance: 97148ea5-398a-4974-b8f6-62290e519794] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 534.308613] env[61923]: ERROR nova.compute.manager [instance: 97148ea5-398a-4974-b8f6-62290e519794] result = function(*args, **kwargs) [ 534.308613] env[61923]: ERROR nova.compute.manager [instance: 97148ea5-398a-4974-b8f6-62290e519794] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 534.308613] env[61923]: ERROR nova.compute.manager [instance: 97148ea5-398a-4974-b8f6-62290e519794] return func(*args, **kwargs) [ 534.308613] env[61923]: ERROR nova.compute.manager [instance: 97148ea5-398a-4974-b8f6-62290e519794] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 534.308613] env[61923]: ERROR nova.compute.manager [instance: 97148ea5-398a-4974-b8f6-62290e519794] raise e [ 534.308613] env[61923]: ERROR nova.compute.manager [instance: 97148ea5-398a-4974-b8f6-62290e519794] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 534.308613] env[61923]: ERROR nova.compute.manager [instance: 97148ea5-398a-4974-b8f6-62290e519794] nwinfo = self.network_api.allocate_for_instance( [ 534.308613] env[61923]: ERROR nova.compute.manager [instance: 97148ea5-398a-4974-b8f6-62290e519794] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 534.308613] env[61923]: ERROR nova.compute.manager [instance: 97148ea5-398a-4974-b8f6-62290e519794] created_port_ids = self._update_ports_for_instance( [ 534.308925] env[61923]: ERROR nova.compute.manager [instance: 97148ea5-398a-4974-b8f6-62290e519794] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 534.308925] env[61923]: ERROR nova.compute.manager [instance: 97148ea5-398a-4974-b8f6-62290e519794] with excutils.save_and_reraise_exception(): [ 534.308925] env[61923]: ERROR nova.compute.manager [instance: 97148ea5-398a-4974-b8f6-62290e519794] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 534.308925] env[61923]: ERROR nova.compute.manager [instance: 97148ea5-398a-4974-b8f6-62290e519794] self.force_reraise() [ 534.308925] env[61923]: ERROR nova.compute.manager [instance: 97148ea5-398a-4974-b8f6-62290e519794] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 534.308925] env[61923]: ERROR nova.compute.manager [instance: 97148ea5-398a-4974-b8f6-62290e519794] raise self.value [ 534.308925] env[61923]: ERROR nova.compute.manager [instance: 97148ea5-398a-4974-b8f6-62290e519794] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 534.308925] env[61923]: ERROR nova.compute.manager [instance: 97148ea5-398a-4974-b8f6-62290e519794] updated_port = self._update_port( [ 534.308925] env[61923]: ERROR nova.compute.manager [instance: 97148ea5-398a-4974-b8f6-62290e519794] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 534.308925] env[61923]: ERROR nova.compute.manager [instance: 97148ea5-398a-4974-b8f6-62290e519794] _ensure_no_port_binding_failure(port) [ 534.308925] env[61923]: ERROR nova.compute.manager [instance: 97148ea5-398a-4974-b8f6-62290e519794] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 534.308925] env[61923]: ERROR nova.compute.manager [instance: 97148ea5-398a-4974-b8f6-62290e519794] raise exception.PortBindingFailed(port_id=port['id']) [ 534.309233] env[61923]: ERROR nova.compute.manager [instance: 97148ea5-398a-4974-b8f6-62290e519794] nova.exception.PortBindingFailed: Binding failed for port 7ec53f9e-f055-4a57-8228-7efbcdb7360f, please check neutron logs for more information. [ 534.309233] env[61923]: ERROR nova.compute.manager [instance: 97148ea5-398a-4974-b8f6-62290e519794] [ 534.309233] env[61923]: INFO nova.compute.manager [None req-7781f812-6fae-4bd6-9d91-ec6098c06372 tempest-InstanceActionsTestJSON-2015629344 tempest-InstanceActionsTestJSON-2015629344-project-member] [instance: 97148ea5-398a-4974-b8f6-62290e519794] Terminating instance [ 534.311219] env[61923]: DEBUG oslo_concurrency.lockutils [None req-7781f812-6fae-4bd6-9d91-ec6098c06372 tempest-InstanceActionsTestJSON-2015629344 tempest-InstanceActionsTestJSON-2015629344-project-member] Acquiring lock "refresh_cache-97148ea5-398a-4974-b8f6-62290e519794" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 534.311597] env[61923]: DEBUG oslo_concurrency.lockutils [None req-7781f812-6fae-4bd6-9d91-ec6098c06372 tempest-InstanceActionsTestJSON-2015629344 tempest-InstanceActionsTestJSON-2015629344-project-member] Acquired lock "refresh_cache-97148ea5-398a-4974-b8f6-62290e519794" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 534.311815] env[61923]: DEBUG nova.network.neutron [None req-7781f812-6fae-4bd6-9d91-ec6098c06372 tempest-InstanceActionsTestJSON-2015629344 tempest-InstanceActionsTestJSON-2015629344-project-member] [instance: 97148ea5-398a-4974-b8f6-62290e519794] Building network info cache for instance {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 534.367788] env[61923]: DEBUG nova.network.neutron [req-f4863623-308d-474c-9e99-25719cffeb5c req-d8001bb8-760a-4b78-918d-8fd6cbdd3480 service nova] [instance: b0a45314-eed8-4d0a-bea2-d4b47dc0bc80] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 534.435431] env[61923]: DEBUG nova.network.neutron [-] [instance: b0a45314-eed8-4d0a-bea2-d4b47dc0bc80] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 534.561560] env[61923]: DEBUG nova.network.neutron [req-f4863623-308d-474c-9e99-25719cffeb5c req-d8001bb8-760a-4b78-918d-8fd6cbdd3480 service nova] [instance: b0a45314-eed8-4d0a-bea2-d4b47dc0bc80] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 534.568788] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1da60c58-ab2a-48fa-b1e0-a669ee8f1164 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 534.581559] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8fbdecf9-89ba-4170-81e6-0d880e2ea73f {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 534.633059] env[61923]: DEBUG oslo_concurrency.lockutils [None req-a2070053-8e36-4abe-a723-0b93480b3283 tempest-ServerDiagnosticsTest-1095156200 tempest-ServerDiagnosticsTest-1095156200-project-member] Releasing lock "refresh_cache-a134ab6d-b261-4df6-b440-d32a3e578645" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 534.635815] env[61923]: DEBUG nova.compute.manager [None req-a2070053-8e36-4abe-a723-0b93480b3283 tempest-ServerDiagnosticsTest-1095156200 tempest-ServerDiagnosticsTest-1095156200-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61923) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 534.636076] env[61923]: DEBUG nova.compute.manager [None req-a2070053-8e36-4abe-a723-0b93480b3283 tempest-ServerDiagnosticsTest-1095156200 tempest-ServerDiagnosticsTest-1095156200-project-member] [instance: a134ab6d-b261-4df6-b440-d32a3e578645] Deallocating network for instance {{(pid=61923) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 534.636270] env[61923]: DEBUG nova.network.neutron [None req-a2070053-8e36-4abe-a723-0b93480b3283 tempest-ServerDiagnosticsTest-1095156200 tempest-ServerDiagnosticsTest-1095156200-project-member] [instance: a134ab6d-b261-4df6-b440-d32a3e578645] deallocate_for_instance() {{(pid=61923) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 534.638988] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48262339-87f7-4570-9bd1-3e36f459aa90 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 534.649022] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6a80090-9464-4564-9c30-682fdafbb99f {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 534.672092] env[61923]: DEBUG nova.compute.provider_tree [None req-f30f4038-8fd0-4bbd-8b35-7fb8e0dea3a4 tempest-SecurityGroupsTestJSON-509490211 tempest-SecurityGroupsTestJSON-509490211-project-member] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 534.675689] env[61923]: DEBUG nova.network.neutron [None req-a2070053-8e36-4abe-a723-0b93480b3283 tempest-ServerDiagnosticsTest-1095156200 tempest-ServerDiagnosticsTest-1095156200-project-member] [instance: a134ab6d-b261-4df6-b440-d32a3e578645] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 534.791046] env[61923]: DEBUG nova.compute.manager [req-76ee35b9-46da-4193-a86e-02ac4b60c220 req-d35462f9-f5ce-4d33-afcf-a9ac08040144 service nova] [instance: 97148ea5-398a-4974-b8f6-62290e519794] Received event network-changed-7ec53f9e-f055-4a57-8228-7efbcdb7360f {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 534.791046] env[61923]: DEBUG nova.compute.manager [req-76ee35b9-46da-4193-a86e-02ac4b60c220 req-d35462f9-f5ce-4d33-afcf-a9ac08040144 service nova] [instance: 97148ea5-398a-4974-b8f6-62290e519794] Refreshing instance network info cache due to event network-changed-7ec53f9e-f055-4a57-8228-7efbcdb7360f. {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 534.791046] env[61923]: DEBUG oslo_concurrency.lockutils [req-76ee35b9-46da-4193-a86e-02ac4b60c220 req-d35462f9-f5ce-4d33-afcf-a9ac08040144 service nova] Acquiring lock "refresh_cache-97148ea5-398a-4974-b8f6-62290e519794" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 534.941066] env[61923]: INFO nova.compute.manager [-] [instance: b0a45314-eed8-4d0a-bea2-d4b47dc0bc80] Took 1.08 seconds to deallocate network for instance. [ 534.945206] env[61923]: DEBUG nova.compute.claims [None req-8a6b6bc2-54c9-4389-ab30-057158d86227 tempest-ServerDiagnosticsNegativeTest-664316905 tempest-ServerDiagnosticsNegativeTest-664316905-project-member] [instance: b0a45314-eed8-4d0a-bea2-d4b47dc0bc80] Aborting claim: {{(pid=61923) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 534.945388] env[61923]: DEBUG oslo_concurrency.lockutils [None req-8a6b6bc2-54c9-4389-ab30-057158d86227 tempest-ServerDiagnosticsNegativeTest-664316905 tempest-ServerDiagnosticsNegativeTest-664316905-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 534.999251] env[61923]: DEBUG nova.network.neutron [None req-7781f812-6fae-4bd6-9d91-ec6098c06372 tempest-InstanceActionsTestJSON-2015629344 tempest-InstanceActionsTestJSON-2015629344-project-member] [instance: 97148ea5-398a-4974-b8f6-62290e519794] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 535.066229] env[61923]: DEBUG oslo_concurrency.lockutils [req-f4863623-308d-474c-9e99-25719cffeb5c req-d8001bb8-760a-4b78-918d-8fd6cbdd3480 service nova] Releasing lock "refresh_cache-b0a45314-eed8-4d0a-bea2-d4b47dc0bc80" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 535.176277] env[61923]: DEBUG nova.scheduler.client.report [None req-f30f4038-8fd0-4bbd-8b35-7fb8e0dea3a4 tempest-SecurityGroupsTestJSON-509490211 tempest-SecurityGroupsTestJSON-509490211-project-member] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 535.181126] env[61923]: DEBUG nova.network.neutron [None req-a2070053-8e36-4abe-a723-0b93480b3283 tempest-ServerDiagnosticsTest-1095156200 tempest-ServerDiagnosticsTest-1095156200-project-member] [instance: a134ab6d-b261-4df6-b440-d32a3e578645] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 535.321740] env[61923]: DEBUG nova.network.neutron [None req-7781f812-6fae-4bd6-9d91-ec6098c06372 tempest-InstanceActionsTestJSON-2015629344 tempest-InstanceActionsTestJSON-2015629344-project-member] [instance: 97148ea5-398a-4974-b8f6-62290e519794] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 535.685298] env[61923]: DEBUG oslo_concurrency.lockutils [None req-f30f4038-8fd0-4bbd-8b35-7fb8e0dea3a4 tempest-SecurityGroupsTestJSON-509490211 tempest-SecurityGroupsTestJSON-509490211-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.347s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 535.685808] env[61923]: DEBUG nova.compute.manager [None req-f30f4038-8fd0-4bbd-8b35-7fb8e0dea3a4 tempest-SecurityGroupsTestJSON-509490211 tempest-SecurityGroupsTestJSON-509490211-project-member] [instance: 59fc11e8-27be-4553-9de9-3eb5ea612ec4] Start building networks asynchronously for instance. {{(pid=61923) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 535.690922] env[61923]: DEBUG oslo_concurrency.lockutils [None req-35a5fc5c-7e44-4806-b390-80f5a8a499d8 tempest-DeleteServersAdminTestJSON-590116201 tempest-DeleteServersAdminTestJSON-590116201-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 9.927s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 535.693417] env[61923]: INFO nova.compute.manager [None req-a2070053-8e36-4abe-a723-0b93480b3283 tempest-ServerDiagnosticsTest-1095156200 tempest-ServerDiagnosticsTest-1095156200-project-member] [instance: a134ab6d-b261-4df6-b440-d32a3e578645] Took 1.06 seconds to deallocate network for instance. [ 535.826736] env[61923]: DEBUG oslo_concurrency.lockutils [None req-7781f812-6fae-4bd6-9d91-ec6098c06372 tempest-InstanceActionsTestJSON-2015629344 tempest-InstanceActionsTestJSON-2015629344-project-member] Releasing lock "refresh_cache-97148ea5-398a-4974-b8f6-62290e519794" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 535.827234] env[61923]: DEBUG nova.compute.manager [None req-7781f812-6fae-4bd6-9d91-ec6098c06372 tempest-InstanceActionsTestJSON-2015629344 tempest-InstanceActionsTestJSON-2015629344-project-member] [instance: 97148ea5-398a-4974-b8f6-62290e519794] Start destroying the instance on the hypervisor. {{(pid=61923) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 535.827435] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-7781f812-6fae-4bd6-9d91-ec6098c06372 tempest-InstanceActionsTestJSON-2015629344 tempest-InstanceActionsTestJSON-2015629344-project-member] [instance: 97148ea5-398a-4974-b8f6-62290e519794] Destroying instance {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 535.827747] env[61923]: DEBUG oslo_concurrency.lockutils [req-76ee35b9-46da-4193-a86e-02ac4b60c220 req-d35462f9-f5ce-4d33-afcf-a9ac08040144 service nova] Acquired lock "refresh_cache-97148ea5-398a-4974-b8f6-62290e519794" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 535.827913] env[61923]: DEBUG nova.network.neutron [req-76ee35b9-46da-4193-a86e-02ac4b60c220 req-d35462f9-f5ce-4d33-afcf-a9ac08040144 service nova] [instance: 97148ea5-398a-4974-b8f6-62290e519794] Refreshing network info cache for port 7ec53f9e-f055-4a57-8228-7efbcdb7360f {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 535.830455] env[61923]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-b0aa5dd3-a56e-4557-a8a5-54381c581007 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 535.839966] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-433f871d-ca3c-44a9-9ee7-7e3281822fc1 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 535.869732] env[61923]: WARNING nova.virt.vmwareapi.vmops [None req-7781f812-6fae-4bd6-9d91-ec6098c06372 tempest-InstanceActionsTestJSON-2015629344 tempest-InstanceActionsTestJSON-2015629344-project-member] [instance: 97148ea5-398a-4974-b8f6-62290e519794] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 97148ea5-398a-4974-b8f6-62290e519794 could not be found. [ 535.869967] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-7781f812-6fae-4bd6-9d91-ec6098c06372 tempest-InstanceActionsTestJSON-2015629344 tempest-InstanceActionsTestJSON-2015629344-project-member] [instance: 97148ea5-398a-4974-b8f6-62290e519794] Instance destroyed {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 535.870258] env[61923]: INFO nova.compute.manager [None req-7781f812-6fae-4bd6-9d91-ec6098c06372 tempest-InstanceActionsTestJSON-2015629344 tempest-InstanceActionsTestJSON-2015629344-project-member] [instance: 97148ea5-398a-4974-b8f6-62290e519794] Took 0.04 seconds to destroy the instance on the hypervisor. [ 535.870408] env[61923]: DEBUG oslo.service.loopingcall [None req-7781f812-6fae-4bd6-9d91-ec6098c06372 tempest-InstanceActionsTestJSON-2015629344 tempest-InstanceActionsTestJSON-2015629344-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61923) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 535.870606] env[61923]: DEBUG nova.compute.manager [-] [instance: 97148ea5-398a-4974-b8f6-62290e519794] Deallocating network for instance {{(pid=61923) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 535.870696] env[61923]: DEBUG nova.network.neutron [-] [instance: 97148ea5-398a-4974-b8f6-62290e519794] deallocate_for_instance() {{(pid=61923) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 535.915227] env[61923]: DEBUG nova.network.neutron [-] [instance: 97148ea5-398a-4974-b8f6-62290e519794] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 536.066572] env[61923]: DEBUG oslo_concurrency.lockutils [None req-15b9e1af-fe57-407e-a058-c08239279239 tempest-ServersWithSpecificFlavorTestJSON-1806421899 tempest-ServersWithSpecificFlavorTestJSON-1806421899-project-member] Acquiring lock "4cec7bce-0136-43ad-898f-086058a505f2" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 536.066853] env[61923]: DEBUG oslo_concurrency.lockutils [None req-15b9e1af-fe57-407e-a058-c08239279239 tempest-ServersWithSpecificFlavorTestJSON-1806421899 tempest-ServersWithSpecificFlavorTestJSON-1806421899-project-member] Lock "4cec7bce-0136-43ad-898f-086058a505f2" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.002s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 536.192587] env[61923]: DEBUG nova.compute.utils [None req-f30f4038-8fd0-4bbd-8b35-7fb8e0dea3a4 tempest-SecurityGroupsTestJSON-509490211 tempest-SecurityGroupsTestJSON-509490211-project-member] Using /dev/sd instead of None {{(pid=61923) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 536.196339] env[61923]: DEBUG nova.compute.manager [None req-f30f4038-8fd0-4bbd-8b35-7fb8e0dea3a4 tempest-SecurityGroupsTestJSON-509490211 tempest-SecurityGroupsTestJSON-509490211-project-member] [instance: 59fc11e8-27be-4553-9de9-3eb5ea612ec4] Allocating IP information in the background. {{(pid=61923) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 536.196525] env[61923]: DEBUG nova.network.neutron [None req-f30f4038-8fd0-4bbd-8b35-7fb8e0dea3a4 tempest-SecurityGroupsTestJSON-509490211 tempest-SecurityGroupsTestJSON-509490211-project-member] [instance: 59fc11e8-27be-4553-9de9-3eb5ea612ec4] allocate_for_instance() {{(pid=61923) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 536.393370] env[61923]: DEBUG nova.policy [None req-f30f4038-8fd0-4bbd-8b35-7fb8e0dea3a4 tempest-SecurityGroupsTestJSON-509490211 tempest-SecurityGroupsTestJSON-509490211-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0237dcc9ff3e4b5e919f8c921d16268e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ebd1ac0807c34d0597971c074f4b6853', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61923) authorize /opt/stack/nova/nova/policy.py:201}} [ 536.396727] env[61923]: DEBUG nova.network.neutron [req-76ee35b9-46da-4193-a86e-02ac4b60c220 req-d35462f9-f5ce-4d33-afcf-a9ac08040144 service nova] [instance: 97148ea5-398a-4974-b8f6-62290e519794] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 536.409392] env[61923]: DEBUG nova.compute.manager [req-eca52c0f-a4c1-4d96-aaa2-2499cb9f3857 req-99c8ec0a-65ec-438e-898d-afa7646b07a5 service nova] [instance: b0a45314-eed8-4d0a-bea2-d4b47dc0bc80] Received event network-vif-deleted-6bf512ed-8713-43d4-bc7e-ce2b9a514f52 {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 536.417153] env[61923]: DEBUG nova.network.neutron [-] [instance: 97148ea5-398a-4974-b8f6-62290e519794] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 536.438389] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-188cd127-88d2-438c-bb0f-805a544de6ac {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 536.449684] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e19a441-d6b8-4f60-ab43-495e02df23bb {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 536.481537] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd523e4e-60ee-4584-afdc-48edf2c40225 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 536.490956] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ba3aa33-b898-49b9-bfc8-c5ccc161d141 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 536.510024] env[61923]: DEBUG nova.compute.provider_tree [None req-35a5fc5c-7e44-4806-b390-80f5a8a499d8 tempest-DeleteServersAdminTestJSON-590116201 tempest-DeleteServersAdminTestJSON-590116201-project-member] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 536.702879] env[61923]: DEBUG nova.compute.manager [None req-f30f4038-8fd0-4bbd-8b35-7fb8e0dea3a4 tempest-SecurityGroupsTestJSON-509490211 tempest-SecurityGroupsTestJSON-509490211-project-member] [instance: 59fc11e8-27be-4553-9de9-3eb5ea612ec4] Start building block device mappings for instance. {{(pid=61923) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 536.745389] env[61923]: INFO nova.scheduler.client.report [None req-a2070053-8e36-4abe-a723-0b93480b3283 tempest-ServerDiagnosticsTest-1095156200 tempest-ServerDiagnosticsTest-1095156200-project-member] Deleted allocations for instance a134ab6d-b261-4df6-b440-d32a3e578645 [ 536.921114] env[61923]: INFO nova.compute.manager [-] [instance: 97148ea5-398a-4974-b8f6-62290e519794] Took 1.05 seconds to deallocate network for instance. [ 536.927020] env[61923]: DEBUG nova.compute.claims [None req-7781f812-6fae-4bd6-9d91-ec6098c06372 tempest-InstanceActionsTestJSON-2015629344 tempest-InstanceActionsTestJSON-2015629344-project-member] [instance: 97148ea5-398a-4974-b8f6-62290e519794] Aborting claim: {{(pid=61923) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 536.927020] env[61923]: DEBUG oslo_concurrency.lockutils [None req-7781f812-6fae-4bd6-9d91-ec6098c06372 tempest-InstanceActionsTestJSON-2015629344 tempest-InstanceActionsTestJSON-2015629344-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 536.933969] env[61923]: ERROR nova.compute.manager [None req-844a2215-1038-4a30-8a79-40c1c3117bde tempest-FloatingIPsAssociationNegativeTestJSON-1301210771 tempest-FloatingIPsAssociationNegativeTestJSON-1301210771-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 4f3fea2e-b3d9-4c89-b291-071aa6707d27, please check neutron logs for more information. [ 536.933969] env[61923]: ERROR nova.compute.manager Traceback (most recent call last): [ 536.933969] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 536.933969] env[61923]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 536.933969] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 536.933969] env[61923]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 536.933969] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 536.933969] env[61923]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 536.933969] env[61923]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 536.933969] env[61923]: ERROR nova.compute.manager self.force_reraise() [ 536.933969] env[61923]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 536.933969] env[61923]: ERROR nova.compute.manager raise self.value [ 536.933969] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 536.933969] env[61923]: ERROR nova.compute.manager updated_port = self._update_port( [ 536.933969] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 536.933969] env[61923]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 536.934500] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 536.934500] env[61923]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 536.934500] env[61923]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 4f3fea2e-b3d9-4c89-b291-071aa6707d27, please check neutron logs for more information. [ 536.934500] env[61923]: ERROR nova.compute.manager [ 536.934500] env[61923]: Traceback (most recent call last): [ 536.934500] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 536.934500] env[61923]: listener.cb(fileno) [ 536.934500] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 536.934500] env[61923]: result = function(*args, **kwargs) [ 536.934500] env[61923]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 536.934500] env[61923]: return func(*args, **kwargs) [ 536.934500] env[61923]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 536.934500] env[61923]: raise e [ 536.934500] env[61923]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 536.934500] env[61923]: nwinfo = self.network_api.allocate_for_instance( [ 536.934500] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 536.934500] env[61923]: created_port_ids = self._update_ports_for_instance( [ 536.934500] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 536.934500] env[61923]: with excutils.save_and_reraise_exception(): [ 536.934500] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 536.934500] env[61923]: self.force_reraise() [ 536.934500] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 536.934500] env[61923]: raise self.value [ 536.934500] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 536.934500] env[61923]: updated_port = self._update_port( [ 536.934500] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 536.934500] env[61923]: _ensure_no_port_binding_failure(port) [ 536.934500] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 536.934500] env[61923]: raise exception.PortBindingFailed(port_id=port['id']) [ 536.935334] env[61923]: nova.exception.PortBindingFailed: Binding failed for port 4f3fea2e-b3d9-4c89-b291-071aa6707d27, please check neutron logs for more information. [ 536.935334] env[61923]: Removing descriptor: 17 [ 536.935334] env[61923]: ERROR nova.compute.manager [None req-844a2215-1038-4a30-8a79-40c1c3117bde tempest-FloatingIPsAssociationNegativeTestJSON-1301210771 tempest-FloatingIPsAssociationNegativeTestJSON-1301210771-project-member] [instance: 5d57303a-ae17-4b92-8f12-303677eb0d7e] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 4f3fea2e-b3d9-4c89-b291-071aa6707d27, please check neutron logs for more information. [ 536.935334] env[61923]: ERROR nova.compute.manager [instance: 5d57303a-ae17-4b92-8f12-303677eb0d7e] Traceback (most recent call last): [ 536.935334] env[61923]: ERROR nova.compute.manager [instance: 5d57303a-ae17-4b92-8f12-303677eb0d7e] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 536.935334] env[61923]: ERROR nova.compute.manager [instance: 5d57303a-ae17-4b92-8f12-303677eb0d7e] yield resources [ 536.935334] env[61923]: ERROR nova.compute.manager [instance: 5d57303a-ae17-4b92-8f12-303677eb0d7e] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 536.935334] env[61923]: ERROR nova.compute.manager [instance: 5d57303a-ae17-4b92-8f12-303677eb0d7e] self.driver.spawn(context, instance, image_meta, [ 536.935334] env[61923]: ERROR nova.compute.manager [instance: 5d57303a-ae17-4b92-8f12-303677eb0d7e] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 536.935334] env[61923]: ERROR nova.compute.manager [instance: 5d57303a-ae17-4b92-8f12-303677eb0d7e] self._vmops.spawn(context, instance, image_meta, injected_files, [ 536.935334] env[61923]: ERROR nova.compute.manager [instance: 5d57303a-ae17-4b92-8f12-303677eb0d7e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 536.935334] env[61923]: ERROR nova.compute.manager [instance: 5d57303a-ae17-4b92-8f12-303677eb0d7e] vm_ref = self.build_virtual_machine(instance, [ 536.935677] env[61923]: ERROR nova.compute.manager [instance: 5d57303a-ae17-4b92-8f12-303677eb0d7e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 536.935677] env[61923]: ERROR nova.compute.manager [instance: 5d57303a-ae17-4b92-8f12-303677eb0d7e] vif_infos = vmwarevif.get_vif_info(self._session, [ 536.935677] env[61923]: ERROR nova.compute.manager [instance: 5d57303a-ae17-4b92-8f12-303677eb0d7e] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 536.935677] env[61923]: ERROR nova.compute.manager [instance: 5d57303a-ae17-4b92-8f12-303677eb0d7e] for vif in network_info: [ 536.935677] env[61923]: ERROR nova.compute.manager [instance: 5d57303a-ae17-4b92-8f12-303677eb0d7e] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 536.935677] env[61923]: ERROR nova.compute.manager [instance: 5d57303a-ae17-4b92-8f12-303677eb0d7e] return self._sync_wrapper(fn, *args, **kwargs) [ 536.935677] env[61923]: ERROR nova.compute.manager [instance: 5d57303a-ae17-4b92-8f12-303677eb0d7e] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 536.935677] env[61923]: ERROR nova.compute.manager [instance: 5d57303a-ae17-4b92-8f12-303677eb0d7e] self.wait() [ 536.935677] env[61923]: ERROR nova.compute.manager [instance: 5d57303a-ae17-4b92-8f12-303677eb0d7e] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 536.935677] env[61923]: ERROR nova.compute.manager [instance: 5d57303a-ae17-4b92-8f12-303677eb0d7e] self[:] = self._gt.wait() [ 536.935677] env[61923]: ERROR nova.compute.manager [instance: 5d57303a-ae17-4b92-8f12-303677eb0d7e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 536.935677] env[61923]: ERROR nova.compute.manager [instance: 5d57303a-ae17-4b92-8f12-303677eb0d7e] return self._exit_event.wait() [ 536.935677] env[61923]: ERROR nova.compute.manager [instance: 5d57303a-ae17-4b92-8f12-303677eb0d7e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 536.937220] env[61923]: ERROR nova.compute.manager [instance: 5d57303a-ae17-4b92-8f12-303677eb0d7e] result = hub.switch() [ 536.937220] env[61923]: ERROR nova.compute.manager [instance: 5d57303a-ae17-4b92-8f12-303677eb0d7e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 536.937220] env[61923]: ERROR nova.compute.manager [instance: 5d57303a-ae17-4b92-8f12-303677eb0d7e] return self.greenlet.switch() [ 536.937220] env[61923]: ERROR nova.compute.manager [instance: 5d57303a-ae17-4b92-8f12-303677eb0d7e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 536.937220] env[61923]: ERROR nova.compute.manager [instance: 5d57303a-ae17-4b92-8f12-303677eb0d7e] result = function(*args, **kwargs) [ 536.937220] env[61923]: ERROR nova.compute.manager [instance: 5d57303a-ae17-4b92-8f12-303677eb0d7e] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 536.937220] env[61923]: ERROR nova.compute.manager [instance: 5d57303a-ae17-4b92-8f12-303677eb0d7e] return func(*args, **kwargs) [ 536.937220] env[61923]: ERROR nova.compute.manager [instance: 5d57303a-ae17-4b92-8f12-303677eb0d7e] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 536.937220] env[61923]: ERROR nova.compute.manager [instance: 5d57303a-ae17-4b92-8f12-303677eb0d7e] raise e [ 536.937220] env[61923]: ERROR nova.compute.manager [instance: 5d57303a-ae17-4b92-8f12-303677eb0d7e] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 536.937220] env[61923]: ERROR nova.compute.manager [instance: 5d57303a-ae17-4b92-8f12-303677eb0d7e] nwinfo = self.network_api.allocate_for_instance( [ 536.937220] env[61923]: ERROR nova.compute.manager [instance: 5d57303a-ae17-4b92-8f12-303677eb0d7e] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 536.937220] env[61923]: ERROR nova.compute.manager [instance: 5d57303a-ae17-4b92-8f12-303677eb0d7e] created_port_ids = self._update_ports_for_instance( [ 536.937729] env[61923]: ERROR nova.compute.manager [instance: 5d57303a-ae17-4b92-8f12-303677eb0d7e] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 536.937729] env[61923]: ERROR nova.compute.manager [instance: 5d57303a-ae17-4b92-8f12-303677eb0d7e] with excutils.save_and_reraise_exception(): [ 536.937729] env[61923]: ERROR nova.compute.manager [instance: 5d57303a-ae17-4b92-8f12-303677eb0d7e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 536.937729] env[61923]: ERROR nova.compute.manager [instance: 5d57303a-ae17-4b92-8f12-303677eb0d7e] self.force_reraise() [ 536.937729] env[61923]: ERROR nova.compute.manager [instance: 5d57303a-ae17-4b92-8f12-303677eb0d7e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 536.937729] env[61923]: ERROR nova.compute.manager [instance: 5d57303a-ae17-4b92-8f12-303677eb0d7e] raise self.value [ 536.937729] env[61923]: ERROR nova.compute.manager [instance: 5d57303a-ae17-4b92-8f12-303677eb0d7e] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 536.937729] env[61923]: ERROR nova.compute.manager [instance: 5d57303a-ae17-4b92-8f12-303677eb0d7e] updated_port = self._update_port( [ 536.937729] env[61923]: ERROR nova.compute.manager [instance: 5d57303a-ae17-4b92-8f12-303677eb0d7e] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 536.937729] env[61923]: ERROR nova.compute.manager [instance: 5d57303a-ae17-4b92-8f12-303677eb0d7e] _ensure_no_port_binding_failure(port) [ 536.937729] env[61923]: ERROR nova.compute.manager [instance: 5d57303a-ae17-4b92-8f12-303677eb0d7e] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 536.937729] env[61923]: ERROR nova.compute.manager [instance: 5d57303a-ae17-4b92-8f12-303677eb0d7e] raise exception.PortBindingFailed(port_id=port['id']) [ 536.938059] env[61923]: ERROR nova.compute.manager [instance: 5d57303a-ae17-4b92-8f12-303677eb0d7e] nova.exception.PortBindingFailed: Binding failed for port 4f3fea2e-b3d9-4c89-b291-071aa6707d27, please check neutron logs for more information. [ 536.938059] env[61923]: ERROR nova.compute.manager [instance: 5d57303a-ae17-4b92-8f12-303677eb0d7e] [ 536.938059] env[61923]: INFO nova.compute.manager [None req-844a2215-1038-4a30-8a79-40c1c3117bde tempest-FloatingIPsAssociationNegativeTestJSON-1301210771 tempest-FloatingIPsAssociationNegativeTestJSON-1301210771-project-member] [instance: 5d57303a-ae17-4b92-8f12-303677eb0d7e] Terminating instance [ 536.940913] env[61923]: DEBUG oslo_concurrency.lockutils [None req-844a2215-1038-4a30-8a79-40c1c3117bde tempest-FloatingIPsAssociationNegativeTestJSON-1301210771 tempest-FloatingIPsAssociationNegativeTestJSON-1301210771-project-member] Acquiring lock "refresh_cache-5d57303a-ae17-4b92-8f12-303677eb0d7e" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 536.940913] env[61923]: DEBUG oslo_concurrency.lockutils [None req-844a2215-1038-4a30-8a79-40c1c3117bde tempest-FloatingIPsAssociationNegativeTestJSON-1301210771 tempest-FloatingIPsAssociationNegativeTestJSON-1301210771-project-member] Acquired lock "refresh_cache-5d57303a-ae17-4b92-8f12-303677eb0d7e" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 536.940913] env[61923]: DEBUG nova.network.neutron [None req-844a2215-1038-4a30-8a79-40c1c3117bde tempest-FloatingIPsAssociationNegativeTestJSON-1301210771 tempest-FloatingIPsAssociationNegativeTestJSON-1301210771-project-member] [instance: 5d57303a-ae17-4b92-8f12-303677eb0d7e] Building network info cache for instance {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 537.011232] env[61923]: DEBUG nova.scheduler.client.report [None req-35a5fc5c-7e44-4806-b390-80f5a8a499d8 tempest-DeleteServersAdminTestJSON-590116201 tempest-DeleteServersAdminTestJSON-590116201-project-member] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 537.072897] env[61923]: DEBUG nova.network.neutron [req-76ee35b9-46da-4193-a86e-02ac4b60c220 req-d35462f9-f5ce-4d33-afcf-a9ac08040144 service nova] [instance: 97148ea5-398a-4974-b8f6-62290e519794] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 537.119724] env[61923]: DEBUG nova.network.neutron [None req-f30f4038-8fd0-4bbd-8b35-7fb8e0dea3a4 tempest-SecurityGroupsTestJSON-509490211 tempest-SecurityGroupsTestJSON-509490211-project-member] [instance: 59fc11e8-27be-4553-9de9-3eb5ea612ec4] Successfully created port: 238265b6-446c-493c-b510-5f51005c1787 {{(pid=61923) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 537.258934] env[61923]: DEBUG oslo_concurrency.lockutils [None req-a2070053-8e36-4abe-a723-0b93480b3283 tempest-ServerDiagnosticsTest-1095156200 tempest-ServerDiagnosticsTest-1095156200-project-member] Lock "a134ab6d-b261-4df6-b440-d32a3e578645" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 23.818s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 537.496994] env[61923]: DEBUG nova.network.neutron [None req-844a2215-1038-4a30-8a79-40c1c3117bde tempest-FloatingIPsAssociationNegativeTestJSON-1301210771 tempest-FloatingIPsAssociationNegativeTestJSON-1301210771-project-member] [instance: 5d57303a-ae17-4b92-8f12-303677eb0d7e] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 537.525407] env[61923]: DEBUG oslo_concurrency.lockutils [None req-35a5fc5c-7e44-4806-b390-80f5a8a499d8 tempest-DeleteServersAdminTestJSON-590116201 tempest-DeleteServersAdminTestJSON-590116201-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.835s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 537.525407] env[61923]: ERROR nova.compute.manager [None req-35a5fc5c-7e44-4806-b390-80f5a8a499d8 tempest-DeleteServersAdminTestJSON-590116201 tempest-DeleteServersAdminTestJSON-590116201-project-member] [instance: 9ea2600a-d439-4fc3-a131-9bb93250d8cd] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port a1ecfb50-9098-4408-b9bd-8db498ae622e, please check neutron logs for more information. [ 537.525407] env[61923]: ERROR nova.compute.manager [instance: 9ea2600a-d439-4fc3-a131-9bb93250d8cd] Traceback (most recent call last): [ 537.525407] env[61923]: ERROR nova.compute.manager [instance: 9ea2600a-d439-4fc3-a131-9bb93250d8cd] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 537.525407] env[61923]: ERROR nova.compute.manager [instance: 9ea2600a-d439-4fc3-a131-9bb93250d8cd] self.driver.spawn(context, instance, image_meta, [ 537.525407] env[61923]: ERROR nova.compute.manager [instance: 9ea2600a-d439-4fc3-a131-9bb93250d8cd] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 537.525407] env[61923]: ERROR nova.compute.manager [instance: 9ea2600a-d439-4fc3-a131-9bb93250d8cd] self._vmops.spawn(context, instance, image_meta, injected_files, [ 537.525407] env[61923]: ERROR nova.compute.manager [instance: 9ea2600a-d439-4fc3-a131-9bb93250d8cd] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 537.525407] env[61923]: ERROR nova.compute.manager [instance: 9ea2600a-d439-4fc3-a131-9bb93250d8cd] vm_ref = self.build_virtual_machine(instance, [ 537.525656] env[61923]: ERROR nova.compute.manager [instance: 9ea2600a-d439-4fc3-a131-9bb93250d8cd] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 537.525656] env[61923]: ERROR nova.compute.manager [instance: 9ea2600a-d439-4fc3-a131-9bb93250d8cd] vif_infos = vmwarevif.get_vif_info(self._session, [ 537.525656] env[61923]: ERROR nova.compute.manager [instance: 9ea2600a-d439-4fc3-a131-9bb93250d8cd] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 537.525656] env[61923]: ERROR nova.compute.manager [instance: 9ea2600a-d439-4fc3-a131-9bb93250d8cd] for vif in network_info: [ 537.525656] env[61923]: ERROR nova.compute.manager [instance: 9ea2600a-d439-4fc3-a131-9bb93250d8cd] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 537.525656] env[61923]: ERROR nova.compute.manager [instance: 9ea2600a-d439-4fc3-a131-9bb93250d8cd] return self._sync_wrapper(fn, *args, **kwargs) [ 537.525656] env[61923]: ERROR nova.compute.manager [instance: 9ea2600a-d439-4fc3-a131-9bb93250d8cd] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 537.525656] env[61923]: ERROR nova.compute.manager [instance: 9ea2600a-d439-4fc3-a131-9bb93250d8cd] self.wait() [ 537.525656] env[61923]: ERROR nova.compute.manager [instance: 9ea2600a-d439-4fc3-a131-9bb93250d8cd] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 537.525656] env[61923]: ERROR nova.compute.manager [instance: 9ea2600a-d439-4fc3-a131-9bb93250d8cd] self[:] = self._gt.wait() [ 537.525656] env[61923]: ERROR nova.compute.manager [instance: 9ea2600a-d439-4fc3-a131-9bb93250d8cd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 537.525656] env[61923]: ERROR nova.compute.manager [instance: 9ea2600a-d439-4fc3-a131-9bb93250d8cd] return self._exit_event.wait() [ 537.525656] env[61923]: ERROR nova.compute.manager [instance: 9ea2600a-d439-4fc3-a131-9bb93250d8cd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 537.526860] env[61923]: ERROR nova.compute.manager [instance: 9ea2600a-d439-4fc3-a131-9bb93250d8cd] result = hub.switch() [ 537.526860] env[61923]: ERROR nova.compute.manager [instance: 9ea2600a-d439-4fc3-a131-9bb93250d8cd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 537.526860] env[61923]: ERROR nova.compute.manager [instance: 9ea2600a-d439-4fc3-a131-9bb93250d8cd] return self.greenlet.switch() [ 537.526860] env[61923]: ERROR nova.compute.manager [instance: 9ea2600a-d439-4fc3-a131-9bb93250d8cd] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 537.526860] env[61923]: ERROR nova.compute.manager [instance: 9ea2600a-d439-4fc3-a131-9bb93250d8cd] result = function(*args, **kwargs) [ 537.526860] env[61923]: ERROR nova.compute.manager [instance: 9ea2600a-d439-4fc3-a131-9bb93250d8cd] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 537.526860] env[61923]: ERROR nova.compute.manager [instance: 9ea2600a-d439-4fc3-a131-9bb93250d8cd] return func(*args, **kwargs) [ 537.526860] env[61923]: ERROR nova.compute.manager [instance: 9ea2600a-d439-4fc3-a131-9bb93250d8cd] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 537.526860] env[61923]: ERROR nova.compute.manager [instance: 9ea2600a-d439-4fc3-a131-9bb93250d8cd] raise e [ 537.526860] env[61923]: ERROR nova.compute.manager [instance: 9ea2600a-d439-4fc3-a131-9bb93250d8cd] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 537.526860] env[61923]: ERROR nova.compute.manager [instance: 9ea2600a-d439-4fc3-a131-9bb93250d8cd] nwinfo = self.network_api.allocate_for_instance( [ 537.526860] env[61923]: ERROR nova.compute.manager [instance: 9ea2600a-d439-4fc3-a131-9bb93250d8cd] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 537.526860] env[61923]: ERROR nova.compute.manager [instance: 9ea2600a-d439-4fc3-a131-9bb93250d8cd] created_port_ids = self._update_ports_for_instance( [ 537.527289] env[61923]: ERROR nova.compute.manager [instance: 9ea2600a-d439-4fc3-a131-9bb93250d8cd] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 537.527289] env[61923]: ERROR nova.compute.manager [instance: 9ea2600a-d439-4fc3-a131-9bb93250d8cd] with excutils.save_and_reraise_exception(): [ 537.527289] env[61923]: ERROR nova.compute.manager [instance: 9ea2600a-d439-4fc3-a131-9bb93250d8cd] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 537.527289] env[61923]: ERROR nova.compute.manager [instance: 9ea2600a-d439-4fc3-a131-9bb93250d8cd] self.force_reraise() [ 537.527289] env[61923]: ERROR nova.compute.manager [instance: 9ea2600a-d439-4fc3-a131-9bb93250d8cd] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 537.527289] env[61923]: ERROR nova.compute.manager [instance: 9ea2600a-d439-4fc3-a131-9bb93250d8cd] raise self.value [ 537.527289] env[61923]: ERROR nova.compute.manager [instance: 9ea2600a-d439-4fc3-a131-9bb93250d8cd] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 537.527289] env[61923]: ERROR nova.compute.manager [instance: 9ea2600a-d439-4fc3-a131-9bb93250d8cd] updated_port = self._update_port( [ 537.527289] env[61923]: ERROR nova.compute.manager [instance: 9ea2600a-d439-4fc3-a131-9bb93250d8cd] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 537.527289] env[61923]: ERROR nova.compute.manager [instance: 9ea2600a-d439-4fc3-a131-9bb93250d8cd] _ensure_no_port_binding_failure(port) [ 537.527289] env[61923]: ERROR nova.compute.manager [instance: 9ea2600a-d439-4fc3-a131-9bb93250d8cd] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 537.527289] env[61923]: ERROR nova.compute.manager [instance: 9ea2600a-d439-4fc3-a131-9bb93250d8cd] raise exception.PortBindingFailed(port_id=port['id']) [ 537.527577] env[61923]: ERROR nova.compute.manager [instance: 9ea2600a-d439-4fc3-a131-9bb93250d8cd] nova.exception.PortBindingFailed: Binding failed for port a1ecfb50-9098-4408-b9bd-8db498ae622e, please check neutron logs for more information. [ 537.527577] env[61923]: ERROR nova.compute.manager [instance: 9ea2600a-d439-4fc3-a131-9bb93250d8cd] [ 537.527577] env[61923]: DEBUG nova.compute.utils [None req-35a5fc5c-7e44-4806-b390-80f5a8a499d8 tempest-DeleteServersAdminTestJSON-590116201 tempest-DeleteServersAdminTestJSON-590116201-project-member] [instance: 9ea2600a-d439-4fc3-a131-9bb93250d8cd] Binding failed for port a1ecfb50-9098-4408-b9bd-8db498ae622e, please check neutron logs for more information. {{(pid=61923) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 537.527577] env[61923]: DEBUG oslo_concurrency.lockutils [None req-36da9865-5d98-4c2a-a964-964bee74523c tempest-ListServerFiltersTestJSON-483030130 tempest-ListServerFiltersTestJSON-483030130-project-member] Acquiring lock "f3f872ce-cbbe-4407-b4e0-3341bb598825" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 537.527577] env[61923]: DEBUG oslo_concurrency.lockutils [None req-36da9865-5d98-4c2a-a964-964bee74523c tempest-ListServerFiltersTestJSON-483030130 tempest-ListServerFiltersTestJSON-483030130-project-member] Lock "f3f872ce-cbbe-4407-b4e0-3341bb598825" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 537.528216] env[61923]: DEBUG nova.compute.manager [None req-35a5fc5c-7e44-4806-b390-80f5a8a499d8 tempest-DeleteServersAdminTestJSON-590116201 tempest-DeleteServersAdminTestJSON-590116201-project-member] [instance: 9ea2600a-d439-4fc3-a131-9bb93250d8cd] Build of instance 9ea2600a-d439-4fc3-a131-9bb93250d8cd was re-scheduled: Binding failed for port a1ecfb50-9098-4408-b9bd-8db498ae622e, please check neutron logs for more information. {{(pid=61923) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 537.528447] env[61923]: DEBUG nova.compute.manager [None req-35a5fc5c-7e44-4806-b390-80f5a8a499d8 tempest-DeleteServersAdminTestJSON-590116201 tempest-DeleteServersAdminTestJSON-590116201-project-member] [instance: 9ea2600a-d439-4fc3-a131-9bb93250d8cd] Unplugging VIFs for instance {{(pid=61923) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 537.528651] env[61923]: DEBUG oslo_concurrency.lockutils [None req-35a5fc5c-7e44-4806-b390-80f5a8a499d8 tempest-DeleteServersAdminTestJSON-590116201 tempest-DeleteServersAdminTestJSON-590116201-project-member] Acquiring lock "refresh_cache-9ea2600a-d439-4fc3-a131-9bb93250d8cd" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 537.528789] env[61923]: DEBUG oslo_concurrency.lockutils [None req-35a5fc5c-7e44-4806-b390-80f5a8a499d8 tempest-DeleteServersAdminTestJSON-590116201 tempest-DeleteServersAdminTestJSON-590116201-project-member] Acquired lock "refresh_cache-9ea2600a-d439-4fc3-a131-9bb93250d8cd" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 537.529212] env[61923]: DEBUG nova.network.neutron [None req-35a5fc5c-7e44-4806-b390-80f5a8a499d8 tempest-DeleteServersAdminTestJSON-590116201 tempest-DeleteServersAdminTestJSON-590116201-project-member] [instance: 9ea2600a-d439-4fc3-a131-9bb93250d8cd] Building network info cache for instance {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 537.529892] env[61923]: DEBUG oslo_concurrency.lockutils [None req-dcb477a6-85d7-4838-be85-b9b2ecf96877 tempest-ImagesOneServerNegativeTestJSON-1560548781 tempest-ImagesOneServerNegativeTestJSON-1560548781-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 10.693s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 537.535788] env[61923]: INFO nova.compute.claims [None req-dcb477a6-85d7-4838-be85-b9b2ecf96877 tempest-ImagesOneServerNegativeTestJSON-1560548781 tempest-ImagesOneServerNegativeTestJSON-1560548781-project-member] [instance: ca62f959-43b9-4838-8349-11b7d6a7f0a7] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 537.576700] env[61923]: DEBUG oslo_concurrency.lockutils [req-76ee35b9-46da-4193-a86e-02ac4b60c220 req-d35462f9-f5ce-4d33-afcf-a9ac08040144 service nova] Releasing lock "refresh_cache-97148ea5-398a-4974-b8f6-62290e519794" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 537.714588] env[61923]: DEBUG nova.compute.manager [None req-f30f4038-8fd0-4bbd-8b35-7fb8e0dea3a4 tempest-SecurityGroupsTestJSON-509490211 tempest-SecurityGroupsTestJSON-509490211-project-member] [instance: 59fc11e8-27be-4553-9de9-3eb5ea612ec4] Start spawning the instance on the hypervisor. {{(pid=61923) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 537.752752] env[61923]: DEBUG nova.virt.hardware [None req-f30f4038-8fd0-4bbd-8b35-7fb8e0dea3a4 tempest-SecurityGroupsTestJSON-509490211 tempest-SecurityGroupsTestJSON-509490211-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-29T20:07:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-29T20:07:35Z,direct_url=,disk_format='vmdk',id=0f153f63-ae0a-45b1-b7f5-7f9b673c947f,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='4e7b5e3b3c3443c8bd5c70f8a15739f5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-29T20:07:36Z,virtual_size=,visibility=), allow threads: False {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 537.753179] env[61923]: DEBUG nova.virt.hardware [None req-f30f4038-8fd0-4bbd-8b35-7fb8e0dea3a4 tempest-SecurityGroupsTestJSON-509490211 tempest-SecurityGroupsTestJSON-509490211-project-member] Flavor limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 537.753390] env[61923]: DEBUG nova.virt.hardware [None req-f30f4038-8fd0-4bbd-8b35-7fb8e0dea3a4 tempest-SecurityGroupsTestJSON-509490211 tempest-SecurityGroupsTestJSON-509490211-project-member] Image limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 537.753586] env[61923]: DEBUG nova.virt.hardware [None req-f30f4038-8fd0-4bbd-8b35-7fb8e0dea3a4 tempest-SecurityGroupsTestJSON-509490211 tempest-SecurityGroupsTestJSON-509490211-project-member] Flavor pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 537.753732] env[61923]: DEBUG nova.virt.hardware [None req-f30f4038-8fd0-4bbd-8b35-7fb8e0dea3a4 tempest-SecurityGroupsTestJSON-509490211 tempest-SecurityGroupsTestJSON-509490211-project-member] Image pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 537.753953] env[61923]: DEBUG nova.virt.hardware [None req-f30f4038-8fd0-4bbd-8b35-7fb8e0dea3a4 tempest-SecurityGroupsTestJSON-509490211 tempest-SecurityGroupsTestJSON-509490211-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 537.754204] env[61923]: DEBUG nova.virt.hardware [None req-f30f4038-8fd0-4bbd-8b35-7fb8e0dea3a4 tempest-SecurityGroupsTestJSON-509490211 tempest-SecurityGroupsTestJSON-509490211-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 537.754367] env[61923]: DEBUG nova.virt.hardware [None req-f30f4038-8fd0-4bbd-8b35-7fb8e0dea3a4 tempest-SecurityGroupsTestJSON-509490211 tempest-SecurityGroupsTestJSON-509490211-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 537.754890] env[61923]: DEBUG nova.virt.hardware [None req-f30f4038-8fd0-4bbd-8b35-7fb8e0dea3a4 tempest-SecurityGroupsTestJSON-509490211 tempest-SecurityGroupsTestJSON-509490211-project-member] Got 1 possible topologies {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 537.755130] env[61923]: DEBUG nova.virt.hardware [None req-f30f4038-8fd0-4bbd-8b35-7fb8e0dea3a4 tempest-SecurityGroupsTestJSON-509490211 tempest-SecurityGroupsTestJSON-509490211-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 537.755364] env[61923]: DEBUG nova.virt.hardware [None req-f30f4038-8fd0-4bbd-8b35-7fb8e0dea3a4 tempest-SecurityGroupsTestJSON-509490211 tempest-SecurityGroupsTestJSON-509490211-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 537.756293] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aeb6802c-4e54-4d7b-985f-a9f70b1019f4 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 537.766074] env[61923]: DEBUG nova.compute.manager [None req-15b9e1af-fe57-407e-a058-c08239279239 tempest-ServersWithSpecificFlavorTestJSON-1806421899 tempest-ServersWithSpecificFlavorTestJSON-1806421899-project-member] [instance: 4cec7bce-0136-43ad-898f-086058a505f2] Starting instance... {{(pid=61923) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 537.770138] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44f3028b-9a48-4188-bacc-daeadd4bd169 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 537.878032] env[61923]: DEBUG nova.network.neutron [None req-844a2215-1038-4a30-8a79-40c1c3117bde tempest-FloatingIPsAssociationNegativeTestJSON-1301210771 tempest-FloatingIPsAssociationNegativeTestJSON-1301210771-project-member] [instance: 5d57303a-ae17-4b92-8f12-303677eb0d7e] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 538.099056] env[61923]: DEBUG nova.network.neutron [None req-35a5fc5c-7e44-4806-b390-80f5a8a499d8 tempest-DeleteServersAdminTestJSON-590116201 tempest-DeleteServersAdminTestJSON-590116201-project-member] [instance: 9ea2600a-d439-4fc3-a131-9bb93250d8cd] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 538.308137] env[61923]: DEBUG oslo_concurrency.lockutils [None req-15b9e1af-fe57-407e-a058-c08239279239 tempest-ServersWithSpecificFlavorTestJSON-1806421899 tempest-ServersWithSpecificFlavorTestJSON-1806421899-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 538.325432] env[61923]: DEBUG nova.network.neutron [None req-35a5fc5c-7e44-4806-b390-80f5a8a499d8 tempest-DeleteServersAdminTestJSON-590116201 tempest-DeleteServersAdminTestJSON-590116201-project-member] [instance: 9ea2600a-d439-4fc3-a131-9bb93250d8cd] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 538.381217] env[61923]: DEBUG oslo_concurrency.lockutils [None req-844a2215-1038-4a30-8a79-40c1c3117bde tempest-FloatingIPsAssociationNegativeTestJSON-1301210771 tempest-FloatingIPsAssociationNegativeTestJSON-1301210771-project-member] Releasing lock "refresh_cache-5d57303a-ae17-4b92-8f12-303677eb0d7e" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 538.381651] env[61923]: DEBUG nova.compute.manager [None req-844a2215-1038-4a30-8a79-40c1c3117bde tempest-FloatingIPsAssociationNegativeTestJSON-1301210771 tempest-FloatingIPsAssociationNegativeTestJSON-1301210771-project-member] [instance: 5d57303a-ae17-4b92-8f12-303677eb0d7e] Start destroying the instance on the hypervisor. {{(pid=61923) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 538.381838] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-844a2215-1038-4a30-8a79-40c1c3117bde tempest-FloatingIPsAssociationNegativeTestJSON-1301210771 tempest-FloatingIPsAssociationNegativeTestJSON-1301210771-project-member] [instance: 5d57303a-ae17-4b92-8f12-303677eb0d7e] Destroying instance {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 538.382845] env[61923]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-51def5b7-734f-4f84-928b-3a42f80d2210 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 538.393563] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a918ddd-3098-4989-85c1-49ac3302aa22 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 538.426352] env[61923]: WARNING nova.virt.vmwareapi.vmops [None req-844a2215-1038-4a30-8a79-40c1c3117bde tempest-FloatingIPsAssociationNegativeTestJSON-1301210771 tempest-FloatingIPsAssociationNegativeTestJSON-1301210771-project-member] [instance: 5d57303a-ae17-4b92-8f12-303677eb0d7e] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 5d57303a-ae17-4b92-8f12-303677eb0d7e could not be found. [ 538.426595] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-844a2215-1038-4a30-8a79-40c1c3117bde tempest-FloatingIPsAssociationNegativeTestJSON-1301210771 tempest-FloatingIPsAssociationNegativeTestJSON-1301210771-project-member] [instance: 5d57303a-ae17-4b92-8f12-303677eb0d7e] Instance destroyed {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 538.426788] env[61923]: INFO nova.compute.manager [None req-844a2215-1038-4a30-8a79-40c1c3117bde tempest-FloatingIPsAssociationNegativeTestJSON-1301210771 tempest-FloatingIPsAssociationNegativeTestJSON-1301210771-project-member] [instance: 5d57303a-ae17-4b92-8f12-303677eb0d7e] Took 0.04 seconds to destroy the instance on the hypervisor. [ 538.427078] env[61923]: DEBUG oslo.service.loopingcall [None req-844a2215-1038-4a30-8a79-40c1c3117bde tempest-FloatingIPsAssociationNegativeTestJSON-1301210771 tempest-FloatingIPsAssociationNegativeTestJSON-1301210771-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61923) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 538.427970] env[61923]: DEBUG nova.compute.manager [-] [instance: 5d57303a-ae17-4b92-8f12-303677eb0d7e] Deallocating network for instance {{(pid=61923) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 538.427970] env[61923]: DEBUG nova.network.neutron [-] [instance: 5d57303a-ae17-4b92-8f12-303677eb0d7e] deallocate_for_instance() {{(pid=61923) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 538.482770] env[61923]: DEBUG nova.network.neutron [-] [instance: 5d57303a-ae17-4b92-8f12-303677eb0d7e] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 538.789832] env[61923]: DEBUG nova.compute.manager [req-e03017e3-a061-4de3-a93c-40e897649b91 req-56a24fb6-4aad-4c82-99c0-0cf7c95b86f4 service nova] [instance: 97148ea5-398a-4974-b8f6-62290e519794] Received event network-vif-deleted-7ec53f9e-f055-4a57-8228-7efbcdb7360f {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 538.790155] env[61923]: DEBUG nova.compute.manager [req-e03017e3-a061-4de3-a93c-40e897649b91 req-56a24fb6-4aad-4c82-99c0-0cf7c95b86f4 service nova] [instance: 5d57303a-ae17-4b92-8f12-303677eb0d7e] Received event network-changed-4f3fea2e-b3d9-4c89-b291-071aa6707d27 {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 538.794361] env[61923]: DEBUG nova.compute.manager [req-e03017e3-a061-4de3-a93c-40e897649b91 req-56a24fb6-4aad-4c82-99c0-0cf7c95b86f4 service nova] [instance: 5d57303a-ae17-4b92-8f12-303677eb0d7e] Refreshing instance network info cache due to event network-changed-4f3fea2e-b3d9-4c89-b291-071aa6707d27. {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 538.794627] env[61923]: DEBUG oslo_concurrency.lockutils [req-e03017e3-a061-4de3-a93c-40e897649b91 req-56a24fb6-4aad-4c82-99c0-0cf7c95b86f4 service nova] Acquiring lock "refresh_cache-5d57303a-ae17-4b92-8f12-303677eb0d7e" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 538.794800] env[61923]: DEBUG oslo_concurrency.lockutils [req-e03017e3-a061-4de3-a93c-40e897649b91 req-56a24fb6-4aad-4c82-99c0-0cf7c95b86f4 service nova] Acquired lock "refresh_cache-5d57303a-ae17-4b92-8f12-303677eb0d7e" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 538.794892] env[61923]: DEBUG nova.network.neutron [req-e03017e3-a061-4de3-a93c-40e897649b91 req-56a24fb6-4aad-4c82-99c0-0cf7c95b86f4 service nova] [instance: 5d57303a-ae17-4b92-8f12-303677eb0d7e] Refreshing network info cache for port 4f3fea2e-b3d9-4c89-b291-071aa6707d27 {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 538.830838] env[61923]: DEBUG oslo_concurrency.lockutils [None req-35a5fc5c-7e44-4806-b390-80f5a8a499d8 tempest-DeleteServersAdminTestJSON-590116201 tempest-DeleteServersAdminTestJSON-590116201-project-member] Releasing lock "refresh_cache-9ea2600a-d439-4fc3-a131-9bb93250d8cd" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 538.830838] env[61923]: DEBUG nova.compute.manager [None req-35a5fc5c-7e44-4806-b390-80f5a8a499d8 tempest-DeleteServersAdminTestJSON-590116201 tempest-DeleteServersAdminTestJSON-590116201-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61923) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 538.830838] env[61923]: DEBUG nova.compute.manager [None req-35a5fc5c-7e44-4806-b390-80f5a8a499d8 tempest-DeleteServersAdminTestJSON-590116201 tempest-DeleteServersAdminTestJSON-590116201-project-member] [instance: 9ea2600a-d439-4fc3-a131-9bb93250d8cd] Deallocating network for instance {{(pid=61923) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 538.830838] env[61923]: DEBUG nova.network.neutron [None req-35a5fc5c-7e44-4806-b390-80f5a8a499d8 tempest-DeleteServersAdminTestJSON-590116201 tempest-DeleteServersAdminTestJSON-590116201-project-member] [instance: 9ea2600a-d439-4fc3-a131-9bb93250d8cd] deallocate_for_instance() {{(pid=61923) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 538.847764] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-efe36c28-381c-4624-a8fd-db7164f6947b {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 538.858550] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93a27037-177e-435c-b749-92d4b89af26f {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 538.899163] env[61923]: DEBUG nova.network.neutron [None req-35a5fc5c-7e44-4806-b390-80f5a8a499d8 tempest-DeleteServersAdminTestJSON-590116201 tempest-DeleteServersAdminTestJSON-590116201-project-member] [instance: 9ea2600a-d439-4fc3-a131-9bb93250d8cd] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 538.900130] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c4381a0-2e36-45a1-9810-197a78f5ff4a {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 538.909504] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e01b419b-a5eb-4dec-8a01-a840177bd77a {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 538.929316] env[61923]: DEBUG nova.compute.provider_tree [None req-dcb477a6-85d7-4838-be85-b9b2ecf96877 tempest-ImagesOneServerNegativeTestJSON-1560548781 tempest-ImagesOneServerNegativeTestJSON-1560548781-project-member] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 538.968066] env[61923]: DEBUG oslo_concurrency.lockutils [None req-fba27a13-474d-4fe3-9187-c31143f40aa5 tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] Acquiring lock "6a2d7ede-e0cc-4a36-b1b2-0b36b269ecb5" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 538.968289] env[61923]: DEBUG oslo_concurrency.lockutils [None req-fba27a13-474d-4fe3-9187-c31143f40aa5 tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] Lock "6a2d7ede-e0cc-4a36-b1b2-0b36b269ecb5" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 538.988996] env[61923]: DEBUG nova.network.neutron [-] [instance: 5d57303a-ae17-4b92-8f12-303677eb0d7e] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 538.997172] env[61923]: DEBUG oslo_concurrency.lockutils [None req-fba27a13-474d-4fe3-9187-c31143f40aa5 tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] Acquiring lock "9c3e2f94-91da-4255-b4d6-1c01e100ff48" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 538.997423] env[61923]: DEBUG oslo_concurrency.lockutils [None req-fba27a13-474d-4fe3-9187-c31143f40aa5 tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] Lock "9c3e2f94-91da-4255-b4d6-1c01e100ff48" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 539.051400] env[61923]: DEBUG oslo_concurrency.lockutils [None req-fba27a13-474d-4fe3-9187-c31143f40aa5 tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] Acquiring lock "93818ffd-eb85-48f0-aebd-9774467b0a8f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 539.051626] env[61923]: DEBUG oslo_concurrency.lockutils [None req-fba27a13-474d-4fe3-9187-c31143f40aa5 tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] Lock "93818ffd-eb85-48f0-aebd-9774467b0a8f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 539.104618] env[61923]: DEBUG oslo_concurrency.lockutils [None req-543d2693-878a-4ade-aaa2-339016c76eca tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Acquiring lock "045c0b05-333e-4f95-94b0-8f51d87dca0a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 539.105015] env[61923]: DEBUG oslo_concurrency.lockutils [None req-543d2693-878a-4ade-aaa2-339016c76eca tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Lock "045c0b05-333e-4f95-94b0-8f51d87dca0a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 539.340901] env[61923]: ERROR nova.compute.manager [None req-f30f4038-8fd0-4bbd-8b35-7fb8e0dea3a4 tempest-SecurityGroupsTestJSON-509490211 tempest-SecurityGroupsTestJSON-509490211-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 238265b6-446c-493c-b510-5f51005c1787, please check neutron logs for more information. [ 539.340901] env[61923]: ERROR nova.compute.manager Traceback (most recent call last): [ 539.340901] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 539.340901] env[61923]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 539.340901] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 539.340901] env[61923]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 539.340901] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 539.340901] env[61923]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 539.340901] env[61923]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 539.340901] env[61923]: ERROR nova.compute.manager self.force_reraise() [ 539.340901] env[61923]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 539.340901] env[61923]: ERROR nova.compute.manager raise self.value [ 539.340901] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 539.340901] env[61923]: ERROR nova.compute.manager updated_port = self._update_port( [ 539.340901] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 539.340901] env[61923]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 539.341396] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 539.341396] env[61923]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 539.341396] env[61923]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 238265b6-446c-493c-b510-5f51005c1787, please check neutron logs for more information. [ 539.341396] env[61923]: ERROR nova.compute.manager [ 539.341396] env[61923]: Traceback (most recent call last): [ 539.341396] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 539.341396] env[61923]: listener.cb(fileno) [ 539.341396] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 539.341396] env[61923]: result = function(*args, **kwargs) [ 539.341396] env[61923]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 539.341396] env[61923]: return func(*args, **kwargs) [ 539.341396] env[61923]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 539.341396] env[61923]: raise e [ 539.341396] env[61923]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 539.341396] env[61923]: nwinfo = self.network_api.allocate_for_instance( [ 539.341396] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 539.341396] env[61923]: created_port_ids = self._update_ports_for_instance( [ 539.341396] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 539.341396] env[61923]: with excutils.save_and_reraise_exception(): [ 539.341396] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 539.341396] env[61923]: self.force_reraise() [ 539.341396] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 539.341396] env[61923]: raise self.value [ 539.341396] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 539.341396] env[61923]: updated_port = self._update_port( [ 539.341396] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 539.341396] env[61923]: _ensure_no_port_binding_failure(port) [ 539.341396] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 539.341396] env[61923]: raise exception.PortBindingFailed(port_id=port['id']) [ 539.342223] env[61923]: nova.exception.PortBindingFailed: Binding failed for port 238265b6-446c-493c-b510-5f51005c1787, please check neutron logs for more information. [ 539.342223] env[61923]: Removing descriptor: 14 [ 539.342223] env[61923]: ERROR nova.compute.manager [None req-f30f4038-8fd0-4bbd-8b35-7fb8e0dea3a4 tempest-SecurityGroupsTestJSON-509490211 tempest-SecurityGroupsTestJSON-509490211-project-member] [instance: 59fc11e8-27be-4553-9de9-3eb5ea612ec4] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 238265b6-446c-493c-b510-5f51005c1787, please check neutron logs for more information. [ 539.342223] env[61923]: ERROR nova.compute.manager [instance: 59fc11e8-27be-4553-9de9-3eb5ea612ec4] Traceback (most recent call last): [ 539.342223] env[61923]: ERROR nova.compute.manager [instance: 59fc11e8-27be-4553-9de9-3eb5ea612ec4] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 539.342223] env[61923]: ERROR nova.compute.manager [instance: 59fc11e8-27be-4553-9de9-3eb5ea612ec4] yield resources [ 539.342223] env[61923]: ERROR nova.compute.manager [instance: 59fc11e8-27be-4553-9de9-3eb5ea612ec4] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 539.342223] env[61923]: ERROR nova.compute.manager [instance: 59fc11e8-27be-4553-9de9-3eb5ea612ec4] self.driver.spawn(context, instance, image_meta, [ 539.342223] env[61923]: ERROR nova.compute.manager [instance: 59fc11e8-27be-4553-9de9-3eb5ea612ec4] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 539.342223] env[61923]: ERROR nova.compute.manager [instance: 59fc11e8-27be-4553-9de9-3eb5ea612ec4] self._vmops.spawn(context, instance, image_meta, injected_files, [ 539.342223] env[61923]: ERROR nova.compute.manager [instance: 59fc11e8-27be-4553-9de9-3eb5ea612ec4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 539.342223] env[61923]: ERROR nova.compute.manager [instance: 59fc11e8-27be-4553-9de9-3eb5ea612ec4] vm_ref = self.build_virtual_machine(instance, [ 539.342670] env[61923]: ERROR nova.compute.manager [instance: 59fc11e8-27be-4553-9de9-3eb5ea612ec4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 539.342670] env[61923]: ERROR nova.compute.manager [instance: 59fc11e8-27be-4553-9de9-3eb5ea612ec4] vif_infos = vmwarevif.get_vif_info(self._session, [ 539.342670] env[61923]: ERROR nova.compute.manager [instance: 59fc11e8-27be-4553-9de9-3eb5ea612ec4] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 539.342670] env[61923]: ERROR nova.compute.manager [instance: 59fc11e8-27be-4553-9de9-3eb5ea612ec4] for vif in network_info: [ 539.342670] env[61923]: ERROR nova.compute.manager [instance: 59fc11e8-27be-4553-9de9-3eb5ea612ec4] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 539.342670] env[61923]: ERROR nova.compute.manager [instance: 59fc11e8-27be-4553-9de9-3eb5ea612ec4] return self._sync_wrapper(fn, *args, **kwargs) [ 539.342670] env[61923]: ERROR nova.compute.manager [instance: 59fc11e8-27be-4553-9de9-3eb5ea612ec4] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 539.342670] env[61923]: ERROR nova.compute.manager [instance: 59fc11e8-27be-4553-9de9-3eb5ea612ec4] self.wait() [ 539.342670] env[61923]: ERROR nova.compute.manager [instance: 59fc11e8-27be-4553-9de9-3eb5ea612ec4] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 539.342670] env[61923]: ERROR nova.compute.manager [instance: 59fc11e8-27be-4553-9de9-3eb5ea612ec4] self[:] = self._gt.wait() [ 539.342670] env[61923]: ERROR nova.compute.manager [instance: 59fc11e8-27be-4553-9de9-3eb5ea612ec4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 539.342670] env[61923]: ERROR nova.compute.manager [instance: 59fc11e8-27be-4553-9de9-3eb5ea612ec4] return self._exit_event.wait() [ 539.342670] env[61923]: ERROR nova.compute.manager [instance: 59fc11e8-27be-4553-9de9-3eb5ea612ec4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 539.343065] env[61923]: ERROR nova.compute.manager [instance: 59fc11e8-27be-4553-9de9-3eb5ea612ec4] result = hub.switch() [ 539.343065] env[61923]: ERROR nova.compute.manager [instance: 59fc11e8-27be-4553-9de9-3eb5ea612ec4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 539.343065] env[61923]: ERROR nova.compute.manager [instance: 59fc11e8-27be-4553-9de9-3eb5ea612ec4] return self.greenlet.switch() [ 539.343065] env[61923]: ERROR nova.compute.manager [instance: 59fc11e8-27be-4553-9de9-3eb5ea612ec4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 539.343065] env[61923]: ERROR nova.compute.manager [instance: 59fc11e8-27be-4553-9de9-3eb5ea612ec4] result = function(*args, **kwargs) [ 539.343065] env[61923]: ERROR nova.compute.manager [instance: 59fc11e8-27be-4553-9de9-3eb5ea612ec4] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 539.343065] env[61923]: ERROR nova.compute.manager [instance: 59fc11e8-27be-4553-9de9-3eb5ea612ec4] return func(*args, **kwargs) [ 539.343065] env[61923]: ERROR nova.compute.manager [instance: 59fc11e8-27be-4553-9de9-3eb5ea612ec4] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 539.343065] env[61923]: ERROR nova.compute.manager [instance: 59fc11e8-27be-4553-9de9-3eb5ea612ec4] raise e [ 539.343065] env[61923]: ERROR nova.compute.manager [instance: 59fc11e8-27be-4553-9de9-3eb5ea612ec4] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 539.343065] env[61923]: ERROR nova.compute.manager [instance: 59fc11e8-27be-4553-9de9-3eb5ea612ec4] nwinfo = self.network_api.allocate_for_instance( [ 539.343065] env[61923]: ERROR nova.compute.manager [instance: 59fc11e8-27be-4553-9de9-3eb5ea612ec4] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 539.343065] env[61923]: ERROR nova.compute.manager [instance: 59fc11e8-27be-4553-9de9-3eb5ea612ec4] created_port_ids = self._update_ports_for_instance( [ 539.343422] env[61923]: ERROR nova.compute.manager [instance: 59fc11e8-27be-4553-9de9-3eb5ea612ec4] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 539.343422] env[61923]: ERROR nova.compute.manager [instance: 59fc11e8-27be-4553-9de9-3eb5ea612ec4] with excutils.save_and_reraise_exception(): [ 539.343422] env[61923]: ERROR nova.compute.manager [instance: 59fc11e8-27be-4553-9de9-3eb5ea612ec4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 539.343422] env[61923]: ERROR nova.compute.manager [instance: 59fc11e8-27be-4553-9de9-3eb5ea612ec4] self.force_reraise() [ 539.343422] env[61923]: ERROR nova.compute.manager [instance: 59fc11e8-27be-4553-9de9-3eb5ea612ec4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 539.343422] env[61923]: ERROR nova.compute.manager [instance: 59fc11e8-27be-4553-9de9-3eb5ea612ec4] raise self.value [ 539.343422] env[61923]: ERROR nova.compute.manager [instance: 59fc11e8-27be-4553-9de9-3eb5ea612ec4] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 539.343422] env[61923]: ERROR nova.compute.manager [instance: 59fc11e8-27be-4553-9de9-3eb5ea612ec4] updated_port = self._update_port( [ 539.343422] env[61923]: ERROR nova.compute.manager [instance: 59fc11e8-27be-4553-9de9-3eb5ea612ec4] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 539.343422] env[61923]: ERROR nova.compute.manager [instance: 59fc11e8-27be-4553-9de9-3eb5ea612ec4] _ensure_no_port_binding_failure(port) [ 539.343422] env[61923]: ERROR nova.compute.manager [instance: 59fc11e8-27be-4553-9de9-3eb5ea612ec4] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 539.343422] env[61923]: ERROR nova.compute.manager [instance: 59fc11e8-27be-4553-9de9-3eb5ea612ec4] raise exception.PortBindingFailed(port_id=port['id']) [ 539.343774] env[61923]: ERROR nova.compute.manager [instance: 59fc11e8-27be-4553-9de9-3eb5ea612ec4] nova.exception.PortBindingFailed: Binding failed for port 238265b6-446c-493c-b510-5f51005c1787, please check neutron logs for more information. [ 539.343774] env[61923]: ERROR nova.compute.manager [instance: 59fc11e8-27be-4553-9de9-3eb5ea612ec4] [ 539.343774] env[61923]: INFO nova.compute.manager [None req-f30f4038-8fd0-4bbd-8b35-7fb8e0dea3a4 tempest-SecurityGroupsTestJSON-509490211 tempest-SecurityGroupsTestJSON-509490211-project-member] [instance: 59fc11e8-27be-4553-9de9-3eb5ea612ec4] Terminating instance [ 539.350322] env[61923]: DEBUG oslo_concurrency.lockutils [None req-f30f4038-8fd0-4bbd-8b35-7fb8e0dea3a4 tempest-SecurityGroupsTestJSON-509490211 tempest-SecurityGroupsTestJSON-509490211-project-member] Acquiring lock "refresh_cache-59fc11e8-27be-4553-9de9-3eb5ea612ec4" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 539.350472] env[61923]: DEBUG oslo_concurrency.lockutils [None req-f30f4038-8fd0-4bbd-8b35-7fb8e0dea3a4 tempest-SecurityGroupsTestJSON-509490211 tempest-SecurityGroupsTestJSON-509490211-project-member] Acquired lock "refresh_cache-59fc11e8-27be-4553-9de9-3eb5ea612ec4" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 539.350640] env[61923]: DEBUG nova.network.neutron [None req-f30f4038-8fd0-4bbd-8b35-7fb8e0dea3a4 tempest-SecurityGroupsTestJSON-509490211 tempest-SecurityGroupsTestJSON-509490211-project-member] [instance: 59fc11e8-27be-4553-9de9-3eb5ea612ec4] Building network info cache for instance {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 539.366064] env[61923]: DEBUG nova.network.neutron [req-e03017e3-a061-4de3-a93c-40e897649b91 req-56a24fb6-4aad-4c82-99c0-0cf7c95b86f4 service nova] [instance: 5d57303a-ae17-4b92-8f12-303677eb0d7e] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 539.403781] env[61923]: DEBUG nova.network.neutron [None req-35a5fc5c-7e44-4806-b390-80f5a8a499d8 tempest-DeleteServersAdminTestJSON-590116201 tempest-DeleteServersAdminTestJSON-590116201-project-member] [instance: 9ea2600a-d439-4fc3-a131-9bb93250d8cd] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 539.434041] env[61923]: DEBUG nova.scheduler.client.report [None req-dcb477a6-85d7-4838-be85-b9b2ecf96877 tempest-ImagesOneServerNegativeTestJSON-1560548781 tempest-ImagesOneServerNegativeTestJSON-1560548781-project-member] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 539.492844] env[61923]: INFO nova.compute.manager [-] [instance: 5d57303a-ae17-4b92-8f12-303677eb0d7e] Took 1.07 seconds to deallocate network for instance. [ 539.495885] env[61923]: DEBUG nova.compute.claims [None req-844a2215-1038-4a30-8a79-40c1c3117bde tempest-FloatingIPsAssociationNegativeTestJSON-1301210771 tempest-FloatingIPsAssociationNegativeTestJSON-1301210771-project-member] [instance: 5d57303a-ae17-4b92-8f12-303677eb0d7e] Aborting claim: {{(pid=61923) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 539.496132] env[61923]: DEBUG oslo_concurrency.lockutils [None req-844a2215-1038-4a30-8a79-40c1c3117bde tempest-FloatingIPsAssociationNegativeTestJSON-1301210771 tempest-FloatingIPsAssociationNegativeTestJSON-1301210771-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 539.667741] env[61923]: DEBUG nova.network.neutron [req-e03017e3-a061-4de3-a93c-40e897649b91 req-56a24fb6-4aad-4c82-99c0-0cf7c95b86f4 service nova] [instance: 5d57303a-ae17-4b92-8f12-303677eb0d7e] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 539.895563] env[61923]: DEBUG nova.network.neutron [None req-f30f4038-8fd0-4bbd-8b35-7fb8e0dea3a4 tempest-SecurityGroupsTestJSON-509490211 tempest-SecurityGroupsTestJSON-509490211-project-member] [instance: 59fc11e8-27be-4553-9de9-3eb5ea612ec4] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 539.913050] env[61923]: INFO nova.compute.manager [None req-35a5fc5c-7e44-4806-b390-80f5a8a499d8 tempest-DeleteServersAdminTestJSON-590116201 tempest-DeleteServersAdminTestJSON-590116201-project-member] [instance: 9ea2600a-d439-4fc3-a131-9bb93250d8cd] Took 1.08 seconds to deallocate network for instance. [ 539.941575] env[61923]: DEBUG oslo_concurrency.lockutils [None req-dcb477a6-85d7-4838-be85-b9b2ecf96877 tempest-ImagesOneServerNegativeTestJSON-1560548781 tempest-ImagesOneServerNegativeTestJSON-1560548781-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.411s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 539.941959] env[61923]: DEBUG nova.compute.manager [None req-dcb477a6-85d7-4838-be85-b9b2ecf96877 tempest-ImagesOneServerNegativeTestJSON-1560548781 tempest-ImagesOneServerNegativeTestJSON-1560548781-project-member] [instance: ca62f959-43b9-4838-8349-11b7d6a7f0a7] Start building networks asynchronously for instance. {{(pid=61923) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 539.944838] env[61923]: DEBUG oslo_concurrency.lockutils [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 12.529s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 539.945518] env[61923]: DEBUG oslo_concurrency.lockutils [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 539.945518] env[61923]: DEBUG nova.compute.resource_tracker [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61923) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 539.945518] env[61923]: DEBUG oslo_concurrency.lockutils [None req-464a16f8-1845-4d4d-b3ea-fd8ba00de8bb tempest-ServersAdminNegativeTestJSON-1828705899 tempest-ServersAdminNegativeTestJSON-1828705899-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 11.304s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 539.955960] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef75225f-df33-4015-9ea0-a54919864076 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 539.965030] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d92843e-6b47-4c36-a143-6e515a50b7e2 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 539.994238] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64078e15-62bc-49d4-a6ac-6362d3a8a4db {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 540.003368] env[61923]: DEBUG oslo_concurrency.lockutils [None req-58d8b214-0b32-4993-bdd8-d652de7f9397 tempest-ServerActionsTestOtherA-1581037736 tempest-ServerActionsTestOtherA-1581037736-project-member] Acquiring lock "af5feda7-204a-47f1-b7fa-afe44832b9a2" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 540.003618] env[61923]: DEBUG oslo_concurrency.lockutils [None req-58d8b214-0b32-4993-bdd8-d652de7f9397 tempest-ServerActionsTestOtherA-1581037736 tempest-ServerActionsTestOtherA-1581037736-project-member] Lock "af5feda7-204a-47f1-b7fa-afe44832b9a2" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 540.009504] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c570bb9-435a-48d1-9e6b-ba98768437a9 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 540.056701] env[61923]: DEBUG nova.compute.resource_tracker [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181435MB free_disk=178GB free_vcpus=48 pci_devices=None {{(pid=61923) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 540.056853] env[61923]: DEBUG oslo_concurrency.lockutils [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 540.082276] env[61923]: DEBUG oslo_concurrency.lockutils [None req-595ebb22-536b-4375-9cbe-6fc472a5a17d tempest-ListServerFiltersTestJSON-483030130 tempest-ListServerFiltersTestJSON-483030130-project-member] Acquiring lock "71d9eccc-3b90-4b0a-a59a-d6a8e08a9417" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 540.083026] env[61923]: DEBUG oslo_concurrency.lockutils [None req-595ebb22-536b-4375-9cbe-6fc472a5a17d tempest-ListServerFiltersTestJSON-483030130 tempest-ListServerFiltersTestJSON-483030130-project-member] Lock "71d9eccc-3b90-4b0a-a59a-d6a8e08a9417" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 540.116973] env[61923]: DEBUG nova.network.neutron [None req-f30f4038-8fd0-4bbd-8b35-7fb8e0dea3a4 tempest-SecurityGroupsTestJSON-509490211 tempest-SecurityGroupsTestJSON-509490211-project-member] [instance: 59fc11e8-27be-4553-9de9-3eb5ea612ec4] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 540.174534] env[61923]: DEBUG oslo_concurrency.lockutils [req-e03017e3-a061-4de3-a93c-40e897649b91 req-56a24fb6-4aad-4c82-99c0-0cf7c95b86f4 service nova] Releasing lock "refresh_cache-5d57303a-ae17-4b92-8f12-303677eb0d7e" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 540.457728] env[61923]: DEBUG nova.compute.utils [None req-dcb477a6-85d7-4838-be85-b9b2ecf96877 tempest-ImagesOneServerNegativeTestJSON-1560548781 tempest-ImagesOneServerNegativeTestJSON-1560548781-project-member] Using /dev/sd instead of None {{(pid=61923) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 540.459262] env[61923]: DEBUG nova.compute.manager [None req-dcb477a6-85d7-4838-be85-b9b2ecf96877 tempest-ImagesOneServerNegativeTestJSON-1560548781 tempest-ImagesOneServerNegativeTestJSON-1560548781-project-member] [instance: ca62f959-43b9-4838-8349-11b7d6a7f0a7] Allocating IP information in the background. {{(pid=61923) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 540.465259] env[61923]: DEBUG nova.network.neutron [None req-dcb477a6-85d7-4838-be85-b9b2ecf96877 tempest-ImagesOneServerNegativeTestJSON-1560548781 tempest-ImagesOneServerNegativeTestJSON-1560548781-project-member] [instance: ca62f959-43b9-4838-8349-11b7d6a7f0a7] allocate_for_instance() {{(pid=61923) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 540.619458] env[61923]: DEBUG oslo_concurrency.lockutils [None req-f30f4038-8fd0-4bbd-8b35-7fb8e0dea3a4 tempest-SecurityGroupsTestJSON-509490211 tempest-SecurityGroupsTestJSON-509490211-project-member] Releasing lock "refresh_cache-59fc11e8-27be-4553-9de9-3eb5ea612ec4" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 540.620053] env[61923]: DEBUG nova.compute.manager [None req-f30f4038-8fd0-4bbd-8b35-7fb8e0dea3a4 tempest-SecurityGroupsTestJSON-509490211 tempest-SecurityGroupsTestJSON-509490211-project-member] [instance: 59fc11e8-27be-4553-9de9-3eb5ea612ec4] Start destroying the instance on the hypervisor. {{(pid=61923) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 540.620476] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-f30f4038-8fd0-4bbd-8b35-7fb8e0dea3a4 tempest-SecurityGroupsTestJSON-509490211 tempest-SecurityGroupsTestJSON-509490211-project-member] [instance: 59fc11e8-27be-4553-9de9-3eb5ea612ec4] Destroying instance {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 540.621724] env[61923]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-9df290f6-3d6d-4cf0-83d5-40a02bd2b508 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 540.638875] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa9b11a6-379d-40a5-9f2f-b7fbb3b07580 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 540.678227] env[61923]: WARNING nova.virt.vmwareapi.vmops [None req-f30f4038-8fd0-4bbd-8b35-7fb8e0dea3a4 tempest-SecurityGroupsTestJSON-509490211 tempest-SecurityGroupsTestJSON-509490211-project-member] [instance: 59fc11e8-27be-4553-9de9-3eb5ea612ec4] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 59fc11e8-27be-4553-9de9-3eb5ea612ec4 could not be found. [ 540.678595] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-f30f4038-8fd0-4bbd-8b35-7fb8e0dea3a4 tempest-SecurityGroupsTestJSON-509490211 tempest-SecurityGroupsTestJSON-509490211-project-member] [instance: 59fc11e8-27be-4553-9de9-3eb5ea612ec4] Instance destroyed {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 540.678682] env[61923]: INFO nova.compute.manager [None req-f30f4038-8fd0-4bbd-8b35-7fb8e0dea3a4 tempest-SecurityGroupsTestJSON-509490211 tempest-SecurityGroupsTestJSON-509490211-project-member] [instance: 59fc11e8-27be-4553-9de9-3eb5ea612ec4] Took 0.06 seconds to destroy the instance on the hypervisor. [ 540.679821] env[61923]: DEBUG oslo.service.loopingcall [None req-f30f4038-8fd0-4bbd-8b35-7fb8e0dea3a4 tempest-SecurityGroupsTestJSON-509490211 tempest-SecurityGroupsTestJSON-509490211-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61923) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 540.679821] env[61923]: DEBUG nova.compute.manager [-] [instance: 59fc11e8-27be-4553-9de9-3eb5ea612ec4] Deallocating network for instance {{(pid=61923) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 540.679821] env[61923]: DEBUG nova.network.neutron [-] [instance: 59fc11e8-27be-4553-9de9-3eb5ea612ec4] deallocate_for_instance() {{(pid=61923) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 540.713514] env[61923]: DEBUG nova.network.neutron [-] [instance: 59fc11e8-27be-4553-9de9-3eb5ea612ec4] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 540.720165] env[61923]: DEBUG nova.policy [None req-dcb477a6-85d7-4838-be85-b9b2ecf96877 tempest-ImagesOneServerNegativeTestJSON-1560548781 tempest-ImagesOneServerNegativeTestJSON-1560548781-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '08eec39f0de942fda8246c823b202cd4', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '03f31bbddde641f194787c46239eaa71', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61923) authorize /opt/stack/nova/nova/policy.py:201}} [ 540.873337] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b37cab7-2ef7-4e31-8551-d89a99d0be59 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 540.885621] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd846270-7c16-49f8-aa29-13dfb81b7bc5 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 540.946511] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-735cd547-7039-4cd5-834a-afcd07104886 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 540.961921] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6f316ea-ff38-4f7e-a734-742719c7131b {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 540.978694] env[61923]: DEBUG nova.compute.manager [None req-dcb477a6-85d7-4838-be85-b9b2ecf96877 tempest-ImagesOneServerNegativeTestJSON-1560548781 tempest-ImagesOneServerNegativeTestJSON-1560548781-project-member] [instance: ca62f959-43b9-4838-8349-11b7d6a7f0a7] Start building block device mappings for instance. {{(pid=61923) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 540.982112] env[61923]: DEBUG nova.compute.provider_tree [None req-464a16f8-1845-4d4d-b3ea-fd8ba00de8bb tempest-ServersAdminNegativeTestJSON-1828705899 tempest-ServersAdminNegativeTestJSON-1828705899-project-member] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 540.985102] env[61923]: INFO nova.scheduler.client.report [None req-35a5fc5c-7e44-4806-b390-80f5a8a499d8 tempest-DeleteServersAdminTestJSON-590116201 tempest-DeleteServersAdminTestJSON-590116201-project-member] Deleted allocations for instance 9ea2600a-d439-4fc3-a131-9bb93250d8cd [ 541.217753] env[61923]: DEBUG nova.network.neutron [-] [instance: 59fc11e8-27be-4553-9de9-3eb5ea612ec4] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 541.346102] env[61923]: DEBUG oslo_concurrency.lockutils [None req-03cb7270-244b-47b9-9dd5-e358ebdfa946 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Acquiring lock "cb998da7-8db0-4725-bed2-657d7df748b5" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 541.346340] env[61923]: DEBUG oslo_concurrency.lockutils [None req-03cb7270-244b-47b9-9dd5-e358ebdfa946 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Lock "cb998da7-8db0-4725-bed2-657d7df748b5" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 541.488167] env[61923]: DEBUG nova.scheduler.client.report [None req-464a16f8-1845-4d4d-b3ea-fd8ba00de8bb tempest-ServersAdminNegativeTestJSON-1828705899 tempest-ServersAdminNegativeTestJSON-1828705899-project-member] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 541.499321] env[61923]: DEBUG oslo_concurrency.lockutils [None req-35a5fc5c-7e44-4806-b390-80f5a8a499d8 tempest-DeleteServersAdminTestJSON-590116201 tempest-DeleteServersAdminTestJSON-590116201-project-member] Lock "9ea2600a-d439-4fc3-a131-9bb93250d8cd" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 24.277s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 541.577283] env[61923]: DEBUG nova.network.neutron [None req-dcb477a6-85d7-4838-be85-b9b2ecf96877 tempest-ImagesOneServerNegativeTestJSON-1560548781 tempest-ImagesOneServerNegativeTestJSON-1560548781-project-member] [instance: ca62f959-43b9-4838-8349-11b7d6a7f0a7] Successfully created port: 63e8ed78-99f3-41e0-b0a6-e6bc6fe00fd4 {{(pid=61923) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 541.720188] env[61923]: DEBUG oslo_concurrency.lockutils [None req-f3aa5656-98df-4c46-9436-b968a8770dbc tempest-ListServerFiltersTestJSON-483030130 tempest-ListServerFiltersTestJSON-483030130-project-member] Acquiring lock "896dd689-4810-4f23-af0a-d2f557a0796b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 541.720421] env[61923]: DEBUG oslo_concurrency.lockutils [None req-f3aa5656-98df-4c46-9436-b968a8770dbc tempest-ListServerFiltersTestJSON-483030130 tempest-ListServerFiltersTestJSON-483030130-project-member] Lock "896dd689-4810-4f23-af0a-d2f557a0796b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 541.720745] env[61923]: INFO nova.compute.manager [-] [instance: 59fc11e8-27be-4553-9de9-3eb5ea612ec4] Took 1.04 seconds to deallocate network for instance. [ 541.723284] env[61923]: DEBUG nova.compute.claims [None req-f30f4038-8fd0-4bbd-8b35-7fb8e0dea3a4 tempest-SecurityGroupsTestJSON-509490211 tempest-SecurityGroupsTestJSON-509490211-project-member] [instance: 59fc11e8-27be-4553-9de9-3eb5ea612ec4] Aborting claim: {{(pid=61923) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 541.723767] env[61923]: DEBUG oslo_concurrency.lockutils [None req-f30f4038-8fd0-4bbd-8b35-7fb8e0dea3a4 tempest-SecurityGroupsTestJSON-509490211 tempest-SecurityGroupsTestJSON-509490211-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 541.995330] env[61923]: DEBUG oslo_concurrency.lockutils [None req-464a16f8-1845-4d4d-b3ea-fd8ba00de8bb tempest-ServersAdminNegativeTestJSON-1828705899 tempest-ServersAdminNegativeTestJSON-1828705899-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.050s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 541.997836] env[61923]: ERROR nova.compute.manager [None req-464a16f8-1845-4d4d-b3ea-fd8ba00de8bb tempest-ServersAdminNegativeTestJSON-1828705899 tempest-ServersAdminNegativeTestJSON-1828705899-project-member] [instance: 1ade1b37-8319-4d74-9544-6b83e8738498] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port c188410e-a247-4617-a118-46ee79a2683c, please check neutron logs for more information. [ 541.997836] env[61923]: ERROR nova.compute.manager [instance: 1ade1b37-8319-4d74-9544-6b83e8738498] Traceback (most recent call last): [ 541.997836] env[61923]: ERROR nova.compute.manager [instance: 1ade1b37-8319-4d74-9544-6b83e8738498] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 541.997836] env[61923]: ERROR nova.compute.manager [instance: 1ade1b37-8319-4d74-9544-6b83e8738498] self.driver.spawn(context, instance, image_meta, [ 541.997836] env[61923]: ERROR nova.compute.manager [instance: 1ade1b37-8319-4d74-9544-6b83e8738498] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 541.997836] env[61923]: ERROR nova.compute.manager [instance: 1ade1b37-8319-4d74-9544-6b83e8738498] self._vmops.spawn(context, instance, image_meta, injected_files, [ 541.997836] env[61923]: ERROR nova.compute.manager [instance: 1ade1b37-8319-4d74-9544-6b83e8738498] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 541.997836] env[61923]: ERROR nova.compute.manager [instance: 1ade1b37-8319-4d74-9544-6b83e8738498] vm_ref = self.build_virtual_machine(instance, [ 541.997836] env[61923]: ERROR nova.compute.manager [instance: 1ade1b37-8319-4d74-9544-6b83e8738498] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 541.997836] env[61923]: ERROR nova.compute.manager [instance: 1ade1b37-8319-4d74-9544-6b83e8738498] vif_infos = vmwarevif.get_vif_info(self._session, [ 541.997836] env[61923]: ERROR nova.compute.manager [instance: 1ade1b37-8319-4d74-9544-6b83e8738498] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 541.998260] env[61923]: ERROR nova.compute.manager [instance: 1ade1b37-8319-4d74-9544-6b83e8738498] for vif in network_info: [ 541.998260] env[61923]: ERROR nova.compute.manager [instance: 1ade1b37-8319-4d74-9544-6b83e8738498] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 541.998260] env[61923]: ERROR nova.compute.manager [instance: 1ade1b37-8319-4d74-9544-6b83e8738498] return self._sync_wrapper(fn, *args, **kwargs) [ 541.998260] env[61923]: ERROR nova.compute.manager [instance: 1ade1b37-8319-4d74-9544-6b83e8738498] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 541.998260] env[61923]: ERROR nova.compute.manager [instance: 1ade1b37-8319-4d74-9544-6b83e8738498] self.wait() [ 541.998260] env[61923]: ERROR nova.compute.manager [instance: 1ade1b37-8319-4d74-9544-6b83e8738498] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 541.998260] env[61923]: ERROR nova.compute.manager [instance: 1ade1b37-8319-4d74-9544-6b83e8738498] self[:] = self._gt.wait() [ 541.998260] env[61923]: ERROR nova.compute.manager [instance: 1ade1b37-8319-4d74-9544-6b83e8738498] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 541.998260] env[61923]: ERROR nova.compute.manager [instance: 1ade1b37-8319-4d74-9544-6b83e8738498] return self._exit_event.wait() [ 541.998260] env[61923]: ERROR nova.compute.manager [instance: 1ade1b37-8319-4d74-9544-6b83e8738498] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 541.998260] env[61923]: ERROR nova.compute.manager [instance: 1ade1b37-8319-4d74-9544-6b83e8738498] result = hub.switch() [ 541.998260] env[61923]: ERROR nova.compute.manager [instance: 1ade1b37-8319-4d74-9544-6b83e8738498] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 541.998260] env[61923]: ERROR nova.compute.manager [instance: 1ade1b37-8319-4d74-9544-6b83e8738498] return self.greenlet.switch() [ 541.998620] env[61923]: ERROR nova.compute.manager [instance: 1ade1b37-8319-4d74-9544-6b83e8738498] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 541.998620] env[61923]: ERROR nova.compute.manager [instance: 1ade1b37-8319-4d74-9544-6b83e8738498] result = function(*args, **kwargs) [ 541.998620] env[61923]: ERROR nova.compute.manager [instance: 1ade1b37-8319-4d74-9544-6b83e8738498] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 541.998620] env[61923]: ERROR nova.compute.manager [instance: 1ade1b37-8319-4d74-9544-6b83e8738498] return func(*args, **kwargs) [ 541.998620] env[61923]: ERROR nova.compute.manager [instance: 1ade1b37-8319-4d74-9544-6b83e8738498] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 541.998620] env[61923]: ERROR nova.compute.manager [instance: 1ade1b37-8319-4d74-9544-6b83e8738498] raise e [ 541.998620] env[61923]: ERROR nova.compute.manager [instance: 1ade1b37-8319-4d74-9544-6b83e8738498] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 541.998620] env[61923]: ERROR nova.compute.manager [instance: 1ade1b37-8319-4d74-9544-6b83e8738498] nwinfo = self.network_api.allocate_for_instance( [ 541.998620] env[61923]: ERROR nova.compute.manager [instance: 1ade1b37-8319-4d74-9544-6b83e8738498] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 541.998620] env[61923]: ERROR nova.compute.manager [instance: 1ade1b37-8319-4d74-9544-6b83e8738498] created_port_ids = self._update_ports_for_instance( [ 541.998620] env[61923]: ERROR nova.compute.manager [instance: 1ade1b37-8319-4d74-9544-6b83e8738498] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 541.998620] env[61923]: ERROR nova.compute.manager [instance: 1ade1b37-8319-4d74-9544-6b83e8738498] with excutils.save_and_reraise_exception(): [ 541.998620] env[61923]: ERROR nova.compute.manager [instance: 1ade1b37-8319-4d74-9544-6b83e8738498] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 541.999245] env[61923]: ERROR nova.compute.manager [instance: 1ade1b37-8319-4d74-9544-6b83e8738498] self.force_reraise() [ 541.999245] env[61923]: ERROR nova.compute.manager [instance: 1ade1b37-8319-4d74-9544-6b83e8738498] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 541.999245] env[61923]: ERROR nova.compute.manager [instance: 1ade1b37-8319-4d74-9544-6b83e8738498] raise self.value [ 541.999245] env[61923]: ERROR nova.compute.manager [instance: 1ade1b37-8319-4d74-9544-6b83e8738498] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 541.999245] env[61923]: ERROR nova.compute.manager [instance: 1ade1b37-8319-4d74-9544-6b83e8738498] updated_port = self._update_port( [ 541.999245] env[61923]: ERROR nova.compute.manager [instance: 1ade1b37-8319-4d74-9544-6b83e8738498] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 541.999245] env[61923]: ERROR nova.compute.manager [instance: 1ade1b37-8319-4d74-9544-6b83e8738498] _ensure_no_port_binding_failure(port) [ 541.999245] env[61923]: ERROR nova.compute.manager [instance: 1ade1b37-8319-4d74-9544-6b83e8738498] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 541.999245] env[61923]: ERROR nova.compute.manager [instance: 1ade1b37-8319-4d74-9544-6b83e8738498] raise exception.PortBindingFailed(port_id=port['id']) [ 541.999245] env[61923]: ERROR nova.compute.manager [instance: 1ade1b37-8319-4d74-9544-6b83e8738498] nova.exception.PortBindingFailed: Binding failed for port c188410e-a247-4617-a118-46ee79a2683c, please check neutron logs for more information. [ 541.999245] env[61923]: ERROR nova.compute.manager [instance: 1ade1b37-8319-4d74-9544-6b83e8738498] [ 541.999617] env[61923]: DEBUG nova.compute.utils [None req-464a16f8-1845-4d4d-b3ea-fd8ba00de8bb tempest-ServersAdminNegativeTestJSON-1828705899 tempest-ServersAdminNegativeTestJSON-1828705899-project-member] [instance: 1ade1b37-8319-4d74-9544-6b83e8738498] Binding failed for port c188410e-a247-4617-a118-46ee79a2683c, please check neutron logs for more information. {{(pid=61923) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 541.999617] env[61923]: DEBUG nova.compute.manager [None req-dcb477a6-85d7-4838-be85-b9b2ecf96877 tempest-ImagesOneServerNegativeTestJSON-1560548781 tempest-ImagesOneServerNegativeTestJSON-1560548781-project-member] [instance: ca62f959-43b9-4838-8349-11b7d6a7f0a7] Start spawning the instance on the hypervisor. {{(pid=61923) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 542.001264] env[61923]: DEBUG oslo_concurrency.lockutils [None req-3b10f02e-4025-4174-94c7-8515a7e12175 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 9.053s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 542.002803] env[61923]: INFO nova.compute.claims [None req-3b10f02e-4025-4174-94c7-8515a7e12175 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] [instance: f7cc960c-e06a-4c58-9367-ec8771fe09d5] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 542.007031] env[61923]: DEBUG nova.compute.manager [None req-464a16f8-1845-4d4d-b3ea-fd8ba00de8bb tempest-ServersAdminNegativeTestJSON-1828705899 tempest-ServersAdminNegativeTestJSON-1828705899-project-member] [instance: 1ade1b37-8319-4d74-9544-6b83e8738498] Build of instance 1ade1b37-8319-4d74-9544-6b83e8738498 was re-scheduled: Binding failed for port c188410e-a247-4617-a118-46ee79a2683c, please check neutron logs for more information. {{(pid=61923) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 542.007031] env[61923]: DEBUG nova.compute.manager [None req-464a16f8-1845-4d4d-b3ea-fd8ba00de8bb tempest-ServersAdminNegativeTestJSON-1828705899 tempest-ServersAdminNegativeTestJSON-1828705899-project-member] [instance: 1ade1b37-8319-4d74-9544-6b83e8738498] Unplugging VIFs for instance {{(pid=61923) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 542.007031] env[61923]: DEBUG oslo_concurrency.lockutils [None req-464a16f8-1845-4d4d-b3ea-fd8ba00de8bb tempest-ServersAdminNegativeTestJSON-1828705899 tempest-ServersAdminNegativeTestJSON-1828705899-project-member] Acquiring lock "refresh_cache-1ade1b37-8319-4d74-9544-6b83e8738498" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 542.007031] env[61923]: DEBUG oslo_concurrency.lockutils [None req-464a16f8-1845-4d4d-b3ea-fd8ba00de8bb tempest-ServersAdminNegativeTestJSON-1828705899 tempest-ServersAdminNegativeTestJSON-1828705899-project-member] Acquired lock "refresh_cache-1ade1b37-8319-4d74-9544-6b83e8738498" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 542.007475] env[61923]: DEBUG nova.network.neutron [None req-464a16f8-1845-4d4d-b3ea-fd8ba00de8bb tempest-ServersAdminNegativeTestJSON-1828705899 tempest-ServersAdminNegativeTestJSON-1828705899-project-member] [instance: 1ade1b37-8319-4d74-9544-6b83e8738498] Building network info cache for instance {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 542.007751] env[61923]: DEBUG nova.compute.manager [None req-36da9865-5d98-4c2a-a964-964bee74523c tempest-ListServerFiltersTestJSON-483030130 tempest-ListServerFiltersTestJSON-483030130-project-member] [instance: f3f872ce-cbbe-4407-b4e0-3341bb598825] Starting instance... {{(pid=61923) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 542.051897] env[61923]: DEBUG nova.virt.hardware [None req-dcb477a6-85d7-4838-be85-b9b2ecf96877 tempest-ImagesOneServerNegativeTestJSON-1560548781 tempest-ImagesOneServerNegativeTestJSON-1560548781-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-29T20:07:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-29T20:07:35Z,direct_url=,disk_format='vmdk',id=0f153f63-ae0a-45b1-b7f5-7f9b673c947f,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='4e7b5e3b3c3443c8bd5c70f8a15739f5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-29T20:07:36Z,virtual_size=,visibility=), allow threads: False {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 542.052133] env[61923]: DEBUG nova.virt.hardware [None req-dcb477a6-85d7-4838-be85-b9b2ecf96877 tempest-ImagesOneServerNegativeTestJSON-1560548781 tempest-ImagesOneServerNegativeTestJSON-1560548781-project-member] Flavor limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 542.052282] env[61923]: DEBUG nova.virt.hardware [None req-dcb477a6-85d7-4838-be85-b9b2ecf96877 tempest-ImagesOneServerNegativeTestJSON-1560548781 tempest-ImagesOneServerNegativeTestJSON-1560548781-project-member] Image limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 542.052454] env[61923]: DEBUG nova.virt.hardware [None req-dcb477a6-85d7-4838-be85-b9b2ecf96877 tempest-ImagesOneServerNegativeTestJSON-1560548781 tempest-ImagesOneServerNegativeTestJSON-1560548781-project-member] Flavor pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 542.052592] env[61923]: DEBUG nova.virt.hardware [None req-dcb477a6-85d7-4838-be85-b9b2ecf96877 tempest-ImagesOneServerNegativeTestJSON-1560548781 tempest-ImagesOneServerNegativeTestJSON-1560548781-project-member] Image pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 542.052732] env[61923]: DEBUG nova.virt.hardware [None req-dcb477a6-85d7-4838-be85-b9b2ecf96877 tempest-ImagesOneServerNegativeTestJSON-1560548781 tempest-ImagesOneServerNegativeTestJSON-1560548781-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 542.052931] env[61923]: DEBUG nova.virt.hardware [None req-dcb477a6-85d7-4838-be85-b9b2ecf96877 tempest-ImagesOneServerNegativeTestJSON-1560548781 tempest-ImagesOneServerNegativeTestJSON-1560548781-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 542.054569] env[61923]: DEBUG nova.virt.hardware [None req-dcb477a6-85d7-4838-be85-b9b2ecf96877 tempest-ImagesOneServerNegativeTestJSON-1560548781 tempest-ImagesOneServerNegativeTestJSON-1560548781-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 542.055448] env[61923]: DEBUG nova.virt.hardware [None req-dcb477a6-85d7-4838-be85-b9b2ecf96877 tempest-ImagesOneServerNegativeTestJSON-1560548781 tempest-ImagesOneServerNegativeTestJSON-1560548781-project-member] Got 1 possible topologies {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 542.056577] env[61923]: DEBUG nova.virt.hardware [None req-dcb477a6-85d7-4838-be85-b9b2ecf96877 tempest-ImagesOneServerNegativeTestJSON-1560548781 tempest-ImagesOneServerNegativeTestJSON-1560548781-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 542.056833] env[61923]: DEBUG nova.virt.hardware [None req-dcb477a6-85d7-4838-be85-b9b2ecf96877 tempest-ImagesOneServerNegativeTestJSON-1560548781 tempest-ImagesOneServerNegativeTestJSON-1560548781-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 542.060621] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b1c6ec3-b902-4a86-b516-cc190000aee1 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 542.078082] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65a2fb46-950d-4bea-aee6-8246115318e9 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 542.545652] env[61923]: DEBUG nova.network.neutron [None req-464a16f8-1845-4d4d-b3ea-fd8ba00de8bb tempest-ServersAdminNegativeTestJSON-1828705899 tempest-ServersAdminNegativeTestJSON-1828705899-project-member] [instance: 1ade1b37-8319-4d74-9544-6b83e8738498] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 542.561955] env[61923]: DEBUG oslo_concurrency.lockutils [None req-36da9865-5d98-4c2a-a964-964bee74523c tempest-ListServerFiltersTestJSON-483030130 tempest-ListServerFiltersTestJSON-483030130-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 542.633532] env[61923]: DEBUG nova.compute.manager [req-96477099-7a6e-4aa0-8cb1-9c7846906b54 req-693af2cc-6c0e-431a-a09e-1414a879f5ce service nova] [instance: 5d57303a-ae17-4b92-8f12-303677eb0d7e] Received event network-vif-deleted-4f3fea2e-b3d9-4c89-b291-071aa6707d27 {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 542.633532] env[61923]: DEBUG nova.compute.manager [req-96477099-7a6e-4aa0-8cb1-9c7846906b54 req-693af2cc-6c0e-431a-a09e-1414a879f5ce service nova] [instance: 59fc11e8-27be-4553-9de9-3eb5ea612ec4] Received event network-changed-238265b6-446c-493c-b510-5f51005c1787 {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 542.633532] env[61923]: DEBUG nova.compute.manager [req-96477099-7a6e-4aa0-8cb1-9c7846906b54 req-693af2cc-6c0e-431a-a09e-1414a879f5ce service nova] [instance: 59fc11e8-27be-4553-9de9-3eb5ea612ec4] Refreshing instance network info cache due to event network-changed-238265b6-446c-493c-b510-5f51005c1787. {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 542.633834] env[61923]: DEBUG oslo_concurrency.lockutils [req-96477099-7a6e-4aa0-8cb1-9c7846906b54 req-693af2cc-6c0e-431a-a09e-1414a879f5ce service nova] Acquiring lock "refresh_cache-59fc11e8-27be-4553-9de9-3eb5ea612ec4" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 542.635056] env[61923]: DEBUG oslo_concurrency.lockutils [req-96477099-7a6e-4aa0-8cb1-9c7846906b54 req-693af2cc-6c0e-431a-a09e-1414a879f5ce service nova] Acquired lock "refresh_cache-59fc11e8-27be-4553-9de9-3eb5ea612ec4" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 542.635056] env[61923]: DEBUG nova.network.neutron [req-96477099-7a6e-4aa0-8cb1-9c7846906b54 req-693af2cc-6c0e-431a-a09e-1414a879f5ce service nova] [instance: 59fc11e8-27be-4553-9de9-3eb5ea612ec4] Refreshing network info cache for port 238265b6-446c-493c-b510-5f51005c1787 {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 542.640928] env[61923]: DEBUG nova.network.neutron [None req-464a16f8-1845-4d4d-b3ea-fd8ba00de8bb tempest-ServersAdminNegativeTestJSON-1828705899 tempest-ServersAdminNegativeTestJSON-1828705899-project-member] [instance: 1ade1b37-8319-4d74-9544-6b83e8738498] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 542.808978] env[61923]: DEBUG oslo_concurrency.lockutils [None req-b2fe22fb-da8f-41ba-8845-95fc6177390e tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Acquiring lock "119bb4a7-2d71-44d7-ae81-9238512937c8" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 542.810573] env[61923]: DEBUG oslo_concurrency.lockutils [None req-b2fe22fb-da8f-41ba-8845-95fc6177390e tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Lock "119bb4a7-2d71-44d7-ae81-9238512937c8" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 543.148509] env[61923]: DEBUG oslo_concurrency.lockutils [None req-464a16f8-1845-4d4d-b3ea-fd8ba00de8bb tempest-ServersAdminNegativeTestJSON-1828705899 tempest-ServersAdminNegativeTestJSON-1828705899-project-member] Releasing lock "refresh_cache-1ade1b37-8319-4d74-9544-6b83e8738498" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 543.148509] env[61923]: DEBUG nova.compute.manager [None req-464a16f8-1845-4d4d-b3ea-fd8ba00de8bb tempest-ServersAdminNegativeTestJSON-1828705899 tempest-ServersAdminNegativeTestJSON-1828705899-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61923) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 543.148870] env[61923]: DEBUG nova.compute.manager [None req-464a16f8-1845-4d4d-b3ea-fd8ba00de8bb tempest-ServersAdminNegativeTestJSON-1828705899 tempest-ServersAdminNegativeTestJSON-1828705899-project-member] [instance: 1ade1b37-8319-4d74-9544-6b83e8738498] Deallocating network for instance {{(pid=61923) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 543.148870] env[61923]: DEBUG nova.network.neutron [None req-464a16f8-1845-4d4d-b3ea-fd8ba00de8bb tempest-ServersAdminNegativeTestJSON-1828705899 tempest-ServersAdminNegativeTestJSON-1828705899-project-member] [instance: 1ade1b37-8319-4d74-9544-6b83e8738498] deallocate_for_instance() {{(pid=61923) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 543.186108] env[61923]: DEBUG nova.network.neutron [req-96477099-7a6e-4aa0-8cb1-9c7846906b54 req-693af2cc-6c0e-431a-a09e-1414a879f5ce service nova] [instance: 59fc11e8-27be-4553-9de9-3eb5ea612ec4] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 543.221960] env[61923]: DEBUG nova.network.neutron [None req-464a16f8-1845-4d4d-b3ea-fd8ba00de8bb tempest-ServersAdminNegativeTestJSON-1828705899 tempest-ServersAdminNegativeTestJSON-1828705899-project-member] [instance: 1ade1b37-8319-4d74-9544-6b83e8738498] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 543.388050] env[61923]: ERROR nova.compute.manager [None req-dcb477a6-85d7-4838-be85-b9b2ecf96877 tempest-ImagesOneServerNegativeTestJSON-1560548781 tempest-ImagesOneServerNegativeTestJSON-1560548781-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 63e8ed78-99f3-41e0-b0a6-e6bc6fe00fd4, please check neutron logs for more information. [ 543.388050] env[61923]: ERROR nova.compute.manager Traceback (most recent call last): [ 543.388050] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 543.388050] env[61923]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 543.388050] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 543.388050] env[61923]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 543.388050] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 543.388050] env[61923]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 543.388050] env[61923]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 543.388050] env[61923]: ERROR nova.compute.manager self.force_reraise() [ 543.388050] env[61923]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 543.388050] env[61923]: ERROR nova.compute.manager raise self.value [ 543.388050] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 543.388050] env[61923]: ERROR nova.compute.manager updated_port = self._update_port( [ 543.388050] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 543.388050] env[61923]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 543.388551] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 543.388551] env[61923]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 543.388551] env[61923]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 63e8ed78-99f3-41e0-b0a6-e6bc6fe00fd4, please check neutron logs for more information. [ 543.388551] env[61923]: ERROR nova.compute.manager [ 543.388551] env[61923]: Traceback (most recent call last): [ 543.388551] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 543.388551] env[61923]: listener.cb(fileno) [ 543.388551] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 543.388551] env[61923]: result = function(*args, **kwargs) [ 543.388551] env[61923]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 543.388551] env[61923]: return func(*args, **kwargs) [ 543.388551] env[61923]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 543.388551] env[61923]: raise e [ 543.388551] env[61923]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 543.388551] env[61923]: nwinfo = self.network_api.allocate_for_instance( [ 543.388551] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 543.388551] env[61923]: created_port_ids = self._update_ports_for_instance( [ 543.388551] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 543.388551] env[61923]: with excutils.save_and_reraise_exception(): [ 543.388551] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 543.388551] env[61923]: self.force_reraise() [ 543.388551] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 543.388551] env[61923]: raise self.value [ 543.388551] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 543.388551] env[61923]: updated_port = self._update_port( [ 543.388551] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 543.388551] env[61923]: _ensure_no_port_binding_failure(port) [ 543.388551] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 543.388551] env[61923]: raise exception.PortBindingFailed(port_id=port['id']) [ 543.389333] env[61923]: nova.exception.PortBindingFailed: Binding failed for port 63e8ed78-99f3-41e0-b0a6-e6bc6fe00fd4, please check neutron logs for more information. [ 543.389333] env[61923]: Removing descriptor: 15 [ 543.389333] env[61923]: ERROR nova.compute.manager [None req-dcb477a6-85d7-4838-be85-b9b2ecf96877 tempest-ImagesOneServerNegativeTestJSON-1560548781 tempest-ImagesOneServerNegativeTestJSON-1560548781-project-member] [instance: ca62f959-43b9-4838-8349-11b7d6a7f0a7] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 63e8ed78-99f3-41e0-b0a6-e6bc6fe00fd4, please check neutron logs for more information. [ 543.389333] env[61923]: ERROR nova.compute.manager [instance: ca62f959-43b9-4838-8349-11b7d6a7f0a7] Traceback (most recent call last): [ 543.389333] env[61923]: ERROR nova.compute.manager [instance: ca62f959-43b9-4838-8349-11b7d6a7f0a7] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 543.389333] env[61923]: ERROR nova.compute.manager [instance: ca62f959-43b9-4838-8349-11b7d6a7f0a7] yield resources [ 543.389333] env[61923]: ERROR nova.compute.manager [instance: ca62f959-43b9-4838-8349-11b7d6a7f0a7] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 543.389333] env[61923]: ERROR nova.compute.manager [instance: ca62f959-43b9-4838-8349-11b7d6a7f0a7] self.driver.spawn(context, instance, image_meta, [ 543.389333] env[61923]: ERROR nova.compute.manager [instance: ca62f959-43b9-4838-8349-11b7d6a7f0a7] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 543.389333] env[61923]: ERROR nova.compute.manager [instance: ca62f959-43b9-4838-8349-11b7d6a7f0a7] self._vmops.spawn(context, instance, image_meta, injected_files, [ 543.389333] env[61923]: ERROR nova.compute.manager [instance: ca62f959-43b9-4838-8349-11b7d6a7f0a7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 543.389333] env[61923]: ERROR nova.compute.manager [instance: ca62f959-43b9-4838-8349-11b7d6a7f0a7] vm_ref = self.build_virtual_machine(instance, [ 543.389665] env[61923]: ERROR nova.compute.manager [instance: ca62f959-43b9-4838-8349-11b7d6a7f0a7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 543.389665] env[61923]: ERROR nova.compute.manager [instance: ca62f959-43b9-4838-8349-11b7d6a7f0a7] vif_infos = vmwarevif.get_vif_info(self._session, [ 543.389665] env[61923]: ERROR nova.compute.manager [instance: ca62f959-43b9-4838-8349-11b7d6a7f0a7] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 543.389665] env[61923]: ERROR nova.compute.manager [instance: ca62f959-43b9-4838-8349-11b7d6a7f0a7] for vif in network_info: [ 543.389665] env[61923]: ERROR nova.compute.manager [instance: ca62f959-43b9-4838-8349-11b7d6a7f0a7] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 543.389665] env[61923]: ERROR nova.compute.manager [instance: ca62f959-43b9-4838-8349-11b7d6a7f0a7] return self._sync_wrapper(fn, *args, **kwargs) [ 543.389665] env[61923]: ERROR nova.compute.manager [instance: ca62f959-43b9-4838-8349-11b7d6a7f0a7] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 543.389665] env[61923]: ERROR nova.compute.manager [instance: ca62f959-43b9-4838-8349-11b7d6a7f0a7] self.wait() [ 543.389665] env[61923]: ERROR nova.compute.manager [instance: ca62f959-43b9-4838-8349-11b7d6a7f0a7] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 543.389665] env[61923]: ERROR nova.compute.manager [instance: ca62f959-43b9-4838-8349-11b7d6a7f0a7] self[:] = self._gt.wait() [ 543.389665] env[61923]: ERROR nova.compute.manager [instance: ca62f959-43b9-4838-8349-11b7d6a7f0a7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 543.389665] env[61923]: ERROR nova.compute.manager [instance: ca62f959-43b9-4838-8349-11b7d6a7f0a7] return self._exit_event.wait() [ 543.389665] env[61923]: ERROR nova.compute.manager [instance: ca62f959-43b9-4838-8349-11b7d6a7f0a7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 543.390043] env[61923]: ERROR nova.compute.manager [instance: ca62f959-43b9-4838-8349-11b7d6a7f0a7] result = hub.switch() [ 543.390043] env[61923]: ERROR nova.compute.manager [instance: ca62f959-43b9-4838-8349-11b7d6a7f0a7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 543.390043] env[61923]: ERROR nova.compute.manager [instance: ca62f959-43b9-4838-8349-11b7d6a7f0a7] return self.greenlet.switch() [ 543.390043] env[61923]: ERROR nova.compute.manager [instance: ca62f959-43b9-4838-8349-11b7d6a7f0a7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 543.390043] env[61923]: ERROR nova.compute.manager [instance: ca62f959-43b9-4838-8349-11b7d6a7f0a7] result = function(*args, **kwargs) [ 543.390043] env[61923]: ERROR nova.compute.manager [instance: ca62f959-43b9-4838-8349-11b7d6a7f0a7] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 543.390043] env[61923]: ERROR nova.compute.manager [instance: ca62f959-43b9-4838-8349-11b7d6a7f0a7] return func(*args, **kwargs) [ 543.390043] env[61923]: ERROR nova.compute.manager [instance: ca62f959-43b9-4838-8349-11b7d6a7f0a7] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 543.390043] env[61923]: ERROR nova.compute.manager [instance: ca62f959-43b9-4838-8349-11b7d6a7f0a7] raise e [ 543.390043] env[61923]: ERROR nova.compute.manager [instance: ca62f959-43b9-4838-8349-11b7d6a7f0a7] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 543.390043] env[61923]: ERROR nova.compute.manager [instance: ca62f959-43b9-4838-8349-11b7d6a7f0a7] nwinfo = self.network_api.allocate_for_instance( [ 543.390043] env[61923]: ERROR nova.compute.manager [instance: ca62f959-43b9-4838-8349-11b7d6a7f0a7] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 543.390043] env[61923]: ERROR nova.compute.manager [instance: ca62f959-43b9-4838-8349-11b7d6a7f0a7] created_port_ids = self._update_ports_for_instance( [ 543.390407] env[61923]: ERROR nova.compute.manager [instance: ca62f959-43b9-4838-8349-11b7d6a7f0a7] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 543.390407] env[61923]: ERROR nova.compute.manager [instance: ca62f959-43b9-4838-8349-11b7d6a7f0a7] with excutils.save_and_reraise_exception(): [ 543.390407] env[61923]: ERROR nova.compute.manager [instance: ca62f959-43b9-4838-8349-11b7d6a7f0a7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 543.390407] env[61923]: ERROR nova.compute.manager [instance: ca62f959-43b9-4838-8349-11b7d6a7f0a7] self.force_reraise() [ 543.390407] env[61923]: ERROR nova.compute.manager [instance: ca62f959-43b9-4838-8349-11b7d6a7f0a7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 543.390407] env[61923]: ERROR nova.compute.manager [instance: ca62f959-43b9-4838-8349-11b7d6a7f0a7] raise self.value [ 543.390407] env[61923]: ERROR nova.compute.manager [instance: ca62f959-43b9-4838-8349-11b7d6a7f0a7] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 543.390407] env[61923]: ERROR nova.compute.manager [instance: ca62f959-43b9-4838-8349-11b7d6a7f0a7] updated_port = self._update_port( [ 543.390407] env[61923]: ERROR nova.compute.manager [instance: ca62f959-43b9-4838-8349-11b7d6a7f0a7] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 543.390407] env[61923]: ERROR nova.compute.manager [instance: ca62f959-43b9-4838-8349-11b7d6a7f0a7] _ensure_no_port_binding_failure(port) [ 543.390407] env[61923]: ERROR nova.compute.manager [instance: ca62f959-43b9-4838-8349-11b7d6a7f0a7] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 543.390407] env[61923]: ERROR nova.compute.manager [instance: ca62f959-43b9-4838-8349-11b7d6a7f0a7] raise exception.PortBindingFailed(port_id=port['id']) [ 543.390916] env[61923]: ERROR nova.compute.manager [instance: ca62f959-43b9-4838-8349-11b7d6a7f0a7] nova.exception.PortBindingFailed: Binding failed for port 63e8ed78-99f3-41e0-b0a6-e6bc6fe00fd4, please check neutron logs for more information. [ 543.390916] env[61923]: ERROR nova.compute.manager [instance: ca62f959-43b9-4838-8349-11b7d6a7f0a7] [ 543.390916] env[61923]: INFO nova.compute.manager [None req-dcb477a6-85d7-4838-be85-b9b2ecf96877 tempest-ImagesOneServerNegativeTestJSON-1560548781 tempest-ImagesOneServerNegativeTestJSON-1560548781-project-member] [instance: ca62f959-43b9-4838-8349-11b7d6a7f0a7] Terminating instance [ 543.396533] env[61923]: DEBUG oslo_concurrency.lockutils [None req-dcb477a6-85d7-4838-be85-b9b2ecf96877 tempest-ImagesOneServerNegativeTestJSON-1560548781 tempest-ImagesOneServerNegativeTestJSON-1560548781-project-member] Acquiring lock "refresh_cache-ca62f959-43b9-4838-8349-11b7d6a7f0a7" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 543.396533] env[61923]: DEBUG oslo_concurrency.lockutils [None req-dcb477a6-85d7-4838-be85-b9b2ecf96877 tempest-ImagesOneServerNegativeTestJSON-1560548781 tempest-ImagesOneServerNegativeTestJSON-1560548781-project-member] Acquired lock "refresh_cache-ca62f959-43b9-4838-8349-11b7d6a7f0a7" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 543.396533] env[61923]: DEBUG nova.network.neutron [None req-dcb477a6-85d7-4838-be85-b9b2ecf96877 tempest-ImagesOneServerNegativeTestJSON-1560548781 tempest-ImagesOneServerNegativeTestJSON-1560548781-project-member] [instance: ca62f959-43b9-4838-8349-11b7d6a7f0a7] Building network info cache for instance {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 543.398328] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a53d4d54-0f56-4ba6-98a6-721e0575a4d5 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 543.407080] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bedbdc71-7c2c-4651-9567-8784d45ba5fc {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 543.445348] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-377e6dd5-ac03-4acf-8ee6-eb345a1c839c {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 543.454564] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b09a550-f0e4-4e01-9ff3-68b15c4f27e0 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 543.470428] env[61923]: DEBUG nova.compute.provider_tree [None req-3b10f02e-4025-4174-94c7-8515a7e12175 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 543.724229] env[61923]: DEBUG nova.network.neutron [None req-464a16f8-1845-4d4d-b3ea-fd8ba00de8bb tempest-ServersAdminNegativeTestJSON-1828705899 tempest-ServersAdminNegativeTestJSON-1828705899-project-member] [instance: 1ade1b37-8319-4d74-9544-6b83e8738498] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 543.741198] env[61923]: DEBUG nova.network.neutron [req-96477099-7a6e-4aa0-8cb1-9c7846906b54 req-693af2cc-6c0e-431a-a09e-1414a879f5ce service nova] [instance: 59fc11e8-27be-4553-9de9-3eb5ea612ec4] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 543.925192] env[61923]: DEBUG nova.network.neutron [None req-dcb477a6-85d7-4838-be85-b9b2ecf96877 tempest-ImagesOneServerNegativeTestJSON-1560548781 tempest-ImagesOneServerNegativeTestJSON-1560548781-project-member] [instance: ca62f959-43b9-4838-8349-11b7d6a7f0a7] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 543.976334] env[61923]: DEBUG nova.scheduler.client.report [None req-3b10f02e-4025-4174-94c7-8515a7e12175 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 544.124959] env[61923]: DEBUG nova.network.neutron [None req-dcb477a6-85d7-4838-be85-b9b2ecf96877 tempest-ImagesOneServerNegativeTestJSON-1560548781 tempest-ImagesOneServerNegativeTestJSON-1560548781-project-member] [instance: ca62f959-43b9-4838-8349-11b7d6a7f0a7] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 544.234023] env[61923]: INFO nova.compute.manager [None req-464a16f8-1845-4d4d-b3ea-fd8ba00de8bb tempest-ServersAdminNegativeTestJSON-1828705899 tempest-ServersAdminNegativeTestJSON-1828705899-project-member] [instance: 1ade1b37-8319-4d74-9544-6b83e8738498] Took 1.08 seconds to deallocate network for instance. [ 544.245201] env[61923]: DEBUG oslo_concurrency.lockutils [req-96477099-7a6e-4aa0-8cb1-9c7846906b54 req-693af2cc-6c0e-431a-a09e-1414a879f5ce service nova] Releasing lock "refresh_cache-59fc11e8-27be-4553-9de9-3eb5ea612ec4" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 544.245201] env[61923]: DEBUG nova.compute.manager [req-96477099-7a6e-4aa0-8cb1-9c7846906b54 req-693af2cc-6c0e-431a-a09e-1414a879f5ce service nova] [instance: 59fc11e8-27be-4553-9de9-3eb5ea612ec4] Received event network-vif-deleted-238265b6-446c-493c-b510-5f51005c1787 {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 544.295633] env[61923]: DEBUG nova.compute.manager [req-15d197ed-2fb1-43b7-8bf6-e29ee89a6f49 req-f7a9f4da-6279-4d0b-a9f0-5d97e198bbda service nova] [instance: ca62f959-43b9-4838-8349-11b7d6a7f0a7] Received event network-changed-63e8ed78-99f3-41e0-b0a6-e6bc6fe00fd4 {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 544.295633] env[61923]: DEBUG nova.compute.manager [req-15d197ed-2fb1-43b7-8bf6-e29ee89a6f49 req-f7a9f4da-6279-4d0b-a9f0-5d97e198bbda service nova] [instance: ca62f959-43b9-4838-8349-11b7d6a7f0a7] Refreshing instance network info cache due to event network-changed-63e8ed78-99f3-41e0-b0a6-e6bc6fe00fd4. {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 544.295633] env[61923]: DEBUG oslo_concurrency.lockutils [req-15d197ed-2fb1-43b7-8bf6-e29ee89a6f49 req-f7a9f4da-6279-4d0b-a9f0-5d97e198bbda service nova] Acquiring lock "refresh_cache-ca62f959-43b9-4838-8349-11b7d6a7f0a7" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 544.481708] env[61923]: DEBUG oslo_concurrency.lockutils [None req-3b10f02e-4025-4174-94c7-8515a7e12175 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.480s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 544.483512] env[61923]: DEBUG nova.compute.manager [None req-3b10f02e-4025-4174-94c7-8515a7e12175 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] [instance: f7cc960c-e06a-4c58-9367-ec8771fe09d5] Start building networks asynchronously for instance. {{(pid=61923) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 544.485020] env[61923]: DEBUG oslo_concurrency.lockutils [None req-d9e2f878-40e6-4b41-8b07-da611f2f81f1 tempest-VolumesAssistedSnapshotsTest-2043366068 tempest-VolumesAssistedSnapshotsTest-2043366068-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 10.240s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 544.489317] env[61923]: INFO nova.compute.claims [None req-d9e2f878-40e6-4b41-8b07-da611f2f81f1 tempest-VolumesAssistedSnapshotsTest-2043366068 tempest-VolumesAssistedSnapshotsTest-2043366068-project-member] [instance: 06bda5ee-c7a8-47d1-9e0c-a5af67d40006] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 544.627670] env[61923]: DEBUG oslo_concurrency.lockutils [None req-dcb477a6-85d7-4838-be85-b9b2ecf96877 tempest-ImagesOneServerNegativeTestJSON-1560548781 tempest-ImagesOneServerNegativeTestJSON-1560548781-project-member] Releasing lock "refresh_cache-ca62f959-43b9-4838-8349-11b7d6a7f0a7" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 544.628135] env[61923]: DEBUG nova.compute.manager [None req-dcb477a6-85d7-4838-be85-b9b2ecf96877 tempest-ImagesOneServerNegativeTestJSON-1560548781 tempest-ImagesOneServerNegativeTestJSON-1560548781-project-member] [instance: ca62f959-43b9-4838-8349-11b7d6a7f0a7] Start destroying the instance on the hypervisor. {{(pid=61923) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 544.628671] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-dcb477a6-85d7-4838-be85-b9b2ecf96877 tempest-ImagesOneServerNegativeTestJSON-1560548781 tempest-ImagesOneServerNegativeTestJSON-1560548781-project-member] [instance: ca62f959-43b9-4838-8349-11b7d6a7f0a7] Destroying instance {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 544.630940] env[61923]: DEBUG oslo_concurrency.lockutils [req-15d197ed-2fb1-43b7-8bf6-e29ee89a6f49 req-f7a9f4da-6279-4d0b-a9f0-5d97e198bbda service nova] Acquired lock "refresh_cache-ca62f959-43b9-4838-8349-11b7d6a7f0a7" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 544.630940] env[61923]: DEBUG nova.network.neutron [req-15d197ed-2fb1-43b7-8bf6-e29ee89a6f49 req-f7a9f4da-6279-4d0b-a9f0-5d97e198bbda service nova] [instance: ca62f959-43b9-4838-8349-11b7d6a7f0a7] Refreshing network info cache for port 63e8ed78-99f3-41e0-b0a6-e6bc6fe00fd4 {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 544.632033] env[61923]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-44dde225-c5f7-4a99-8f6d-05148e5b7169 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 544.642902] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-520efdf6-ddce-4e50-a134-7c16ab634ab6 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 544.667738] env[61923]: WARNING nova.virt.vmwareapi.vmops [None req-dcb477a6-85d7-4838-be85-b9b2ecf96877 tempest-ImagesOneServerNegativeTestJSON-1560548781 tempest-ImagesOneServerNegativeTestJSON-1560548781-project-member] [instance: ca62f959-43b9-4838-8349-11b7d6a7f0a7] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance ca62f959-43b9-4838-8349-11b7d6a7f0a7 could not be found. [ 544.667968] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-dcb477a6-85d7-4838-be85-b9b2ecf96877 tempest-ImagesOneServerNegativeTestJSON-1560548781 tempest-ImagesOneServerNegativeTestJSON-1560548781-project-member] [instance: ca62f959-43b9-4838-8349-11b7d6a7f0a7] Instance destroyed {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 544.669145] env[61923]: INFO nova.compute.manager [None req-dcb477a6-85d7-4838-be85-b9b2ecf96877 tempest-ImagesOneServerNegativeTestJSON-1560548781 tempest-ImagesOneServerNegativeTestJSON-1560548781-project-member] [instance: ca62f959-43b9-4838-8349-11b7d6a7f0a7] Took 0.04 seconds to destroy the instance on the hypervisor. [ 544.669145] env[61923]: DEBUG oslo.service.loopingcall [None req-dcb477a6-85d7-4838-be85-b9b2ecf96877 tempest-ImagesOneServerNegativeTestJSON-1560548781 tempest-ImagesOneServerNegativeTestJSON-1560548781-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61923) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 544.669145] env[61923]: DEBUG nova.compute.manager [-] [instance: ca62f959-43b9-4838-8349-11b7d6a7f0a7] Deallocating network for instance {{(pid=61923) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 544.669145] env[61923]: DEBUG nova.network.neutron [-] [instance: ca62f959-43b9-4838-8349-11b7d6a7f0a7] deallocate_for_instance() {{(pid=61923) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 544.705984] env[61923]: DEBUG nova.network.neutron [-] [instance: ca62f959-43b9-4838-8349-11b7d6a7f0a7] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 544.919728] env[61923]: DEBUG oslo_concurrency.lockutils [None req-aa9c464b-ac6b-44fe-9384-28cd9bdfecbc tempest-ImagesOneServerTestJSON-63657725 tempest-ImagesOneServerTestJSON-63657725-project-member] Acquiring lock "0c1b8117-199f-4101-ab81-4ae7ef0d1251" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 544.920128] env[61923]: DEBUG oslo_concurrency.lockutils [None req-aa9c464b-ac6b-44fe-9384-28cd9bdfecbc tempest-ImagesOneServerTestJSON-63657725 tempest-ImagesOneServerTestJSON-63657725-project-member] Lock "0c1b8117-199f-4101-ab81-4ae7ef0d1251" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 544.991597] env[61923]: DEBUG nova.compute.utils [None req-3b10f02e-4025-4174-94c7-8515a7e12175 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] Using /dev/sd instead of None {{(pid=61923) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 544.994720] env[61923]: DEBUG nova.compute.manager [None req-3b10f02e-4025-4174-94c7-8515a7e12175 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] [instance: f7cc960c-e06a-4c58-9367-ec8771fe09d5] Not allocating networking since 'none' was specified. {{(pid=61923) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 545.162281] env[61923]: DEBUG nova.network.neutron [req-15d197ed-2fb1-43b7-8bf6-e29ee89a6f49 req-f7a9f4da-6279-4d0b-a9f0-5d97e198bbda service nova] [instance: ca62f959-43b9-4838-8349-11b7d6a7f0a7] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 545.209142] env[61923]: DEBUG nova.network.neutron [-] [instance: ca62f959-43b9-4838-8349-11b7d6a7f0a7] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 545.251918] env[61923]: DEBUG nova.network.neutron [req-15d197ed-2fb1-43b7-8bf6-e29ee89a6f49 req-f7a9f4da-6279-4d0b-a9f0-5d97e198bbda service nova] [instance: ca62f959-43b9-4838-8349-11b7d6a7f0a7] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 545.284785] env[61923]: INFO nova.scheduler.client.report [None req-464a16f8-1845-4d4d-b3ea-fd8ba00de8bb tempest-ServersAdminNegativeTestJSON-1828705899 tempest-ServersAdminNegativeTestJSON-1828705899-project-member] Deleted allocations for instance 1ade1b37-8319-4d74-9544-6b83e8738498 [ 545.496074] env[61923]: DEBUG nova.compute.manager [None req-3b10f02e-4025-4174-94c7-8515a7e12175 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] [instance: f7cc960c-e06a-4c58-9367-ec8771fe09d5] Start building block device mappings for instance. {{(pid=61923) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 545.714520] env[61923]: INFO nova.compute.manager [-] [instance: ca62f959-43b9-4838-8349-11b7d6a7f0a7] Took 1.05 seconds to deallocate network for instance. [ 545.718199] env[61923]: DEBUG nova.compute.claims [None req-dcb477a6-85d7-4838-be85-b9b2ecf96877 tempest-ImagesOneServerNegativeTestJSON-1560548781 tempest-ImagesOneServerNegativeTestJSON-1560548781-project-member] [instance: ca62f959-43b9-4838-8349-11b7d6a7f0a7] Aborting claim: {{(pid=61923) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 545.718442] env[61923]: DEBUG oslo_concurrency.lockutils [None req-dcb477a6-85d7-4838-be85-b9b2ecf96877 tempest-ImagesOneServerNegativeTestJSON-1560548781 tempest-ImagesOneServerNegativeTestJSON-1560548781-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 545.758296] env[61923]: DEBUG oslo_concurrency.lockutils [req-15d197ed-2fb1-43b7-8bf6-e29ee89a6f49 req-f7a9f4da-6279-4d0b-a9f0-5d97e198bbda service nova] Releasing lock "refresh_cache-ca62f959-43b9-4838-8349-11b7d6a7f0a7" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 545.795370] env[61923]: DEBUG oslo_concurrency.lockutils [None req-464a16f8-1845-4d4d-b3ea-fd8ba00de8bb tempest-ServersAdminNegativeTestJSON-1828705899 tempest-ServersAdminNegativeTestJSON-1828705899-project-member] Lock "1ade1b37-8319-4d74-9544-6b83e8738498" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 28.578s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 545.849494] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7258321c-0d0f-44d9-9fbf-bd49acbc5eee {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 545.861323] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a437955-6757-4d68-b4a8-209df12bba11 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 545.896982] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ffcbfee6-bd38-4936-9439-3a33ad7df61e {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 545.910427] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-512c2098-e910-41f8-a8c6-8fc5528a05a8 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 545.922929] env[61923]: DEBUG nova.compute.provider_tree [None req-d9e2f878-40e6-4b41-8b07-da611f2f81f1 tempest-VolumesAssistedSnapshotsTest-2043366068 tempest-VolumesAssistedSnapshotsTest-2043366068-project-member] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 546.301212] env[61923]: DEBUG nova.compute.manager [None req-fba27a13-474d-4fe3-9187-c31143f40aa5 tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] [instance: 6a2d7ede-e0cc-4a36-b1b2-0b36b269ecb5] Starting instance... {{(pid=61923) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 546.429976] env[61923]: DEBUG nova.scheduler.client.report [None req-d9e2f878-40e6-4b41-8b07-da611f2f81f1 tempest-VolumesAssistedSnapshotsTest-2043366068 tempest-VolumesAssistedSnapshotsTest-2043366068-project-member] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 546.471717] env[61923]: DEBUG oslo_concurrency.lockutils [None req-0c55849d-766e-4fac-b1c4-e4e97ba688a6 tempest-DeleteServersAdminTestJSON-590116201 tempest-DeleteServersAdminTestJSON-590116201-project-member] Acquiring lock "1b509288-2df3-4927-afcf-cb76918d3da3" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 546.471717] env[61923]: DEBUG oslo_concurrency.lockutils [None req-0c55849d-766e-4fac-b1c4-e4e97ba688a6 tempest-DeleteServersAdminTestJSON-590116201 tempest-DeleteServersAdminTestJSON-590116201-project-member] Lock "1b509288-2df3-4927-afcf-cb76918d3da3" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 546.510035] env[61923]: DEBUG nova.compute.manager [None req-3b10f02e-4025-4174-94c7-8515a7e12175 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] [instance: f7cc960c-e06a-4c58-9367-ec8771fe09d5] Start spawning the instance on the hypervisor. {{(pid=61923) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 546.539753] env[61923]: DEBUG nova.virt.hardware [None req-3b10f02e-4025-4174-94c7-8515a7e12175 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-29T20:07:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-29T20:07:35Z,direct_url=,disk_format='vmdk',id=0f153f63-ae0a-45b1-b7f5-7f9b673c947f,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='4e7b5e3b3c3443c8bd5c70f8a15739f5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-29T20:07:36Z,virtual_size=,visibility=), allow threads: False {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 546.540345] env[61923]: DEBUG nova.virt.hardware [None req-3b10f02e-4025-4174-94c7-8515a7e12175 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] Flavor limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 546.540660] env[61923]: DEBUG nova.virt.hardware [None req-3b10f02e-4025-4174-94c7-8515a7e12175 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] Image limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 546.541225] env[61923]: DEBUG nova.virt.hardware [None req-3b10f02e-4025-4174-94c7-8515a7e12175 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] Flavor pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 546.541225] env[61923]: DEBUG nova.virt.hardware [None req-3b10f02e-4025-4174-94c7-8515a7e12175 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] Image pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 546.541541] env[61923]: DEBUG nova.virt.hardware [None req-3b10f02e-4025-4174-94c7-8515a7e12175 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 546.541602] env[61923]: DEBUG nova.virt.hardware [None req-3b10f02e-4025-4174-94c7-8515a7e12175 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 546.541901] env[61923]: DEBUG nova.virt.hardware [None req-3b10f02e-4025-4174-94c7-8515a7e12175 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 546.542110] env[61923]: DEBUG nova.virt.hardware [None req-3b10f02e-4025-4174-94c7-8515a7e12175 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] Got 1 possible topologies {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 546.542285] env[61923]: DEBUG nova.virt.hardware [None req-3b10f02e-4025-4174-94c7-8515a7e12175 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 546.542455] env[61923]: DEBUG nova.virt.hardware [None req-3b10f02e-4025-4174-94c7-8515a7e12175 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 546.545148] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8d58566-39da-43d4-80cc-41f76964c913 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 546.554676] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7660e1e-d668-4ca7-bbbe-775e047111d0 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 546.569965] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-3b10f02e-4025-4174-94c7-8515a7e12175 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] [instance: f7cc960c-e06a-4c58-9367-ec8771fe09d5] Instance VIF info [] {{(pid=61923) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 546.579365] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-3b10f02e-4025-4174-94c7-8515a7e12175 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] Creating folder: OpenStack. Parent ref: group-v4. {{(pid=61923) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 546.579706] env[61923]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-ed1b3c03-d703-4140-84a1-b390a5350a90 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 546.592373] env[61923]: INFO nova.virt.vmwareapi.vm_util [None req-3b10f02e-4025-4174-94c7-8515a7e12175 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] Created folder: OpenStack in parent group-v4. [ 546.592373] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-3b10f02e-4025-4174-94c7-8515a7e12175 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] Creating folder: Project (6be832751b484c6e898fc7f6d49c69c8). Parent ref: group-v292629. {{(pid=61923) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 546.592511] env[61923]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-840a7071-6585-414a-880c-e0fa0c7cd0ec {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 546.602906] env[61923]: INFO nova.virt.vmwareapi.vm_util [None req-3b10f02e-4025-4174-94c7-8515a7e12175 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] Created folder: Project (6be832751b484c6e898fc7f6d49c69c8) in parent group-v292629. [ 546.602906] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-3b10f02e-4025-4174-94c7-8515a7e12175 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] Creating folder: Instances. Parent ref: group-v292630. {{(pid=61923) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 546.603207] env[61923]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-86fc6a26-9cfa-4c5e-b974-e41a4f012b41 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 546.614016] env[61923]: INFO nova.virt.vmwareapi.vm_util [None req-3b10f02e-4025-4174-94c7-8515a7e12175 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] Created folder: Instances in parent group-v292630. [ 546.614016] env[61923]: DEBUG oslo.service.loopingcall [None req-3b10f02e-4025-4174-94c7-8515a7e12175 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61923) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 546.614016] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f7cc960c-e06a-4c58-9367-ec8771fe09d5] Creating VM on the ESX host {{(pid=61923) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 546.614016] env[61923]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-2cc05f36-e99c-42b5-85ef-28f498ceec81 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 546.631049] env[61923]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 546.631049] env[61923]: value = "task-1377425" [ 546.631049] env[61923]: _type = "Task" [ 546.631049] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 546.638534] env[61923]: DEBUG oslo_vmware.api [-] Task: {'id': task-1377425, 'name': CreateVM_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 546.832249] env[61923]: DEBUG oslo_concurrency.lockutils [None req-fba27a13-474d-4fe3-9187-c31143f40aa5 tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 546.872713] env[61923]: DEBUG oslo_concurrency.lockutils [None req-95b9b70a-a835-4901-9a6d-c0fa1b2217f3 tempest-MigrationsAdminTest-1311988845 tempest-MigrationsAdminTest-1311988845-project-member] Acquiring lock "a4b54a7a-3f3d-49ae-ba43-5a10c262d4e5" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 546.872933] env[61923]: DEBUG oslo_concurrency.lockutils [None req-95b9b70a-a835-4901-9a6d-c0fa1b2217f3 tempest-MigrationsAdminTest-1311988845 tempest-MigrationsAdminTest-1311988845-project-member] Lock "a4b54a7a-3f3d-49ae-ba43-5a10c262d4e5" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 546.937692] env[61923]: DEBUG oslo_concurrency.lockutils [None req-d9e2f878-40e6-4b41-8b07-da611f2f81f1 tempest-VolumesAssistedSnapshotsTest-2043366068 tempest-VolumesAssistedSnapshotsTest-2043366068-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.453s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 546.938512] env[61923]: DEBUG nova.compute.manager [None req-d9e2f878-40e6-4b41-8b07-da611f2f81f1 tempest-VolumesAssistedSnapshotsTest-2043366068 tempest-VolumesAssistedSnapshotsTest-2043366068-project-member] [instance: 06bda5ee-c7a8-47d1-9e0c-a5af67d40006] Start building networks asynchronously for instance. {{(pid=61923) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 546.941864] env[61923]: DEBUG oslo_concurrency.lockutils [None req-8a6b6bc2-54c9-4389-ab30-057158d86227 tempest-ServerDiagnosticsNegativeTest-664316905 tempest-ServerDiagnosticsNegativeTest-664316905-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 11.996s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 547.141095] env[61923]: DEBUG oslo_vmware.api [-] Task: {'id': task-1377425, 'name': CreateVM_Task, 'duration_secs': 0.305459} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 547.141333] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f7cc960c-e06a-4c58-9367-ec8771fe09d5] Created VM on the ESX host {{(pid=61923) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 547.142325] env[61923]: DEBUG oslo_vmware.service [None req-3b10f02e-4025-4174-94c7-8515a7e12175 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d61fe763-ee8b-4a0b-aa6e-4346168b5d1f {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 547.147928] env[61923]: DEBUG oslo_concurrency.lockutils [None req-3b10f02e-4025-4174-94c7-8515a7e12175 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 547.148106] env[61923]: DEBUG oslo_concurrency.lockutils [None req-3b10f02e-4025-4174-94c7-8515a7e12175 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 547.149075] env[61923]: DEBUG oslo_concurrency.lockutils [None req-3b10f02e-4025-4174-94c7-8515a7e12175 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 547.149075] env[61923]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f9630ceb-71c5-4332-a60b-0f17ee0f8f0f {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 547.155244] env[61923]: DEBUG oslo_vmware.api [None req-3b10f02e-4025-4174-94c7-8515a7e12175 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] Waiting for the task: (returnval){ [ 547.155244] env[61923]: value = "session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52a682d2-dc52-8abe-3809-9b670ffd37f9" [ 547.155244] env[61923]: _type = "Task" [ 547.155244] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 547.165560] env[61923]: DEBUG oslo_vmware.api [None req-3b10f02e-4025-4174-94c7-8515a7e12175 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52a682d2-dc52-8abe-3809-9b670ffd37f9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 547.213150] env[61923]: DEBUG nova.compute.manager [req-81b240bc-bf85-4e06-b287-7182cbb7d537 req-a0abf508-ddf1-44a3-aeee-641652d29d0b service nova] [instance: ca62f959-43b9-4838-8349-11b7d6a7f0a7] Received event network-vif-deleted-63e8ed78-99f3-41e0-b0a6-e6bc6fe00fd4 {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 547.447266] env[61923]: DEBUG nova.compute.utils [None req-d9e2f878-40e6-4b41-8b07-da611f2f81f1 tempest-VolumesAssistedSnapshotsTest-2043366068 tempest-VolumesAssistedSnapshotsTest-2043366068-project-member] Using /dev/sd instead of None {{(pid=61923) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 547.455970] env[61923]: DEBUG nova.compute.manager [None req-d9e2f878-40e6-4b41-8b07-da611f2f81f1 tempest-VolumesAssistedSnapshotsTest-2043366068 tempest-VolumesAssistedSnapshotsTest-2043366068-project-member] [instance: 06bda5ee-c7a8-47d1-9e0c-a5af67d40006] Allocating IP information in the background. {{(pid=61923) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 547.456198] env[61923]: DEBUG nova.network.neutron [None req-d9e2f878-40e6-4b41-8b07-da611f2f81f1 tempest-VolumesAssistedSnapshotsTest-2043366068 tempest-VolumesAssistedSnapshotsTest-2043366068-project-member] [instance: 06bda5ee-c7a8-47d1-9e0c-a5af67d40006] allocate_for_instance() {{(pid=61923) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 547.530484] env[61923]: DEBUG nova.policy [None req-d9e2f878-40e6-4b41-8b07-da611f2f81f1 tempest-VolumesAssistedSnapshotsTest-2043366068 tempest-VolumesAssistedSnapshotsTest-2043366068-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '3eb6415a9af24132ae706d907f5cc640', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e448c3eab37f4d65a24b5a68d3bb6067', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61923) authorize /opt/stack/nova/nova/policy.py:201}} [ 547.669192] env[61923]: DEBUG oslo_concurrency.lockutils [None req-3b10f02e-4025-4174-94c7-8515a7e12175 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 547.669192] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-3b10f02e-4025-4174-94c7-8515a7e12175 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] [instance: f7cc960c-e06a-4c58-9367-ec8771fe09d5] Processing image 0f153f63-ae0a-45b1-b7f5-7f9b673c947f {{(pid=61923) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 547.669192] env[61923]: DEBUG oslo_concurrency.lockutils [None req-3b10f02e-4025-4174-94c7-8515a7e12175 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 547.669192] env[61923]: DEBUG oslo_concurrency.lockutils [None req-3b10f02e-4025-4174-94c7-8515a7e12175 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 547.669347] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-3b10f02e-4025-4174-94c7-8515a7e12175 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61923) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 547.669347] env[61923]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-67682c55-353b-4ca0-b44d-d413872ef872 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 547.802866] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d213861-a78b-4ed3-8623-3cccb591cb82 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 547.820036] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57e49692-dce8-40af-b20b-3c951e26597d {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 547.859049] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3067cc60-da57-49c9-bb3b-11013694629c {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 547.866840] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d29d357d-ed62-4950-9805-95427297e497 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 547.888637] env[61923]: DEBUG nova.compute.provider_tree [None req-8a6b6bc2-54c9-4389-ab30-057158d86227 tempest-ServerDiagnosticsNegativeTest-664316905 tempest-ServerDiagnosticsNegativeTest-664316905-project-member] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 547.935959] env[61923]: DEBUG oslo_concurrency.lockutils [None req-cb394400-3d4b-4149-a290-413641bb6f18 tempest-InstanceActionsV221TestJSON-1730893270 tempest-InstanceActionsV221TestJSON-1730893270-project-member] Acquiring lock "dc8cc22d-aba9-4dba-b456-1b323013d234" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 547.936257] env[61923]: DEBUG oslo_concurrency.lockutils [None req-cb394400-3d4b-4149-a290-413641bb6f18 tempest-InstanceActionsV221TestJSON-1730893270 tempest-InstanceActionsV221TestJSON-1730893270-project-member] Lock "dc8cc22d-aba9-4dba-b456-1b323013d234" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 547.951712] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-3b10f02e-4025-4174-94c7-8515a7e12175 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61923) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 547.952146] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-3b10f02e-4025-4174-94c7-8515a7e12175 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61923) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 547.952715] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4eeb6b7c-13c8-4325-b756-7190245f2fec {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 547.956274] env[61923]: DEBUG nova.compute.manager [None req-d9e2f878-40e6-4b41-8b07-da611f2f81f1 tempest-VolumesAssistedSnapshotsTest-2043366068 tempest-VolumesAssistedSnapshotsTest-2043366068-project-member] [instance: 06bda5ee-c7a8-47d1-9e0c-a5af67d40006] Start building block device mappings for instance. {{(pid=61923) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 547.961493] env[61923]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-060a4183-ae3a-428a-afd1-fdb25b754e90 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 547.966539] env[61923]: DEBUG oslo_vmware.api [None req-3b10f02e-4025-4174-94c7-8515a7e12175 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] Waiting for the task: (returnval){ [ 547.966539] env[61923]: value = "session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]525a834b-bf55-541e-12e1-df9122afe4c0" [ 547.966539] env[61923]: _type = "Task" [ 547.966539] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 547.974583] env[61923]: DEBUG oslo_vmware.api [None req-3b10f02e-4025-4174-94c7-8515a7e12175 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]525a834b-bf55-541e-12e1-df9122afe4c0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 548.180616] env[61923]: DEBUG nova.network.neutron [None req-d9e2f878-40e6-4b41-8b07-da611f2f81f1 tempest-VolumesAssistedSnapshotsTest-2043366068 tempest-VolumesAssistedSnapshotsTest-2043366068-project-member] [instance: 06bda5ee-c7a8-47d1-9e0c-a5af67d40006] Successfully created port: 48788634-af14-4531-8cff-07f9a6f55f4c {{(pid=61923) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 548.392105] env[61923]: DEBUG nova.scheduler.client.report [None req-8a6b6bc2-54c9-4389-ab30-057158d86227 tempest-ServerDiagnosticsNegativeTest-664316905 tempest-ServerDiagnosticsNegativeTest-664316905-project-member] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 548.479020] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-3b10f02e-4025-4174-94c7-8515a7e12175 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] [instance: f7cc960c-e06a-4c58-9367-ec8771fe09d5] Preparing fetch location {{(pid=61923) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 548.479020] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-3b10f02e-4025-4174-94c7-8515a7e12175 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] Creating directory with path [datastore1] vmware_temp/458df3c4-aac7-4022-909f-21ebff462779/0f153f63-ae0a-45b1-b7f5-7f9b673c947f {{(pid=61923) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 548.479020] env[61923]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-6c398fd9-ae4a-4834-98fd-bdbd396a0721 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 548.502018] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-3b10f02e-4025-4174-94c7-8515a7e12175 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] Created directory with path [datastore1] vmware_temp/458df3c4-aac7-4022-909f-21ebff462779/0f153f63-ae0a-45b1-b7f5-7f9b673c947f {{(pid=61923) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 548.502018] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-3b10f02e-4025-4174-94c7-8515a7e12175 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] [instance: f7cc960c-e06a-4c58-9367-ec8771fe09d5] Fetch image to [datastore1] vmware_temp/458df3c4-aac7-4022-909f-21ebff462779/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/tmp-sparse.vmdk {{(pid=61923) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 548.502018] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-3b10f02e-4025-4174-94c7-8515a7e12175 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] [instance: f7cc960c-e06a-4c58-9367-ec8771fe09d5] Downloading image file data 0f153f63-ae0a-45b1-b7f5-7f9b673c947f to [datastore1] vmware_temp/458df3c4-aac7-4022-909f-21ebff462779/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/tmp-sparse.vmdk on the data store datastore1 {{(pid=61923) _fetch_image_as_file /opt/stack/nova/nova/virt/vmwareapi/vmops.py:399}} [ 548.502018] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af95870d-52a5-4595-b1fd-b457541f4388 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 548.513379] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5007621c-cbe4-4660-ab78-0250a849f603 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 548.533175] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1630762-46f9-447f-ba28-0fd5982d0f50 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 548.569593] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f5d84f4-ac16-483a-a6fb-72cf8bb897a2 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 548.576431] env[61923]: DEBUG oslo_vmware.service [-] Invoking SessionManager.AcquireGenericServiceTicket with opID=oslo.vmware-f8b2dc7a-a44a-421a-b74f-c31cf82ab2ba {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 548.611405] env[61923]: DEBUG nova.virt.vmwareapi.images [None req-3b10f02e-4025-4174-94c7-8515a7e12175 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] [instance: f7cc960c-e06a-4c58-9367-ec8771fe09d5] Downloading image file data 0f153f63-ae0a-45b1-b7f5-7f9b673c947f to the data store datastore1 {{(pid=61923) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:245}} [ 548.686966] env[61923]: DEBUG oslo_vmware.rw_handles [None req-3b10f02e-4025-4174-94c7-8515a7e12175 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] Creating HTTP connection to write to file with size = 21318656 and URL = https://esx7c2n1.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/458df3c4-aac7-4022-909f-21ebff462779/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore1. {{(pid=61923) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 548.897089] env[61923]: DEBUG oslo_concurrency.lockutils [None req-8a6b6bc2-54c9-4389-ab30-057158d86227 tempest-ServerDiagnosticsNegativeTest-664316905 tempest-ServerDiagnosticsNegativeTest-664316905-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.956s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 548.897900] env[61923]: ERROR nova.compute.manager [None req-8a6b6bc2-54c9-4389-ab30-057158d86227 tempest-ServerDiagnosticsNegativeTest-664316905 tempest-ServerDiagnosticsNegativeTest-664316905-project-member] [instance: b0a45314-eed8-4d0a-bea2-d4b47dc0bc80] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 6bf512ed-8713-43d4-bc7e-ce2b9a514f52, please check neutron logs for more information. [ 548.897900] env[61923]: ERROR nova.compute.manager [instance: b0a45314-eed8-4d0a-bea2-d4b47dc0bc80] Traceback (most recent call last): [ 548.897900] env[61923]: ERROR nova.compute.manager [instance: b0a45314-eed8-4d0a-bea2-d4b47dc0bc80] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 548.897900] env[61923]: ERROR nova.compute.manager [instance: b0a45314-eed8-4d0a-bea2-d4b47dc0bc80] self.driver.spawn(context, instance, image_meta, [ 548.897900] env[61923]: ERROR nova.compute.manager [instance: b0a45314-eed8-4d0a-bea2-d4b47dc0bc80] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 548.897900] env[61923]: ERROR nova.compute.manager [instance: b0a45314-eed8-4d0a-bea2-d4b47dc0bc80] self._vmops.spawn(context, instance, image_meta, injected_files, [ 548.897900] env[61923]: ERROR nova.compute.manager [instance: b0a45314-eed8-4d0a-bea2-d4b47dc0bc80] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 548.897900] env[61923]: ERROR nova.compute.manager [instance: b0a45314-eed8-4d0a-bea2-d4b47dc0bc80] vm_ref = self.build_virtual_machine(instance, [ 548.897900] env[61923]: ERROR nova.compute.manager [instance: b0a45314-eed8-4d0a-bea2-d4b47dc0bc80] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 548.897900] env[61923]: ERROR nova.compute.manager [instance: b0a45314-eed8-4d0a-bea2-d4b47dc0bc80] vif_infos = vmwarevif.get_vif_info(self._session, [ 548.897900] env[61923]: ERROR nova.compute.manager [instance: b0a45314-eed8-4d0a-bea2-d4b47dc0bc80] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 548.898541] env[61923]: ERROR nova.compute.manager [instance: b0a45314-eed8-4d0a-bea2-d4b47dc0bc80] for vif in network_info: [ 548.898541] env[61923]: ERROR nova.compute.manager [instance: b0a45314-eed8-4d0a-bea2-d4b47dc0bc80] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 548.898541] env[61923]: ERROR nova.compute.manager [instance: b0a45314-eed8-4d0a-bea2-d4b47dc0bc80] return self._sync_wrapper(fn, *args, **kwargs) [ 548.898541] env[61923]: ERROR nova.compute.manager [instance: b0a45314-eed8-4d0a-bea2-d4b47dc0bc80] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 548.898541] env[61923]: ERROR nova.compute.manager [instance: b0a45314-eed8-4d0a-bea2-d4b47dc0bc80] self.wait() [ 548.898541] env[61923]: ERROR nova.compute.manager [instance: b0a45314-eed8-4d0a-bea2-d4b47dc0bc80] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 548.898541] env[61923]: ERROR nova.compute.manager [instance: b0a45314-eed8-4d0a-bea2-d4b47dc0bc80] self[:] = self._gt.wait() [ 548.898541] env[61923]: ERROR nova.compute.manager [instance: b0a45314-eed8-4d0a-bea2-d4b47dc0bc80] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 548.898541] env[61923]: ERROR nova.compute.manager [instance: b0a45314-eed8-4d0a-bea2-d4b47dc0bc80] return self._exit_event.wait() [ 548.898541] env[61923]: ERROR nova.compute.manager [instance: b0a45314-eed8-4d0a-bea2-d4b47dc0bc80] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 548.898541] env[61923]: ERROR nova.compute.manager [instance: b0a45314-eed8-4d0a-bea2-d4b47dc0bc80] result = hub.switch() [ 548.898541] env[61923]: ERROR nova.compute.manager [instance: b0a45314-eed8-4d0a-bea2-d4b47dc0bc80] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 548.898541] env[61923]: ERROR nova.compute.manager [instance: b0a45314-eed8-4d0a-bea2-d4b47dc0bc80] return self.greenlet.switch() [ 548.898963] env[61923]: ERROR nova.compute.manager [instance: b0a45314-eed8-4d0a-bea2-d4b47dc0bc80] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 548.898963] env[61923]: ERROR nova.compute.manager [instance: b0a45314-eed8-4d0a-bea2-d4b47dc0bc80] result = function(*args, **kwargs) [ 548.898963] env[61923]: ERROR nova.compute.manager [instance: b0a45314-eed8-4d0a-bea2-d4b47dc0bc80] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 548.898963] env[61923]: ERROR nova.compute.manager [instance: b0a45314-eed8-4d0a-bea2-d4b47dc0bc80] return func(*args, **kwargs) [ 548.898963] env[61923]: ERROR nova.compute.manager [instance: b0a45314-eed8-4d0a-bea2-d4b47dc0bc80] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 548.898963] env[61923]: ERROR nova.compute.manager [instance: b0a45314-eed8-4d0a-bea2-d4b47dc0bc80] raise e [ 548.898963] env[61923]: ERROR nova.compute.manager [instance: b0a45314-eed8-4d0a-bea2-d4b47dc0bc80] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 548.898963] env[61923]: ERROR nova.compute.manager [instance: b0a45314-eed8-4d0a-bea2-d4b47dc0bc80] nwinfo = self.network_api.allocate_for_instance( [ 548.898963] env[61923]: ERROR nova.compute.manager [instance: b0a45314-eed8-4d0a-bea2-d4b47dc0bc80] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 548.898963] env[61923]: ERROR nova.compute.manager [instance: b0a45314-eed8-4d0a-bea2-d4b47dc0bc80] created_port_ids = self._update_ports_for_instance( [ 548.898963] env[61923]: ERROR nova.compute.manager [instance: b0a45314-eed8-4d0a-bea2-d4b47dc0bc80] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 548.898963] env[61923]: ERROR nova.compute.manager [instance: b0a45314-eed8-4d0a-bea2-d4b47dc0bc80] with excutils.save_and_reraise_exception(): [ 548.898963] env[61923]: ERROR nova.compute.manager [instance: b0a45314-eed8-4d0a-bea2-d4b47dc0bc80] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 548.899809] env[61923]: ERROR nova.compute.manager [instance: b0a45314-eed8-4d0a-bea2-d4b47dc0bc80] self.force_reraise() [ 548.899809] env[61923]: ERROR nova.compute.manager [instance: b0a45314-eed8-4d0a-bea2-d4b47dc0bc80] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 548.899809] env[61923]: ERROR nova.compute.manager [instance: b0a45314-eed8-4d0a-bea2-d4b47dc0bc80] raise self.value [ 548.899809] env[61923]: ERROR nova.compute.manager [instance: b0a45314-eed8-4d0a-bea2-d4b47dc0bc80] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 548.899809] env[61923]: ERROR nova.compute.manager [instance: b0a45314-eed8-4d0a-bea2-d4b47dc0bc80] updated_port = self._update_port( [ 548.899809] env[61923]: ERROR nova.compute.manager [instance: b0a45314-eed8-4d0a-bea2-d4b47dc0bc80] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 548.899809] env[61923]: ERROR nova.compute.manager [instance: b0a45314-eed8-4d0a-bea2-d4b47dc0bc80] _ensure_no_port_binding_failure(port) [ 548.899809] env[61923]: ERROR nova.compute.manager [instance: b0a45314-eed8-4d0a-bea2-d4b47dc0bc80] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 548.899809] env[61923]: ERROR nova.compute.manager [instance: b0a45314-eed8-4d0a-bea2-d4b47dc0bc80] raise exception.PortBindingFailed(port_id=port['id']) [ 548.899809] env[61923]: ERROR nova.compute.manager [instance: b0a45314-eed8-4d0a-bea2-d4b47dc0bc80] nova.exception.PortBindingFailed: Binding failed for port 6bf512ed-8713-43d4-bc7e-ce2b9a514f52, please check neutron logs for more information. [ 548.899809] env[61923]: ERROR nova.compute.manager [instance: b0a45314-eed8-4d0a-bea2-d4b47dc0bc80] [ 548.900199] env[61923]: DEBUG nova.compute.utils [None req-8a6b6bc2-54c9-4389-ab30-057158d86227 tempest-ServerDiagnosticsNegativeTest-664316905 tempest-ServerDiagnosticsNegativeTest-664316905-project-member] [instance: b0a45314-eed8-4d0a-bea2-d4b47dc0bc80] Binding failed for port 6bf512ed-8713-43d4-bc7e-ce2b9a514f52, please check neutron logs for more information. {{(pid=61923) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 548.900860] env[61923]: DEBUG oslo_concurrency.lockutils [None req-7781f812-6fae-4bd6-9d91-ec6098c06372 tempest-InstanceActionsTestJSON-2015629344 tempest-InstanceActionsTestJSON-2015629344-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 11.976s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 548.904745] env[61923]: DEBUG nova.compute.manager [None req-8a6b6bc2-54c9-4389-ab30-057158d86227 tempest-ServerDiagnosticsNegativeTest-664316905 tempest-ServerDiagnosticsNegativeTest-664316905-project-member] [instance: b0a45314-eed8-4d0a-bea2-d4b47dc0bc80] Build of instance b0a45314-eed8-4d0a-bea2-d4b47dc0bc80 was re-scheduled: Binding failed for port 6bf512ed-8713-43d4-bc7e-ce2b9a514f52, please check neutron logs for more information. {{(pid=61923) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 548.904745] env[61923]: DEBUG nova.compute.manager [None req-8a6b6bc2-54c9-4389-ab30-057158d86227 tempest-ServerDiagnosticsNegativeTest-664316905 tempest-ServerDiagnosticsNegativeTest-664316905-project-member] [instance: b0a45314-eed8-4d0a-bea2-d4b47dc0bc80] Unplugging VIFs for instance {{(pid=61923) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 548.904745] env[61923]: DEBUG oslo_concurrency.lockutils [None req-8a6b6bc2-54c9-4389-ab30-057158d86227 tempest-ServerDiagnosticsNegativeTest-664316905 tempest-ServerDiagnosticsNegativeTest-664316905-project-member] Acquiring lock "refresh_cache-b0a45314-eed8-4d0a-bea2-d4b47dc0bc80" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 548.905090] env[61923]: DEBUG oslo_concurrency.lockutils [None req-8a6b6bc2-54c9-4389-ab30-057158d86227 tempest-ServerDiagnosticsNegativeTest-664316905 tempest-ServerDiagnosticsNegativeTest-664316905-project-member] Acquired lock "refresh_cache-b0a45314-eed8-4d0a-bea2-d4b47dc0bc80" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 548.905090] env[61923]: DEBUG nova.network.neutron [None req-8a6b6bc2-54c9-4389-ab30-057158d86227 tempest-ServerDiagnosticsNegativeTest-664316905 tempest-ServerDiagnosticsNegativeTest-664316905-project-member] [instance: b0a45314-eed8-4d0a-bea2-d4b47dc0bc80] Building network info cache for instance {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 548.969213] env[61923]: DEBUG nova.compute.manager [None req-d9e2f878-40e6-4b41-8b07-da611f2f81f1 tempest-VolumesAssistedSnapshotsTest-2043366068 tempest-VolumesAssistedSnapshotsTest-2043366068-project-member] [instance: 06bda5ee-c7a8-47d1-9e0c-a5af67d40006] Start spawning the instance on the hypervisor. {{(pid=61923) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 549.001974] env[61923]: DEBUG nova.virt.hardware [None req-d9e2f878-40e6-4b41-8b07-da611f2f81f1 tempest-VolumesAssistedSnapshotsTest-2043366068 tempest-VolumesAssistedSnapshotsTest-2043366068-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-29T20:07:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-29T20:07:35Z,direct_url=,disk_format='vmdk',id=0f153f63-ae0a-45b1-b7f5-7f9b673c947f,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='4e7b5e3b3c3443c8bd5c70f8a15739f5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-29T20:07:36Z,virtual_size=,visibility=), allow threads: False {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 549.002250] env[61923]: DEBUG nova.virt.hardware [None req-d9e2f878-40e6-4b41-8b07-da611f2f81f1 tempest-VolumesAssistedSnapshotsTest-2043366068 tempest-VolumesAssistedSnapshotsTest-2043366068-project-member] Flavor limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 549.002406] env[61923]: DEBUG nova.virt.hardware [None req-d9e2f878-40e6-4b41-8b07-da611f2f81f1 tempest-VolumesAssistedSnapshotsTest-2043366068 tempest-VolumesAssistedSnapshotsTest-2043366068-project-member] Image limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 549.002582] env[61923]: DEBUG nova.virt.hardware [None req-d9e2f878-40e6-4b41-8b07-da611f2f81f1 tempest-VolumesAssistedSnapshotsTest-2043366068 tempest-VolumesAssistedSnapshotsTest-2043366068-project-member] Flavor pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 549.002732] env[61923]: DEBUG nova.virt.hardware [None req-d9e2f878-40e6-4b41-8b07-da611f2f81f1 tempest-VolumesAssistedSnapshotsTest-2043366068 tempest-VolumesAssistedSnapshotsTest-2043366068-project-member] Image pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 549.002902] env[61923]: DEBUG nova.virt.hardware [None req-d9e2f878-40e6-4b41-8b07-da611f2f81f1 tempest-VolumesAssistedSnapshotsTest-2043366068 tempest-VolumesAssistedSnapshotsTest-2043366068-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 549.003148] env[61923]: DEBUG nova.virt.hardware [None req-d9e2f878-40e6-4b41-8b07-da611f2f81f1 tempest-VolumesAssistedSnapshotsTest-2043366068 tempest-VolumesAssistedSnapshotsTest-2043366068-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 549.003317] env[61923]: DEBUG nova.virt.hardware [None req-d9e2f878-40e6-4b41-8b07-da611f2f81f1 tempest-VolumesAssistedSnapshotsTest-2043366068 tempest-VolumesAssistedSnapshotsTest-2043366068-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 549.003480] env[61923]: DEBUG nova.virt.hardware [None req-d9e2f878-40e6-4b41-8b07-da611f2f81f1 tempest-VolumesAssistedSnapshotsTest-2043366068 tempest-VolumesAssistedSnapshotsTest-2043366068-project-member] Got 1 possible topologies {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 549.003639] env[61923]: DEBUG nova.virt.hardware [None req-d9e2f878-40e6-4b41-8b07-da611f2f81f1 tempest-VolumesAssistedSnapshotsTest-2043366068 tempest-VolumesAssistedSnapshotsTest-2043366068-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 549.004702] env[61923]: DEBUG nova.virt.hardware [None req-d9e2f878-40e6-4b41-8b07-da611f2f81f1 tempest-VolumesAssistedSnapshotsTest-2043366068 tempest-VolumesAssistedSnapshotsTest-2043366068-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 549.004830] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57456393-ff6d-4b70-aea7-cc21a31a6d88 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 549.016727] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7734720-7760-47a9-a1c8-9f81772196fb {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 549.266534] env[61923]: DEBUG oslo_concurrency.lockutils [None req-753531fa-39a1-4c82-9915-d195bf0fe1b2 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] Acquiring lock "03100310-86ce-4739-8636-cd2aa1b542ff" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 549.266787] env[61923]: DEBUG oslo_concurrency.lockutils [None req-753531fa-39a1-4c82-9915-d195bf0fe1b2 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] Lock "03100310-86ce-4739-8636-cd2aa1b542ff" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 549.355517] env[61923]: DEBUG oslo_vmware.rw_handles [None req-3b10f02e-4025-4174-94c7-8515a7e12175 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] Completed reading data from the image iterator. {{(pid=61923) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 549.355745] env[61923]: DEBUG oslo_vmware.rw_handles [None req-3b10f02e-4025-4174-94c7-8515a7e12175 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] Closing write handle for https://esx7c2n1.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/458df3c4-aac7-4022-909f-21ebff462779/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore1. {{(pid=61923) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:281}} [ 549.429328] env[61923]: DEBUG nova.network.neutron [None req-8a6b6bc2-54c9-4389-ab30-057158d86227 tempest-ServerDiagnosticsNegativeTest-664316905 tempest-ServerDiagnosticsNegativeTest-664316905-project-member] [instance: b0a45314-eed8-4d0a-bea2-d4b47dc0bc80] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 549.481127] env[61923]: DEBUG nova.network.neutron [None req-8a6b6bc2-54c9-4389-ab30-057158d86227 tempest-ServerDiagnosticsNegativeTest-664316905 tempest-ServerDiagnosticsNegativeTest-664316905-project-member] [instance: b0a45314-eed8-4d0a-bea2-d4b47dc0bc80] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 549.492838] env[61923]: DEBUG nova.virt.vmwareapi.images [None req-3b10f02e-4025-4174-94c7-8515a7e12175 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] [instance: f7cc960c-e06a-4c58-9367-ec8771fe09d5] Downloaded image file data 0f153f63-ae0a-45b1-b7f5-7f9b673c947f to vmware_temp/458df3c4-aac7-4022-909f-21ebff462779/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/tmp-sparse.vmdk on the data store datastore1 {{(pid=61923) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:258}} [ 549.494934] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-3b10f02e-4025-4174-94c7-8515a7e12175 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] [instance: f7cc960c-e06a-4c58-9367-ec8771fe09d5] Caching image {{(pid=61923) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 549.495192] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-3b10f02e-4025-4174-94c7-8515a7e12175 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] Copying Virtual Disk [datastore1] vmware_temp/458df3c4-aac7-4022-909f-21ebff462779/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/tmp-sparse.vmdk to [datastore1] vmware_temp/458df3c4-aac7-4022-909f-21ebff462779/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk {{(pid=61923) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 549.495780] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-3eb31865-56fb-424a-a0aa-16bdb76cd5b3 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 549.507724] env[61923]: DEBUG oslo_vmware.api [None req-3b10f02e-4025-4174-94c7-8515a7e12175 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] Waiting for the task: (returnval){ [ 549.507724] env[61923]: value = "task-1377426" [ 549.507724] env[61923]: _type = "Task" [ 549.507724] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 549.523052] env[61923]: DEBUG oslo_vmware.api [None req-3b10f02e-4025-4174-94c7-8515a7e12175 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] Task: {'id': task-1377426, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 549.732228] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6496b70-f10b-4901-b749-d9ba5f52f6b5 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 549.740251] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ad41557-64b9-4071-9755-b3600af09d69 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 549.774613] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e447a150-babc-4c61-8d37-211ada74f96c {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 549.782458] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bac4e095-73a2-49ec-8444-5bf1af594874 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 549.796820] env[61923]: DEBUG nova.compute.provider_tree [None req-7781f812-6fae-4bd6-9d91-ec6098c06372 tempest-InstanceActionsTestJSON-2015629344 tempest-InstanceActionsTestJSON-2015629344-project-member] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 549.818836] env[61923]: ERROR nova.compute.manager [None req-d9e2f878-40e6-4b41-8b07-da611f2f81f1 tempest-VolumesAssistedSnapshotsTest-2043366068 tempest-VolumesAssistedSnapshotsTest-2043366068-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 48788634-af14-4531-8cff-07f9a6f55f4c, please check neutron logs for more information. [ 549.818836] env[61923]: ERROR nova.compute.manager Traceback (most recent call last): [ 549.818836] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 549.818836] env[61923]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 549.818836] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 549.818836] env[61923]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 549.818836] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 549.818836] env[61923]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 549.818836] env[61923]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 549.818836] env[61923]: ERROR nova.compute.manager self.force_reraise() [ 549.818836] env[61923]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 549.818836] env[61923]: ERROR nova.compute.manager raise self.value [ 549.818836] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 549.818836] env[61923]: ERROR nova.compute.manager updated_port = self._update_port( [ 549.818836] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 549.818836] env[61923]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 549.819387] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 549.819387] env[61923]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 549.819387] env[61923]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 48788634-af14-4531-8cff-07f9a6f55f4c, please check neutron logs for more information. [ 549.819387] env[61923]: ERROR nova.compute.manager [ 549.819387] env[61923]: Traceback (most recent call last): [ 549.819387] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 549.819387] env[61923]: listener.cb(fileno) [ 549.819387] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 549.819387] env[61923]: result = function(*args, **kwargs) [ 549.819387] env[61923]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 549.819387] env[61923]: return func(*args, **kwargs) [ 549.819387] env[61923]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 549.819387] env[61923]: raise e [ 549.819387] env[61923]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 549.819387] env[61923]: nwinfo = self.network_api.allocate_for_instance( [ 549.819387] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 549.819387] env[61923]: created_port_ids = self._update_ports_for_instance( [ 549.819387] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 549.819387] env[61923]: with excutils.save_and_reraise_exception(): [ 549.819387] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 549.819387] env[61923]: self.force_reraise() [ 549.819387] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 549.819387] env[61923]: raise self.value [ 549.819387] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 549.819387] env[61923]: updated_port = self._update_port( [ 549.819387] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 549.819387] env[61923]: _ensure_no_port_binding_failure(port) [ 549.819387] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 549.819387] env[61923]: raise exception.PortBindingFailed(port_id=port['id']) [ 549.820242] env[61923]: nova.exception.PortBindingFailed: Binding failed for port 48788634-af14-4531-8cff-07f9a6f55f4c, please check neutron logs for more information. [ 549.820242] env[61923]: Removing descriptor: 14 [ 549.820242] env[61923]: ERROR nova.compute.manager [None req-d9e2f878-40e6-4b41-8b07-da611f2f81f1 tempest-VolumesAssistedSnapshotsTest-2043366068 tempest-VolumesAssistedSnapshotsTest-2043366068-project-member] [instance: 06bda5ee-c7a8-47d1-9e0c-a5af67d40006] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 48788634-af14-4531-8cff-07f9a6f55f4c, please check neutron logs for more information. [ 549.820242] env[61923]: ERROR nova.compute.manager [instance: 06bda5ee-c7a8-47d1-9e0c-a5af67d40006] Traceback (most recent call last): [ 549.820242] env[61923]: ERROR nova.compute.manager [instance: 06bda5ee-c7a8-47d1-9e0c-a5af67d40006] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 549.820242] env[61923]: ERROR nova.compute.manager [instance: 06bda5ee-c7a8-47d1-9e0c-a5af67d40006] yield resources [ 549.820242] env[61923]: ERROR nova.compute.manager [instance: 06bda5ee-c7a8-47d1-9e0c-a5af67d40006] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 549.820242] env[61923]: ERROR nova.compute.manager [instance: 06bda5ee-c7a8-47d1-9e0c-a5af67d40006] self.driver.spawn(context, instance, image_meta, [ 549.820242] env[61923]: ERROR nova.compute.manager [instance: 06bda5ee-c7a8-47d1-9e0c-a5af67d40006] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 549.820242] env[61923]: ERROR nova.compute.manager [instance: 06bda5ee-c7a8-47d1-9e0c-a5af67d40006] self._vmops.spawn(context, instance, image_meta, injected_files, [ 549.820242] env[61923]: ERROR nova.compute.manager [instance: 06bda5ee-c7a8-47d1-9e0c-a5af67d40006] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 549.820242] env[61923]: ERROR nova.compute.manager [instance: 06bda5ee-c7a8-47d1-9e0c-a5af67d40006] vm_ref = self.build_virtual_machine(instance, [ 549.820553] env[61923]: ERROR nova.compute.manager [instance: 06bda5ee-c7a8-47d1-9e0c-a5af67d40006] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 549.820553] env[61923]: ERROR nova.compute.manager [instance: 06bda5ee-c7a8-47d1-9e0c-a5af67d40006] vif_infos = vmwarevif.get_vif_info(self._session, [ 549.820553] env[61923]: ERROR nova.compute.manager [instance: 06bda5ee-c7a8-47d1-9e0c-a5af67d40006] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 549.820553] env[61923]: ERROR nova.compute.manager [instance: 06bda5ee-c7a8-47d1-9e0c-a5af67d40006] for vif in network_info: [ 549.820553] env[61923]: ERROR nova.compute.manager [instance: 06bda5ee-c7a8-47d1-9e0c-a5af67d40006] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 549.820553] env[61923]: ERROR nova.compute.manager [instance: 06bda5ee-c7a8-47d1-9e0c-a5af67d40006] return self._sync_wrapper(fn, *args, **kwargs) [ 549.820553] env[61923]: ERROR nova.compute.manager [instance: 06bda5ee-c7a8-47d1-9e0c-a5af67d40006] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 549.820553] env[61923]: ERROR nova.compute.manager [instance: 06bda5ee-c7a8-47d1-9e0c-a5af67d40006] self.wait() [ 549.820553] env[61923]: ERROR nova.compute.manager [instance: 06bda5ee-c7a8-47d1-9e0c-a5af67d40006] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 549.820553] env[61923]: ERROR nova.compute.manager [instance: 06bda5ee-c7a8-47d1-9e0c-a5af67d40006] self[:] = self._gt.wait() [ 549.820553] env[61923]: ERROR nova.compute.manager [instance: 06bda5ee-c7a8-47d1-9e0c-a5af67d40006] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 549.820553] env[61923]: ERROR nova.compute.manager [instance: 06bda5ee-c7a8-47d1-9e0c-a5af67d40006] return self._exit_event.wait() [ 549.820553] env[61923]: ERROR nova.compute.manager [instance: 06bda5ee-c7a8-47d1-9e0c-a5af67d40006] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 549.821254] env[61923]: ERROR nova.compute.manager [instance: 06bda5ee-c7a8-47d1-9e0c-a5af67d40006] result = hub.switch() [ 549.821254] env[61923]: ERROR nova.compute.manager [instance: 06bda5ee-c7a8-47d1-9e0c-a5af67d40006] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 549.821254] env[61923]: ERROR nova.compute.manager [instance: 06bda5ee-c7a8-47d1-9e0c-a5af67d40006] return self.greenlet.switch() [ 549.821254] env[61923]: ERROR nova.compute.manager [instance: 06bda5ee-c7a8-47d1-9e0c-a5af67d40006] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 549.821254] env[61923]: ERROR nova.compute.manager [instance: 06bda5ee-c7a8-47d1-9e0c-a5af67d40006] result = function(*args, **kwargs) [ 549.821254] env[61923]: ERROR nova.compute.manager [instance: 06bda5ee-c7a8-47d1-9e0c-a5af67d40006] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 549.821254] env[61923]: ERROR nova.compute.manager [instance: 06bda5ee-c7a8-47d1-9e0c-a5af67d40006] return func(*args, **kwargs) [ 549.821254] env[61923]: ERROR nova.compute.manager [instance: 06bda5ee-c7a8-47d1-9e0c-a5af67d40006] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 549.821254] env[61923]: ERROR nova.compute.manager [instance: 06bda5ee-c7a8-47d1-9e0c-a5af67d40006] raise e [ 549.821254] env[61923]: ERROR nova.compute.manager [instance: 06bda5ee-c7a8-47d1-9e0c-a5af67d40006] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 549.821254] env[61923]: ERROR nova.compute.manager [instance: 06bda5ee-c7a8-47d1-9e0c-a5af67d40006] nwinfo = self.network_api.allocate_for_instance( [ 549.821254] env[61923]: ERROR nova.compute.manager [instance: 06bda5ee-c7a8-47d1-9e0c-a5af67d40006] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 549.821254] env[61923]: ERROR nova.compute.manager [instance: 06bda5ee-c7a8-47d1-9e0c-a5af67d40006] created_port_ids = self._update_ports_for_instance( [ 549.821669] env[61923]: ERROR nova.compute.manager [instance: 06bda5ee-c7a8-47d1-9e0c-a5af67d40006] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 549.821669] env[61923]: ERROR nova.compute.manager [instance: 06bda5ee-c7a8-47d1-9e0c-a5af67d40006] with excutils.save_and_reraise_exception(): [ 549.821669] env[61923]: ERROR nova.compute.manager [instance: 06bda5ee-c7a8-47d1-9e0c-a5af67d40006] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 549.821669] env[61923]: ERROR nova.compute.manager [instance: 06bda5ee-c7a8-47d1-9e0c-a5af67d40006] self.force_reraise() [ 549.821669] env[61923]: ERROR nova.compute.manager [instance: 06bda5ee-c7a8-47d1-9e0c-a5af67d40006] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 549.821669] env[61923]: ERROR nova.compute.manager [instance: 06bda5ee-c7a8-47d1-9e0c-a5af67d40006] raise self.value [ 549.821669] env[61923]: ERROR nova.compute.manager [instance: 06bda5ee-c7a8-47d1-9e0c-a5af67d40006] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 549.821669] env[61923]: ERROR nova.compute.manager [instance: 06bda5ee-c7a8-47d1-9e0c-a5af67d40006] updated_port = self._update_port( [ 549.821669] env[61923]: ERROR nova.compute.manager [instance: 06bda5ee-c7a8-47d1-9e0c-a5af67d40006] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 549.821669] env[61923]: ERROR nova.compute.manager [instance: 06bda5ee-c7a8-47d1-9e0c-a5af67d40006] _ensure_no_port_binding_failure(port) [ 549.821669] env[61923]: ERROR nova.compute.manager [instance: 06bda5ee-c7a8-47d1-9e0c-a5af67d40006] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 549.821669] env[61923]: ERROR nova.compute.manager [instance: 06bda5ee-c7a8-47d1-9e0c-a5af67d40006] raise exception.PortBindingFailed(port_id=port['id']) [ 549.822316] env[61923]: ERROR nova.compute.manager [instance: 06bda5ee-c7a8-47d1-9e0c-a5af67d40006] nova.exception.PortBindingFailed: Binding failed for port 48788634-af14-4531-8cff-07f9a6f55f4c, please check neutron logs for more information. [ 549.822316] env[61923]: ERROR nova.compute.manager [instance: 06bda5ee-c7a8-47d1-9e0c-a5af67d40006] [ 549.822316] env[61923]: INFO nova.compute.manager [None req-d9e2f878-40e6-4b41-8b07-da611f2f81f1 tempest-VolumesAssistedSnapshotsTest-2043366068 tempest-VolumesAssistedSnapshotsTest-2043366068-project-member] [instance: 06bda5ee-c7a8-47d1-9e0c-a5af67d40006] Terminating instance [ 549.822900] env[61923]: DEBUG oslo_concurrency.lockutils [None req-d9e2f878-40e6-4b41-8b07-da611f2f81f1 tempest-VolumesAssistedSnapshotsTest-2043366068 tempest-VolumesAssistedSnapshotsTest-2043366068-project-member] Acquiring lock "refresh_cache-06bda5ee-c7a8-47d1-9e0c-a5af67d40006" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 549.823076] env[61923]: DEBUG oslo_concurrency.lockutils [None req-d9e2f878-40e6-4b41-8b07-da611f2f81f1 tempest-VolumesAssistedSnapshotsTest-2043366068 tempest-VolumesAssistedSnapshotsTest-2043366068-project-member] Acquired lock "refresh_cache-06bda5ee-c7a8-47d1-9e0c-a5af67d40006" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 549.823249] env[61923]: DEBUG nova.network.neutron [None req-d9e2f878-40e6-4b41-8b07-da611f2f81f1 tempest-VolumesAssistedSnapshotsTest-2043366068 tempest-VolumesAssistedSnapshotsTest-2043366068-project-member] [instance: 06bda5ee-c7a8-47d1-9e0c-a5af67d40006] Building network info cache for instance {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 549.927177] env[61923]: DEBUG nova.compute.manager [req-b3ca27af-ac7c-41ac-a4e1-309fd5a73cd7 req-950cb064-57db-45e8-b7e1-cb07560f3eea service nova] [instance: 06bda5ee-c7a8-47d1-9e0c-a5af67d40006] Received event network-changed-48788634-af14-4531-8cff-07f9a6f55f4c {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 549.927177] env[61923]: DEBUG nova.compute.manager [req-b3ca27af-ac7c-41ac-a4e1-309fd5a73cd7 req-950cb064-57db-45e8-b7e1-cb07560f3eea service nova] [instance: 06bda5ee-c7a8-47d1-9e0c-a5af67d40006] Refreshing instance network info cache due to event network-changed-48788634-af14-4531-8cff-07f9a6f55f4c. {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 549.927177] env[61923]: DEBUG oslo_concurrency.lockutils [req-b3ca27af-ac7c-41ac-a4e1-309fd5a73cd7 req-950cb064-57db-45e8-b7e1-cb07560f3eea service nova] Acquiring lock "refresh_cache-06bda5ee-c7a8-47d1-9e0c-a5af67d40006" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 549.987690] env[61923]: DEBUG oslo_concurrency.lockutils [None req-8a6b6bc2-54c9-4389-ab30-057158d86227 tempest-ServerDiagnosticsNegativeTest-664316905 tempest-ServerDiagnosticsNegativeTest-664316905-project-member] Releasing lock "refresh_cache-b0a45314-eed8-4d0a-bea2-d4b47dc0bc80" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 549.987800] env[61923]: DEBUG nova.compute.manager [None req-8a6b6bc2-54c9-4389-ab30-057158d86227 tempest-ServerDiagnosticsNegativeTest-664316905 tempest-ServerDiagnosticsNegativeTest-664316905-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61923) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 549.988732] env[61923]: DEBUG nova.compute.manager [None req-8a6b6bc2-54c9-4389-ab30-057158d86227 tempest-ServerDiagnosticsNegativeTest-664316905 tempest-ServerDiagnosticsNegativeTest-664316905-project-member] [instance: b0a45314-eed8-4d0a-bea2-d4b47dc0bc80] Deallocating network for instance {{(pid=61923) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 549.988732] env[61923]: DEBUG nova.network.neutron [None req-8a6b6bc2-54c9-4389-ab30-057158d86227 tempest-ServerDiagnosticsNegativeTest-664316905 tempest-ServerDiagnosticsNegativeTest-664316905-project-member] [instance: b0a45314-eed8-4d0a-bea2-d4b47dc0bc80] deallocate_for_instance() {{(pid=61923) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 550.008967] env[61923]: DEBUG nova.network.neutron [None req-8a6b6bc2-54c9-4389-ab30-057158d86227 tempest-ServerDiagnosticsNegativeTest-664316905 tempest-ServerDiagnosticsNegativeTest-664316905-project-member] [instance: b0a45314-eed8-4d0a-bea2-d4b47dc0bc80] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 550.019973] env[61923]: DEBUG oslo_vmware.api [None req-3b10f02e-4025-4174-94c7-8515a7e12175 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] Task: {'id': task-1377426, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 550.299809] env[61923]: DEBUG nova.scheduler.client.report [None req-7781f812-6fae-4bd6-9d91-ec6098c06372 tempest-InstanceActionsTestJSON-2015629344 tempest-InstanceActionsTestJSON-2015629344-project-member] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 550.347570] env[61923]: DEBUG nova.network.neutron [None req-d9e2f878-40e6-4b41-8b07-da611f2f81f1 tempest-VolumesAssistedSnapshotsTest-2043366068 tempest-VolumesAssistedSnapshotsTest-2043366068-project-member] [instance: 06bda5ee-c7a8-47d1-9e0c-a5af67d40006] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 550.439084] env[61923]: DEBUG nova.network.neutron [None req-d9e2f878-40e6-4b41-8b07-da611f2f81f1 tempest-VolumesAssistedSnapshotsTest-2043366068 tempest-VolumesAssistedSnapshotsTest-2043366068-project-member] [instance: 06bda5ee-c7a8-47d1-9e0c-a5af67d40006] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 550.516295] env[61923]: DEBUG nova.network.neutron [None req-8a6b6bc2-54c9-4389-ab30-057158d86227 tempest-ServerDiagnosticsNegativeTest-664316905 tempest-ServerDiagnosticsNegativeTest-664316905-project-member] [instance: b0a45314-eed8-4d0a-bea2-d4b47dc0bc80] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 550.524786] env[61923]: DEBUG oslo_vmware.api [None req-3b10f02e-4025-4174-94c7-8515a7e12175 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] Task: {'id': task-1377426, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.826881} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 550.525333] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-3b10f02e-4025-4174-94c7-8515a7e12175 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] Copied Virtual Disk [datastore1] vmware_temp/458df3c4-aac7-4022-909f-21ebff462779/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/tmp-sparse.vmdk to [datastore1] vmware_temp/458df3c4-aac7-4022-909f-21ebff462779/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk {{(pid=61923) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 550.525333] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-3b10f02e-4025-4174-94c7-8515a7e12175 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] Deleting the datastore file [datastore1] vmware_temp/458df3c4-aac7-4022-909f-21ebff462779/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/tmp-sparse.vmdk {{(pid=61923) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 550.525946] env[61923]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-6be77737-49e8-410e-85fd-812138427e2a {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 550.533723] env[61923]: DEBUG oslo_vmware.api [None req-3b10f02e-4025-4174-94c7-8515a7e12175 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] Waiting for the task: (returnval){ [ 550.533723] env[61923]: value = "task-1377427" [ 550.533723] env[61923]: _type = "Task" [ 550.533723] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 550.544965] env[61923]: DEBUG oslo_vmware.api [None req-3b10f02e-4025-4174-94c7-8515a7e12175 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] Task: {'id': task-1377427, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 550.806887] env[61923]: DEBUG oslo_concurrency.lockutils [None req-7781f812-6fae-4bd6-9d91-ec6098c06372 tempest-InstanceActionsTestJSON-2015629344 tempest-InstanceActionsTestJSON-2015629344-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.906s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 550.807274] env[61923]: ERROR nova.compute.manager [None req-7781f812-6fae-4bd6-9d91-ec6098c06372 tempest-InstanceActionsTestJSON-2015629344 tempest-InstanceActionsTestJSON-2015629344-project-member] [instance: 97148ea5-398a-4974-b8f6-62290e519794] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 7ec53f9e-f055-4a57-8228-7efbcdb7360f, please check neutron logs for more information. [ 550.807274] env[61923]: ERROR nova.compute.manager [instance: 97148ea5-398a-4974-b8f6-62290e519794] Traceback (most recent call last): [ 550.807274] env[61923]: ERROR nova.compute.manager [instance: 97148ea5-398a-4974-b8f6-62290e519794] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 550.807274] env[61923]: ERROR nova.compute.manager [instance: 97148ea5-398a-4974-b8f6-62290e519794] self.driver.spawn(context, instance, image_meta, [ 550.807274] env[61923]: ERROR nova.compute.manager [instance: 97148ea5-398a-4974-b8f6-62290e519794] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 550.807274] env[61923]: ERROR nova.compute.manager [instance: 97148ea5-398a-4974-b8f6-62290e519794] self._vmops.spawn(context, instance, image_meta, injected_files, [ 550.807274] env[61923]: ERROR nova.compute.manager [instance: 97148ea5-398a-4974-b8f6-62290e519794] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 550.807274] env[61923]: ERROR nova.compute.manager [instance: 97148ea5-398a-4974-b8f6-62290e519794] vm_ref = self.build_virtual_machine(instance, [ 550.807274] env[61923]: ERROR nova.compute.manager [instance: 97148ea5-398a-4974-b8f6-62290e519794] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 550.807274] env[61923]: ERROR nova.compute.manager [instance: 97148ea5-398a-4974-b8f6-62290e519794] vif_infos = vmwarevif.get_vif_info(self._session, [ 550.807274] env[61923]: ERROR nova.compute.manager [instance: 97148ea5-398a-4974-b8f6-62290e519794] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 550.808137] env[61923]: ERROR nova.compute.manager [instance: 97148ea5-398a-4974-b8f6-62290e519794] for vif in network_info: [ 550.808137] env[61923]: ERROR nova.compute.manager [instance: 97148ea5-398a-4974-b8f6-62290e519794] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 550.808137] env[61923]: ERROR nova.compute.manager [instance: 97148ea5-398a-4974-b8f6-62290e519794] return self._sync_wrapper(fn, *args, **kwargs) [ 550.808137] env[61923]: ERROR nova.compute.manager [instance: 97148ea5-398a-4974-b8f6-62290e519794] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 550.808137] env[61923]: ERROR nova.compute.manager [instance: 97148ea5-398a-4974-b8f6-62290e519794] self.wait() [ 550.808137] env[61923]: ERROR nova.compute.manager [instance: 97148ea5-398a-4974-b8f6-62290e519794] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 550.808137] env[61923]: ERROR nova.compute.manager [instance: 97148ea5-398a-4974-b8f6-62290e519794] self[:] = self._gt.wait() [ 550.808137] env[61923]: ERROR nova.compute.manager [instance: 97148ea5-398a-4974-b8f6-62290e519794] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 550.808137] env[61923]: ERROR nova.compute.manager [instance: 97148ea5-398a-4974-b8f6-62290e519794] return self._exit_event.wait() [ 550.808137] env[61923]: ERROR nova.compute.manager [instance: 97148ea5-398a-4974-b8f6-62290e519794] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 550.808137] env[61923]: ERROR nova.compute.manager [instance: 97148ea5-398a-4974-b8f6-62290e519794] result = hub.switch() [ 550.808137] env[61923]: ERROR nova.compute.manager [instance: 97148ea5-398a-4974-b8f6-62290e519794] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 550.808137] env[61923]: ERROR nova.compute.manager [instance: 97148ea5-398a-4974-b8f6-62290e519794] return self.greenlet.switch() [ 550.809309] env[61923]: ERROR nova.compute.manager [instance: 97148ea5-398a-4974-b8f6-62290e519794] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 550.809309] env[61923]: ERROR nova.compute.manager [instance: 97148ea5-398a-4974-b8f6-62290e519794] result = function(*args, **kwargs) [ 550.809309] env[61923]: ERROR nova.compute.manager [instance: 97148ea5-398a-4974-b8f6-62290e519794] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 550.809309] env[61923]: ERROR nova.compute.manager [instance: 97148ea5-398a-4974-b8f6-62290e519794] return func(*args, **kwargs) [ 550.809309] env[61923]: ERROR nova.compute.manager [instance: 97148ea5-398a-4974-b8f6-62290e519794] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 550.809309] env[61923]: ERROR nova.compute.manager [instance: 97148ea5-398a-4974-b8f6-62290e519794] raise e [ 550.809309] env[61923]: ERROR nova.compute.manager [instance: 97148ea5-398a-4974-b8f6-62290e519794] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 550.809309] env[61923]: ERROR nova.compute.manager [instance: 97148ea5-398a-4974-b8f6-62290e519794] nwinfo = self.network_api.allocate_for_instance( [ 550.809309] env[61923]: ERROR nova.compute.manager [instance: 97148ea5-398a-4974-b8f6-62290e519794] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 550.809309] env[61923]: ERROR nova.compute.manager [instance: 97148ea5-398a-4974-b8f6-62290e519794] created_port_ids = self._update_ports_for_instance( [ 550.809309] env[61923]: ERROR nova.compute.manager [instance: 97148ea5-398a-4974-b8f6-62290e519794] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 550.809309] env[61923]: ERROR nova.compute.manager [instance: 97148ea5-398a-4974-b8f6-62290e519794] with excutils.save_and_reraise_exception(): [ 550.809309] env[61923]: ERROR nova.compute.manager [instance: 97148ea5-398a-4974-b8f6-62290e519794] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 550.809830] env[61923]: ERROR nova.compute.manager [instance: 97148ea5-398a-4974-b8f6-62290e519794] self.force_reraise() [ 550.809830] env[61923]: ERROR nova.compute.manager [instance: 97148ea5-398a-4974-b8f6-62290e519794] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 550.809830] env[61923]: ERROR nova.compute.manager [instance: 97148ea5-398a-4974-b8f6-62290e519794] raise self.value [ 550.809830] env[61923]: ERROR nova.compute.manager [instance: 97148ea5-398a-4974-b8f6-62290e519794] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 550.809830] env[61923]: ERROR nova.compute.manager [instance: 97148ea5-398a-4974-b8f6-62290e519794] updated_port = self._update_port( [ 550.809830] env[61923]: ERROR nova.compute.manager [instance: 97148ea5-398a-4974-b8f6-62290e519794] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 550.809830] env[61923]: ERROR nova.compute.manager [instance: 97148ea5-398a-4974-b8f6-62290e519794] _ensure_no_port_binding_failure(port) [ 550.809830] env[61923]: ERROR nova.compute.manager [instance: 97148ea5-398a-4974-b8f6-62290e519794] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 550.809830] env[61923]: ERROR nova.compute.manager [instance: 97148ea5-398a-4974-b8f6-62290e519794] raise exception.PortBindingFailed(port_id=port['id']) [ 550.809830] env[61923]: ERROR nova.compute.manager [instance: 97148ea5-398a-4974-b8f6-62290e519794] nova.exception.PortBindingFailed: Binding failed for port 7ec53f9e-f055-4a57-8228-7efbcdb7360f, please check neutron logs for more information. [ 550.809830] env[61923]: ERROR nova.compute.manager [instance: 97148ea5-398a-4974-b8f6-62290e519794] [ 550.810199] env[61923]: DEBUG nova.compute.utils [None req-7781f812-6fae-4bd6-9d91-ec6098c06372 tempest-InstanceActionsTestJSON-2015629344 tempest-InstanceActionsTestJSON-2015629344-project-member] [instance: 97148ea5-398a-4974-b8f6-62290e519794] Binding failed for port 7ec53f9e-f055-4a57-8228-7efbcdb7360f, please check neutron logs for more information. {{(pid=61923) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 550.811767] env[61923]: DEBUG oslo_concurrency.lockutils [None req-15b9e1af-fe57-407e-a058-c08239279239 tempest-ServersWithSpecificFlavorTestJSON-1806421899 tempest-ServersWithSpecificFlavorTestJSON-1806421899-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 12.505s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 550.814412] env[61923]: INFO nova.compute.claims [None req-15b9e1af-fe57-407e-a058-c08239279239 tempest-ServersWithSpecificFlavorTestJSON-1806421899 tempest-ServersWithSpecificFlavorTestJSON-1806421899-project-member] [instance: 4cec7bce-0136-43ad-898f-086058a505f2] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 550.817491] env[61923]: DEBUG nova.compute.manager [None req-7781f812-6fae-4bd6-9d91-ec6098c06372 tempest-InstanceActionsTestJSON-2015629344 tempest-InstanceActionsTestJSON-2015629344-project-member] [instance: 97148ea5-398a-4974-b8f6-62290e519794] Build of instance 97148ea5-398a-4974-b8f6-62290e519794 was re-scheduled: Binding failed for port 7ec53f9e-f055-4a57-8228-7efbcdb7360f, please check neutron logs for more information. {{(pid=61923) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 550.817491] env[61923]: DEBUG nova.compute.manager [None req-7781f812-6fae-4bd6-9d91-ec6098c06372 tempest-InstanceActionsTestJSON-2015629344 tempest-InstanceActionsTestJSON-2015629344-project-member] [instance: 97148ea5-398a-4974-b8f6-62290e519794] Unplugging VIFs for instance {{(pid=61923) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 550.817764] env[61923]: DEBUG oslo_concurrency.lockutils [None req-7781f812-6fae-4bd6-9d91-ec6098c06372 tempest-InstanceActionsTestJSON-2015629344 tempest-InstanceActionsTestJSON-2015629344-project-member] Acquiring lock "refresh_cache-97148ea5-398a-4974-b8f6-62290e519794" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 550.817962] env[61923]: DEBUG oslo_concurrency.lockutils [None req-7781f812-6fae-4bd6-9d91-ec6098c06372 tempest-InstanceActionsTestJSON-2015629344 tempest-InstanceActionsTestJSON-2015629344-project-member] Acquired lock "refresh_cache-97148ea5-398a-4974-b8f6-62290e519794" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 550.818188] env[61923]: DEBUG nova.network.neutron [None req-7781f812-6fae-4bd6-9d91-ec6098c06372 tempest-InstanceActionsTestJSON-2015629344 tempest-InstanceActionsTestJSON-2015629344-project-member] [instance: 97148ea5-398a-4974-b8f6-62290e519794] Building network info cache for instance {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 550.942625] env[61923]: DEBUG oslo_concurrency.lockutils [None req-d9e2f878-40e6-4b41-8b07-da611f2f81f1 tempest-VolumesAssistedSnapshotsTest-2043366068 tempest-VolumesAssistedSnapshotsTest-2043366068-project-member] Releasing lock "refresh_cache-06bda5ee-c7a8-47d1-9e0c-a5af67d40006" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 550.943555] env[61923]: DEBUG nova.compute.manager [None req-d9e2f878-40e6-4b41-8b07-da611f2f81f1 tempest-VolumesAssistedSnapshotsTest-2043366068 tempest-VolumesAssistedSnapshotsTest-2043366068-project-member] [instance: 06bda5ee-c7a8-47d1-9e0c-a5af67d40006] Start destroying the instance on the hypervisor. {{(pid=61923) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 550.943555] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-d9e2f878-40e6-4b41-8b07-da611f2f81f1 tempest-VolumesAssistedSnapshotsTest-2043366068 tempest-VolumesAssistedSnapshotsTest-2043366068-project-member] [instance: 06bda5ee-c7a8-47d1-9e0c-a5af67d40006] Destroying instance {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 550.943665] env[61923]: DEBUG oslo_concurrency.lockutils [req-b3ca27af-ac7c-41ac-a4e1-309fd5a73cd7 req-950cb064-57db-45e8-b7e1-cb07560f3eea service nova] Acquired lock "refresh_cache-06bda5ee-c7a8-47d1-9e0c-a5af67d40006" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 550.943754] env[61923]: DEBUG nova.network.neutron [req-b3ca27af-ac7c-41ac-a4e1-309fd5a73cd7 req-950cb064-57db-45e8-b7e1-cb07560f3eea service nova] [instance: 06bda5ee-c7a8-47d1-9e0c-a5af67d40006] Refreshing network info cache for port 48788634-af14-4531-8cff-07f9a6f55f4c {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 550.945142] env[61923]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d0477659-d58d-43a9-9851-4ad76a92efb6 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 550.954462] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c382cd2-e5e2-4cba-ac29-e730abacab5d {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 550.977212] env[61923]: WARNING nova.virt.vmwareapi.vmops [None req-d9e2f878-40e6-4b41-8b07-da611f2f81f1 tempest-VolumesAssistedSnapshotsTest-2043366068 tempest-VolumesAssistedSnapshotsTest-2043366068-project-member] [instance: 06bda5ee-c7a8-47d1-9e0c-a5af67d40006] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 06bda5ee-c7a8-47d1-9e0c-a5af67d40006 could not be found. [ 550.977496] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-d9e2f878-40e6-4b41-8b07-da611f2f81f1 tempest-VolumesAssistedSnapshotsTest-2043366068 tempest-VolumesAssistedSnapshotsTest-2043366068-project-member] [instance: 06bda5ee-c7a8-47d1-9e0c-a5af67d40006] Instance destroyed {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 550.977619] env[61923]: INFO nova.compute.manager [None req-d9e2f878-40e6-4b41-8b07-da611f2f81f1 tempest-VolumesAssistedSnapshotsTest-2043366068 tempest-VolumesAssistedSnapshotsTest-2043366068-project-member] [instance: 06bda5ee-c7a8-47d1-9e0c-a5af67d40006] Took 0.03 seconds to destroy the instance on the hypervisor. [ 550.977848] env[61923]: DEBUG oslo.service.loopingcall [None req-d9e2f878-40e6-4b41-8b07-da611f2f81f1 tempest-VolumesAssistedSnapshotsTest-2043366068 tempest-VolumesAssistedSnapshotsTest-2043366068-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61923) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 550.978084] env[61923]: DEBUG nova.compute.manager [-] [instance: 06bda5ee-c7a8-47d1-9e0c-a5af67d40006] Deallocating network for instance {{(pid=61923) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 550.978181] env[61923]: DEBUG nova.network.neutron [-] [instance: 06bda5ee-c7a8-47d1-9e0c-a5af67d40006] deallocate_for_instance() {{(pid=61923) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 550.994980] env[61923]: DEBUG nova.network.neutron [-] [instance: 06bda5ee-c7a8-47d1-9e0c-a5af67d40006] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 551.019322] env[61923]: INFO nova.compute.manager [None req-8a6b6bc2-54c9-4389-ab30-057158d86227 tempest-ServerDiagnosticsNegativeTest-664316905 tempest-ServerDiagnosticsNegativeTest-664316905-project-member] [instance: b0a45314-eed8-4d0a-bea2-d4b47dc0bc80] Took 1.03 seconds to deallocate network for instance. [ 551.045831] env[61923]: DEBUG oslo_vmware.api [None req-3b10f02e-4025-4174-94c7-8515a7e12175 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] Task: {'id': task-1377427, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.023779} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 551.047643] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-3b10f02e-4025-4174-94c7-8515a7e12175 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] Deleted the datastore file {{(pid=61923) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 551.047643] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-3b10f02e-4025-4174-94c7-8515a7e12175 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] Moving file from [datastore1] vmware_temp/458df3c4-aac7-4022-909f-21ebff462779/0f153f63-ae0a-45b1-b7f5-7f9b673c947f to [datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f. {{(pid=61923) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:334}} [ 551.047643] env[61923]: DEBUG oslo_vmware.service [-] Invoking FileManager.MoveDatastoreFile_Task with opID=oslo.vmware-a29e8a09-be07-4aac-ab49-158aac642247 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 551.054970] env[61923]: DEBUG oslo_vmware.api [None req-3b10f02e-4025-4174-94c7-8515a7e12175 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] Waiting for the task: (returnval){ [ 551.054970] env[61923]: value = "task-1377428" [ 551.054970] env[61923]: _type = "Task" [ 551.054970] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 551.067667] env[61923]: DEBUG oslo_vmware.api [None req-3b10f02e-4025-4174-94c7-8515a7e12175 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] Task: {'id': task-1377428, 'name': MoveDatastoreFile_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 551.348119] env[61923]: DEBUG nova.network.neutron [None req-7781f812-6fae-4bd6-9d91-ec6098c06372 tempest-InstanceActionsTestJSON-2015629344 tempest-InstanceActionsTestJSON-2015629344-project-member] [instance: 97148ea5-398a-4974-b8f6-62290e519794] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 551.471233] env[61923]: DEBUG nova.network.neutron [req-b3ca27af-ac7c-41ac-a4e1-309fd5a73cd7 req-950cb064-57db-45e8-b7e1-cb07560f3eea service nova] [instance: 06bda5ee-c7a8-47d1-9e0c-a5af67d40006] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 551.498107] env[61923]: DEBUG nova.network.neutron [-] [instance: 06bda5ee-c7a8-47d1-9e0c-a5af67d40006] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 551.555720] env[61923]: DEBUG nova.network.neutron [None req-7781f812-6fae-4bd6-9d91-ec6098c06372 tempest-InstanceActionsTestJSON-2015629344 tempest-InstanceActionsTestJSON-2015629344-project-member] [instance: 97148ea5-398a-4974-b8f6-62290e519794] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 551.566259] env[61923]: DEBUG oslo_vmware.api [None req-3b10f02e-4025-4174-94c7-8515a7e12175 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] Task: {'id': task-1377428, 'name': MoveDatastoreFile_Task, 'duration_secs': 0.025532} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 551.566559] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-3b10f02e-4025-4174-94c7-8515a7e12175 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] File moved {{(pid=61923) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:346}} [ 551.566753] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-3b10f02e-4025-4174-94c7-8515a7e12175 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] [instance: f7cc960c-e06a-4c58-9367-ec8771fe09d5] Cleaning up location [datastore1] vmware_temp/458df3c4-aac7-4022-909f-21ebff462779 {{(pid=61923) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 551.566917] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-3b10f02e-4025-4174-94c7-8515a7e12175 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] Deleting the datastore file [datastore1] vmware_temp/458df3c4-aac7-4022-909f-21ebff462779 {{(pid=61923) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 551.567489] env[61923]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-3e9ee20a-6fc6-401d-8307-bc79cb4b219a {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 551.575035] env[61923]: DEBUG oslo_vmware.api [None req-3b10f02e-4025-4174-94c7-8515a7e12175 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] Waiting for the task: (returnval){ [ 551.575035] env[61923]: value = "task-1377429" [ 551.575035] env[61923]: _type = "Task" [ 551.575035] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 551.584824] env[61923]: DEBUG oslo_vmware.api [None req-3b10f02e-4025-4174-94c7-8515a7e12175 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] Task: {'id': task-1377429, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 551.695445] env[61923]: DEBUG nova.network.neutron [req-b3ca27af-ac7c-41ac-a4e1-309fd5a73cd7 req-950cb064-57db-45e8-b7e1-cb07560f3eea service nova] [instance: 06bda5ee-c7a8-47d1-9e0c-a5af67d40006] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 552.001231] env[61923]: INFO nova.compute.manager [-] [instance: 06bda5ee-c7a8-47d1-9e0c-a5af67d40006] Took 1.02 seconds to deallocate network for instance. [ 552.005278] env[61923]: DEBUG nova.compute.claims [None req-d9e2f878-40e6-4b41-8b07-da611f2f81f1 tempest-VolumesAssistedSnapshotsTest-2043366068 tempest-VolumesAssistedSnapshotsTest-2043366068-project-member] [instance: 06bda5ee-c7a8-47d1-9e0c-a5af67d40006] Aborting claim: {{(pid=61923) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 552.005459] env[61923]: DEBUG oslo_concurrency.lockutils [None req-d9e2f878-40e6-4b41-8b07-da611f2f81f1 tempest-VolumesAssistedSnapshotsTest-2043366068 tempest-VolumesAssistedSnapshotsTest-2043366068-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 552.054494] env[61923]: INFO nova.scheduler.client.report [None req-8a6b6bc2-54c9-4389-ab30-057158d86227 tempest-ServerDiagnosticsNegativeTest-664316905 tempest-ServerDiagnosticsNegativeTest-664316905-project-member] Deleted allocations for instance b0a45314-eed8-4d0a-bea2-d4b47dc0bc80 [ 552.061149] env[61923]: DEBUG oslo_concurrency.lockutils [None req-7781f812-6fae-4bd6-9d91-ec6098c06372 tempest-InstanceActionsTestJSON-2015629344 tempest-InstanceActionsTestJSON-2015629344-project-member] Releasing lock "refresh_cache-97148ea5-398a-4974-b8f6-62290e519794" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 552.061378] env[61923]: DEBUG nova.compute.manager [None req-7781f812-6fae-4bd6-9d91-ec6098c06372 tempest-InstanceActionsTestJSON-2015629344 tempest-InstanceActionsTestJSON-2015629344-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61923) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 552.061557] env[61923]: DEBUG nova.compute.manager [None req-7781f812-6fae-4bd6-9d91-ec6098c06372 tempest-InstanceActionsTestJSON-2015629344 tempest-InstanceActionsTestJSON-2015629344-project-member] [instance: 97148ea5-398a-4974-b8f6-62290e519794] Deallocating network for instance {{(pid=61923) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 552.061747] env[61923]: DEBUG nova.network.neutron [None req-7781f812-6fae-4bd6-9d91-ec6098c06372 tempest-InstanceActionsTestJSON-2015629344 tempest-InstanceActionsTestJSON-2015629344-project-member] [instance: 97148ea5-398a-4974-b8f6-62290e519794] deallocate_for_instance() {{(pid=61923) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 552.091733] env[61923]: DEBUG oslo_vmware.api [None req-3b10f02e-4025-4174-94c7-8515a7e12175 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] Task: {'id': task-1377429, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.025989} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 552.095200] env[61923]: DEBUG nova.network.neutron [None req-7781f812-6fae-4bd6-9d91-ec6098c06372 tempest-InstanceActionsTestJSON-2015629344 tempest-InstanceActionsTestJSON-2015629344-project-member] [instance: 97148ea5-398a-4974-b8f6-62290e519794] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 552.097101] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-3b10f02e-4025-4174-94c7-8515a7e12175 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] Deleted the datastore file {{(pid=61923) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 552.098294] env[61923]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6186be3f-4553-4834-818e-88b70b4a61c2 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 552.104751] env[61923]: DEBUG oslo_vmware.api [None req-3b10f02e-4025-4174-94c7-8515a7e12175 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] Waiting for the task: (returnval){ [ 552.104751] env[61923]: value = "session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52159443-b2fe-198b-2ba1-f57c8a79a399" [ 552.104751] env[61923]: _type = "Task" [ 552.104751] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 552.116782] env[61923]: DEBUG oslo_vmware.api [None req-3b10f02e-4025-4174-94c7-8515a7e12175 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52159443-b2fe-198b-2ba1-f57c8a79a399, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 552.200541] env[61923]: DEBUG oslo_concurrency.lockutils [req-b3ca27af-ac7c-41ac-a4e1-309fd5a73cd7 req-950cb064-57db-45e8-b7e1-cb07560f3eea service nova] Releasing lock "refresh_cache-06bda5ee-c7a8-47d1-9e0c-a5af67d40006" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 552.216911] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dab76753-c6e0-4dbb-aa8c-b4c2e8d4fb9b {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 552.225465] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8747ba6d-5ba3-4ff6-ac7d-451eea3c9040 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 552.261237] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb4f7036-089c-4367-92d7-0665648d44c1 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 552.271479] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c718f29-6253-4bbd-b5f1-7f705ea0227d {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 552.285504] env[61923]: DEBUG nova.compute.provider_tree [None req-15b9e1af-fe57-407e-a058-c08239279239 tempest-ServersWithSpecificFlavorTestJSON-1806421899 tempest-ServersWithSpecificFlavorTestJSON-1806421899-project-member] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 552.566959] env[61923]: DEBUG oslo_concurrency.lockutils [None req-8a6b6bc2-54c9-4389-ab30-057158d86227 tempest-ServerDiagnosticsNegativeTest-664316905 tempest-ServerDiagnosticsNegativeTest-664316905-project-member] Lock "b0a45314-eed8-4d0a-bea2-d4b47dc0bc80" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 34.693s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 552.599985] env[61923]: DEBUG nova.network.neutron [None req-7781f812-6fae-4bd6-9d91-ec6098c06372 tempest-InstanceActionsTestJSON-2015629344 tempest-InstanceActionsTestJSON-2015629344-project-member] [instance: 97148ea5-398a-4974-b8f6-62290e519794] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 552.618791] env[61923]: DEBUG oslo_vmware.api [None req-3b10f02e-4025-4174-94c7-8515a7e12175 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52159443-b2fe-198b-2ba1-f57c8a79a399, 'name': SearchDatastore_Task, 'duration_secs': 0.009553} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 552.619872] env[61923]: DEBUG oslo_concurrency.lockutils [None req-3b10f02e-4025-4174-94c7-8515a7e12175 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 552.620287] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-3b10f02e-4025-4174-94c7-8515a7e12175 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk to [datastore1] f7cc960c-e06a-4c58-9367-ec8771fe09d5/f7cc960c-e06a-4c58-9367-ec8771fe09d5.vmdk {{(pid=61923) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 552.620683] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-6a01c578-45e4-4c4b-bf22-3d94759b44ba {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 552.630675] env[61923]: DEBUG oslo_vmware.api [None req-3b10f02e-4025-4174-94c7-8515a7e12175 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] Waiting for the task: (returnval){ [ 552.630675] env[61923]: value = "task-1377430" [ 552.630675] env[61923]: _type = "Task" [ 552.630675] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 552.648904] env[61923]: DEBUG oslo_vmware.api [None req-3b10f02e-4025-4174-94c7-8515a7e12175 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] Task: {'id': task-1377430, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 552.769534] env[61923]: DEBUG nova.compute.manager [req-730ac6ce-751e-473d-9b3c-7a478aa6ffbc req-7550fc6d-6d77-49b9-a848-01d035a77f0a service nova] [instance: 06bda5ee-c7a8-47d1-9e0c-a5af67d40006] Received event network-vif-deleted-48788634-af14-4531-8cff-07f9a6f55f4c {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 552.789509] env[61923]: DEBUG nova.scheduler.client.report [None req-15b9e1af-fe57-407e-a058-c08239279239 tempest-ServersWithSpecificFlavorTestJSON-1806421899 tempest-ServersWithSpecificFlavorTestJSON-1806421899-project-member] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 553.070270] env[61923]: DEBUG nova.compute.manager [None req-fba27a13-474d-4fe3-9187-c31143f40aa5 tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] [instance: 9c3e2f94-91da-4255-b4d6-1c01e100ff48] Starting instance... {{(pid=61923) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 553.107037] env[61923]: INFO nova.compute.manager [None req-7781f812-6fae-4bd6-9d91-ec6098c06372 tempest-InstanceActionsTestJSON-2015629344 tempest-InstanceActionsTestJSON-2015629344-project-member] [instance: 97148ea5-398a-4974-b8f6-62290e519794] Took 1.04 seconds to deallocate network for instance. [ 553.148141] env[61923]: DEBUG oslo_vmware.api [None req-3b10f02e-4025-4174-94c7-8515a7e12175 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] Task: {'id': task-1377430, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 553.298421] env[61923]: DEBUG oslo_concurrency.lockutils [None req-15b9e1af-fe57-407e-a058-c08239279239 tempest-ServersWithSpecificFlavorTestJSON-1806421899 tempest-ServersWithSpecificFlavorTestJSON-1806421899-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.487s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 553.299015] env[61923]: DEBUG nova.compute.manager [None req-15b9e1af-fe57-407e-a058-c08239279239 tempest-ServersWithSpecificFlavorTestJSON-1806421899 tempest-ServersWithSpecificFlavorTestJSON-1806421899-project-member] [instance: 4cec7bce-0136-43ad-898f-086058a505f2] Start building networks asynchronously for instance. {{(pid=61923) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 553.301777] env[61923]: DEBUG oslo_concurrency.lockutils [None req-844a2215-1038-4a30-8a79-40c1c3117bde tempest-FloatingIPsAssociationNegativeTestJSON-1301210771 tempest-FloatingIPsAssociationNegativeTestJSON-1301210771-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 13.806s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 553.599187] env[61923]: DEBUG oslo_concurrency.lockutils [None req-fba27a13-474d-4fe3-9187-c31143f40aa5 tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 553.651953] env[61923]: DEBUG oslo_vmware.api [None req-3b10f02e-4025-4174-94c7-8515a7e12175 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] Task: {'id': task-1377430, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 553.806768] env[61923]: DEBUG nova.compute.utils [None req-15b9e1af-fe57-407e-a058-c08239279239 tempest-ServersWithSpecificFlavorTestJSON-1806421899 tempest-ServersWithSpecificFlavorTestJSON-1806421899-project-member] Using /dev/sd instead of None {{(pid=61923) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 553.813486] env[61923]: DEBUG nova.compute.manager [None req-15b9e1af-fe57-407e-a058-c08239279239 tempest-ServersWithSpecificFlavorTestJSON-1806421899 tempest-ServersWithSpecificFlavorTestJSON-1806421899-project-member] [instance: 4cec7bce-0136-43ad-898f-086058a505f2] Allocating IP information in the background. {{(pid=61923) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 553.814981] env[61923]: DEBUG nova.network.neutron [None req-15b9e1af-fe57-407e-a058-c08239279239 tempest-ServersWithSpecificFlavorTestJSON-1806421899 tempest-ServersWithSpecificFlavorTestJSON-1806421899-project-member] [instance: 4cec7bce-0136-43ad-898f-086058a505f2] allocate_for_instance() {{(pid=61923) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 553.965944] env[61923]: DEBUG nova.policy [None req-15b9e1af-fe57-407e-a058-c08239279239 tempest-ServersWithSpecificFlavorTestJSON-1806421899 tempest-ServersWithSpecificFlavorTestJSON-1806421899-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '95a2937a775445fb9de3bd6612bb304c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0563f123be6344f98ec0fea13deb2b72', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61923) authorize /opt/stack/nova/nova/policy.py:201}} [ 554.151992] env[61923]: DEBUG oslo_vmware.api [None req-3b10f02e-4025-4174-94c7-8515a7e12175 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] Task: {'id': task-1377430, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.038578} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 554.156770] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-3b10f02e-4025-4174-94c7-8515a7e12175 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk to [datastore1] f7cc960c-e06a-4c58-9367-ec8771fe09d5/f7cc960c-e06a-4c58-9367-ec8771fe09d5.vmdk {{(pid=61923) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 554.156770] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-3b10f02e-4025-4174-94c7-8515a7e12175 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] [instance: f7cc960c-e06a-4c58-9367-ec8771fe09d5] Extending root virtual disk to 1048576 {{(pid=61923) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 554.156770] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-7e87a4e9-44a6-4b91-a62e-fc0394caa5e0 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 554.169442] env[61923]: DEBUG oslo_vmware.api [None req-3b10f02e-4025-4174-94c7-8515a7e12175 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] Waiting for the task: (returnval){ [ 554.169442] env[61923]: value = "task-1377431" [ 554.169442] env[61923]: _type = "Task" [ 554.169442] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 554.183982] env[61923]: DEBUG oslo_vmware.api [None req-3b10f02e-4025-4174-94c7-8515a7e12175 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] Task: {'id': task-1377431, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 554.214133] env[61923]: INFO nova.scheduler.client.report [None req-7781f812-6fae-4bd6-9d91-ec6098c06372 tempest-InstanceActionsTestJSON-2015629344 tempest-InstanceActionsTestJSON-2015629344-project-member] Deleted allocations for instance 97148ea5-398a-4974-b8f6-62290e519794 [ 554.317705] env[61923]: DEBUG nova.compute.manager [None req-15b9e1af-fe57-407e-a058-c08239279239 tempest-ServersWithSpecificFlavorTestJSON-1806421899 tempest-ServersWithSpecificFlavorTestJSON-1806421899-project-member] [instance: 4cec7bce-0136-43ad-898f-086058a505f2] Start building block device mappings for instance. {{(pid=61923) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 554.373055] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35ecfb90-acd3-4ab1-a3c0-cf63579898a7 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 554.383545] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-553e624f-f55a-408d-ad27-62e43e7dc072 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 554.421755] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66371935-fa50-4afb-a1d9-06411e121601 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 554.430800] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f5fd63e-7a79-471d-b6ee-2942fdbae6c4 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 554.466269] env[61923]: DEBUG nova.compute.provider_tree [None req-844a2215-1038-4a30-8a79-40c1c3117bde tempest-FloatingIPsAssociationNegativeTestJSON-1301210771 tempest-FloatingIPsAssociationNegativeTestJSON-1301210771-project-member] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 554.685932] env[61923]: DEBUG oslo_vmware.api [None req-3b10f02e-4025-4174-94c7-8515a7e12175 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] Task: {'id': task-1377431, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.074872} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 554.686536] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-3b10f02e-4025-4174-94c7-8515a7e12175 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] [instance: f7cc960c-e06a-4c58-9367-ec8771fe09d5] Extended root virtual disk {{(pid=61923) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 554.688151] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da9cb5b2-57d1-4748-b985-14717d9d8e0a {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 554.715473] env[61923]: DEBUG nova.virt.vmwareapi.volumeops [None req-3b10f02e-4025-4174-94c7-8515a7e12175 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] [instance: f7cc960c-e06a-4c58-9367-ec8771fe09d5] Reconfiguring VM instance instance-0000000a to attach disk [datastore1] f7cc960c-e06a-4c58-9367-ec8771fe09d5/f7cc960c-e06a-4c58-9367-ec8771fe09d5.vmdk or device None with type sparse {{(pid=61923) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 554.715784] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e9b1538d-899f-4569-b4bb-e18254dbf5aa {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 554.731661] env[61923]: DEBUG oslo_concurrency.lockutils [None req-7781f812-6fae-4bd6-9d91-ec6098c06372 tempest-InstanceActionsTestJSON-2015629344 tempest-InstanceActionsTestJSON-2015629344-project-member] Lock "97148ea5-398a-4974-b8f6-62290e519794" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 34.668s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 554.738413] env[61923]: DEBUG oslo_vmware.api [None req-3b10f02e-4025-4174-94c7-8515a7e12175 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] Waiting for the task: (returnval){ [ 554.738413] env[61923]: value = "task-1377432" [ 554.738413] env[61923]: _type = "Task" [ 554.738413] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 554.748910] env[61923]: DEBUG oslo_vmware.api [None req-3b10f02e-4025-4174-94c7-8515a7e12175 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] Task: {'id': task-1377432, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 554.971961] env[61923]: DEBUG nova.scheduler.client.report [None req-844a2215-1038-4a30-8a79-40c1c3117bde tempest-FloatingIPsAssociationNegativeTestJSON-1301210771 tempest-FloatingIPsAssociationNegativeTestJSON-1301210771-project-member] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 555.059317] env[61923]: DEBUG nova.network.neutron [None req-15b9e1af-fe57-407e-a058-c08239279239 tempest-ServersWithSpecificFlavorTestJSON-1806421899 tempest-ServersWithSpecificFlavorTestJSON-1806421899-project-member] [instance: 4cec7bce-0136-43ad-898f-086058a505f2] Successfully created port: fe27e2eb-c2cb-49f1-84e8-03cd81ad3229 {{(pid=61923) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 555.235576] env[61923]: DEBUG nova.compute.manager [None req-fba27a13-474d-4fe3-9187-c31143f40aa5 tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] [instance: 93818ffd-eb85-48f0-aebd-9774467b0a8f] Starting instance... {{(pid=61923) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 555.253745] env[61923]: DEBUG oslo_vmware.api [None req-3b10f02e-4025-4174-94c7-8515a7e12175 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] Task: {'id': task-1377432, 'name': ReconfigVM_Task, 'duration_secs': 0.289197} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 555.254209] env[61923]: DEBUG nova.virt.vmwareapi.volumeops [None req-3b10f02e-4025-4174-94c7-8515a7e12175 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] [instance: f7cc960c-e06a-4c58-9367-ec8771fe09d5] Reconfigured VM instance instance-0000000a to attach disk [datastore1] f7cc960c-e06a-4c58-9367-ec8771fe09d5/f7cc960c-e06a-4c58-9367-ec8771fe09d5.vmdk or device None with type sparse {{(pid=61923) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 555.256182] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b9f306f4-fe9a-435c-80e6-d65ca6077702 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 555.265340] env[61923]: DEBUG oslo_vmware.api [None req-3b10f02e-4025-4174-94c7-8515a7e12175 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] Waiting for the task: (returnval){ [ 555.265340] env[61923]: value = "task-1377433" [ 555.265340] env[61923]: _type = "Task" [ 555.265340] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 555.279689] env[61923]: DEBUG oslo_vmware.api [None req-3b10f02e-4025-4174-94c7-8515a7e12175 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] Task: {'id': task-1377433, 'name': Rename_Task} progress is 6%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 555.333456] env[61923]: DEBUG nova.compute.manager [None req-15b9e1af-fe57-407e-a058-c08239279239 tempest-ServersWithSpecificFlavorTestJSON-1806421899 tempest-ServersWithSpecificFlavorTestJSON-1806421899-project-member] [instance: 4cec7bce-0136-43ad-898f-086058a505f2] Start spawning the instance on the hypervisor. {{(pid=61923) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 555.369178] env[61923]: DEBUG nova.virt.hardware [None req-15b9e1af-fe57-407e-a058-c08239279239 tempest-ServersWithSpecificFlavorTestJSON-1806421899 tempest-ServersWithSpecificFlavorTestJSON-1806421899-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-29T20:09:43Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='1920019676',id=20,is_public=True,memory_mb=192,name='tempest-flavor_with_ephemeral_0-1848395615',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-29T20:07:35Z,direct_url=,disk_format='vmdk',id=0f153f63-ae0a-45b1-b7f5-7f9b673c947f,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='4e7b5e3b3c3443c8bd5c70f8a15739f5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-29T20:07:36Z,virtual_size=,visibility=), allow threads: False {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 555.369618] env[61923]: DEBUG nova.virt.hardware [None req-15b9e1af-fe57-407e-a058-c08239279239 tempest-ServersWithSpecificFlavorTestJSON-1806421899 tempest-ServersWithSpecificFlavorTestJSON-1806421899-project-member] Flavor limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 555.370014] env[61923]: DEBUG nova.virt.hardware [None req-15b9e1af-fe57-407e-a058-c08239279239 tempest-ServersWithSpecificFlavorTestJSON-1806421899 tempest-ServersWithSpecificFlavorTestJSON-1806421899-project-member] Image limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 555.370218] env[61923]: DEBUG nova.virt.hardware [None req-15b9e1af-fe57-407e-a058-c08239279239 tempest-ServersWithSpecificFlavorTestJSON-1806421899 tempest-ServersWithSpecificFlavorTestJSON-1806421899-project-member] Flavor pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 555.370519] env[61923]: DEBUG nova.virt.hardware [None req-15b9e1af-fe57-407e-a058-c08239279239 tempest-ServersWithSpecificFlavorTestJSON-1806421899 tempest-ServersWithSpecificFlavorTestJSON-1806421899-project-member] Image pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 555.370598] env[61923]: DEBUG nova.virt.hardware [None req-15b9e1af-fe57-407e-a058-c08239279239 tempest-ServersWithSpecificFlavorTestJSON-1806421899 tempest-ServersWithSpecificFlavorTestJSON-1806421899-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 555.370773] env[61923]: DEBUG nova.virt.hardware [None req-15b9e1af-fe57-407e-a058-c08239279239 tempest-ServersWithSpecificFlavorTestJSON-1806421899 tempest-ServersWithSpecificFlavorTestJSON-1806421899-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 555.371066] env[61923]: DEBUG nova.virt.hardware [None req-15b9e1af-fe57-407e-a058-c08239279239 tempest-ServersWithSpecificFlavorTestJSON-1806421899 tempest-ServersWithSpecificFlavorTestJSON-1806421899-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 555.371163] env[61923]: DEBUG nova.virt.hardware [None req-15b9e1af-fe57-407e-a058-c08239279239 tempest-ServersWithSpecificFlavorTestJSON-1806421899 tempest-ServersWithSpecificFlavorTestJSON-1806421899-project-member] Got 1 possible topologies {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 555.371246] env[61923]: DEBUG nova.virt.hardware [None req-15b9e1af-fe57-407e-a058-c08239279239 tempest-ServersWithSpecificFlavorTestJSON-1806421899 tempest-ServersWithSpecificFlavorTestJSON-1806421899-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 555.371410] env[61923]: DEBUG nova.virt.hardware [None req-15b9e1af-fe57-407e-a058-c08239279239 tempest-ServersWithSpecificFlavorTestJSON-1806421899 tempest-ServersWithSpecificFlavorTestJSON-1806421899-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 555.372948] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4613d10e-0808-407c-ac4a-08e98aba461e {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 555.379982] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f318ba3d-e422-428b-934f-98ac98f29b8e {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 555.479270] env[61923]: DEBUG oslo_concurrency.lockutils [None req-844a2215-1038-4a30-8a79-40c1c3117bde tempest-FloatingIPsAssociationNegativeTestJSON-1301210771 tempest-FloatingIPsAssociationNegativeTestJSON-1301210771-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.177s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 555.480596] env[61923]: ERROR nova.compute.manager [None req-844a2215-1038-4a30-8a79-40c1c3117bde tempest-FloatingIPsAssociationNegativeTestJSON-1301210771 tempest-FloatingIPsAssociationNegativeTestJSON-1301210771-project-member] [instance: 5d57303a-ae17-4b92-8f12-303677eb0d7e] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 4f3fea2e-b3d9-4c89-b291-071aa6707d27, please check neutron logs for more information. [ 555.480596] env[61923]: ERROR nova.compute.manager [instance: 5d57303a-ae17-4b92-8f12-303677eb0d7e] Traceback (most recent call last): [ 555.480596] env[61923]: ERROR nova.compute.manager [instance: 5d57303a-ae17-4b92-8f12-303677eb0d7e] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 555.480596] env[61923]: ERROR nova.compute.manager [instance: 5d57303a-ae17-4b92-8f12-303677eb0d7e] self.driver.spawn(context, instance, image_meta, [ 555.480596] env[61923]: ERROR nova.compute.manager [instance: 5d57303a-ae17-4b92-8f12-303677eb0d7e] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 555.480596] env[61923]: ERROR nova.compute.manager [instance: 5d57303a-ae17-4b92-8f12-303677eb0d7e] self._vmops.spawn(context, instance, image_meta, injected_files, [ 555.480596] env[61923]: ERROR nova.compute.manager [instance: 5d57303a-ae17-4b92-8f12-303677eb0d7e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 555.480596] env[61923]: ERROR nova.compute.manager [instance: 5d57303a-ae17-4b92-8f12-303677eb0d7e] vm_ref = self.build_virtual_machine(instance, [ 555.480596] env[61923]: ERROR nova.compute.manager [instance: 5d57303a-ae17-4b92-8f12-303677eb0d7e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 555.480596] env[61923]: ERROR nova.compute.manager [instance: 5d57303a-ae17-4b92-8f12-303677eb0d7e] vif_infos = vmwarevif.get_vif_info(self._session, [ 555.480596] env[61923]: ERROR nova.compute.manager [instance: 5d57303a-ae17-4b92-8f12-303677eb0d7e] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 555.481052] env[61923]: ERROR nova.compute.manager [instance: 5d57303a-ae17-4b92-8f12-303677eb0d7e] for vif in network_info: [ 555.481052] env[61923]: ERROR nova.compute.manager [instance: 5d57303a-ae17-4b92-8f12-303677eb0d7e] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 555.481052] env[61923]: ERROR nova.compute.manager [instance: 5d57303a-ae17-4b92-8f12-303677eb0d7e] return self._sync_wrapper(fn, *args, **kwargs) [ 555.481052] env[61923]: ERROR nova.compute.manager [instance: 5d57303a-ae17-4b92-8f12-303677eb0d7e] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 555.481052] env[61923]: ERROR nova.compute.manager [instance: 5d57303a-ae17-4b92-8f12-303677eb0d7e] self.wait() [ 555.481052] env[61923]: ERROR nova.compute.manager [instance: 5d57303a-ae17-4b92-8f12-303677eb0d7e] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 555.481052] env[61923]: ERROR nova.compute.manager [instance: 5d57303a-ae17-4b92-8f12-303677eb0d7e] self[:] = self._gt.wait() [ 555.481052] env[61923]: ERROR nova.compute.manager [instance: 5d57303a-ae17-4b92-8f12-303677eb0d7e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 555.481052] env[61923]: ERROR nova.compute.manager [instance: 5d57303a-ae17-4b92-8f12-303677eb0d7e] return self._exit_event.wait() [ 555.481052] env[61923]: ERROR nova.compute.manager [instance: 5d57303a-ae17-4b92-8f12-303677eb0d7e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 555.481052] env[61923]: ERROR nova.compute.manager [instance: 5d57303a-ae17-4b92-8f12-303677eb0d7e] result = hub.switch() [ 555.481052] env[61923]: ERROR nova.compute.manager [instance: 5d57303a-ae17-4b92-8f12-303677eb0d7e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 555.481052] env[61923]: ERROR nova.compute.manager [instance: 5d57303a-ae17-4b92-8f12-303677eb0d7e] return self.greenlet.switch() [ 555.481416] env[61923]: ERROR nova.compute.manager [instance: 5d57303a-ae17-4b92-8f12-303677eb0d7e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 555.481416] env[61923]: ERROR nova.compute.manager [instance: 5d57303a-ae17-4b92-8f12-303677eb0d7e] result = function(*args, **kwargs) [ 555.481416] env[61923]: ERROR nova.compute.manager [instance: 5d57303a-ae17-4b92-8f12-303677eb0d7e] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 555.481416] env[61923]: ERROR nova.compute.manager [instance: 5d57303a-ae17-4b92-8f12-303677eb0d7e] return func(*args, **kwargs) [ 555.481416] env[61923]: ERROR nova.compute.manager [instance: 5d57303a-ae17-4b92-8f12-303677eb0d7e] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 555.481416] env[61923]: ERROR nova.compute.manager [instance: 5d57303a-ae17-4b92-8f12-303677eb0d7e] raise e [ 555.481416] env[61923]: ERROR nova.compute.manager [instance: 5d57303a-ae17-4b92-8f12-303677eb0d7e] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 555.481416] env[61923]: ERROR nova.compute.manager [instance: 5d57303a-ae17-4b92-8f12-303677eb0d7e] nwinfo = self.network_api.allocate_for_instance( [ 555.481416] env[61923]: ERROR nova.compute.manager [instance: 5d57303a-ae17-4b92-8f12-303677eb0d7e] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 555.481416] env[61923]: ERROR nova.compute.manager [instance: 5d57303a-ae17-4b92-8f12-303677eb0d7e] created_port_ids = self._update_ports_for_instance( [ 555.481416] env[61923]: ERROR nova.compute.manager [instance: 5d57303a-ae17-4b92-8f12-303677eb0d7e] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 555.481416] env[61923]: ERROR nova.compute.manager [instance: 5d57303a-ae17-4b92-8f12-303677eb0d7e] with excutils.save_and_reraise_exception(): [ 555.481416] env[61923]: ERROR nova.compute.manager [instance: 5d57303a-ae17-4b92-8f12-303677eb0d7e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 555.481800] env[61923]: ERROR nova.compute.manager [instance: 5d57303a-ae17-4b92-8f12-303677eb0d7e] self.force_reraise() [ 555.481800] env[61923]: ERROR nova.compute.manager [instance: 5d57303a-ae17-4b92-8f12-303677eb0d7e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 555.481800] env[61923]: ERROR nova.compute.manager [instance: 5d57303a-ae17-4b92-8f12-303677eb0d7e] raise self.value [ 555.481800] env[61923]: ERROR nova.compute.manager [instance: 5d57303a-ae17-4b92-8f12-303677eb0d7e] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 555.481800] env[61923]: ERROR nova.compute.manager [instance: 5d57303a-ae17-4b92-8f12-303677eb0d7e] updated_port = self._update_port( [ 555.481800] env[61923]: ERROR nova.compute.manager [instance: 5d57303a-ae17-4b92-8f12-303677eb0d7e] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 555.481800] env[61923]: ERROR nova.compute.manager [instance: 5d57303a-ae17-4b92-8f12-303677eb0d7e] _ensure_no_port_binding_failure(port) [ 555.481800] env[61923]: ERROR nova.compute.manager [instance: 5d57303a-ae17-4b92-8f12-303677eb0d7e] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 555.481800] env[61923]: ERROR nova.compute.manager [instance: 5d57303a-ae17-4b92-8f12-303677eb0d7e] raise exception.PortBindingFailed(port_id=port['id']) [ 555.481800] env[61923]: ERROR nova.compute.manager [instance: 5d57303a-ae17-4b92-8f12-303677eb0d7e] nova.exception.PortBindingFailed: Binding failed for port 4f3fea2e-b3d9-4c89-b291-071aa6707d27, please check neutron logs for more information. [ 555.481800] env[61923]: ERROR nova.compute.manager [instance: 5d57303a-ae17-4b92-8f12-303677eb0d7e] [ 555.482620] env[61923]: DEBUG nova.compute.utils [None req-844a2215-1038-4a30-8a79-40c1c3117bde tempest-FloatingIPsAssociationNegativeTestJSON-1301210771 tempest-FloatingIPsAssociationNegativeTestJSON-1301210771-project-member] [instance: 5d57303a-ae17-4b92-8f12-303677eb0d7e] Binding failed for port 4f3fea2e-b3d9-4c89-b291-071aa6707d27, please check neutron logs for more information. {{(pid=61923) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 555.483919] env[61923]: DEBUG oslo_concurrency.lockutils [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 15.427s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 555.485672] env[61923]: DEBUG nova.compute.manager [None req-844a2215-1038-4a30-8a79-40c1c3117bde tempest-FloatingIPsAssociationNegativeTestJSON-1301210771 tempest-FloatingIPsAssociationNegativeTestJSON-1301210771-project-member] [instance: 5d57303a-ae17-4b92-8f12-303677eb0d7e] Build of instance 5d57303a-ae17-4b92-8f12-303677eb0d7e was re-scheduled: Binding failed for port 4f3fea2e-b3d9-4c89-b291-071aa6707d27, please check neutron logs for more information. {{(pid=61923) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 555.486324] env[61923]: DEBUG nova.compute.manager [None req-844a2215-1038-4a30-8a79-40c1c3117bde tempest-FloatingIPsAssociationNegativeTestJSON-1301210771 tempest-FloatingIPsAssociationNegativeTestJSON-1301210771-project-member] [instance: 5d57303a-ae17-4b92-8f12-303677eb0d7e] Unplugging VIFs for instance {{(pid=61923) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 555.486614] env[61923]: DEBUG oslo_concurrency.lockutils [None req-844a2215-1038-4a30-8a79-40c1c3117bde tempest-FloatingIPsAssociationNegativeTestJSON-1301210771 tempest-FloatingIPsAssociationNegativeTestJSON-1301210771-project-member] Acquiring lock "refresh_cache-5d57303a-ae17-4b92-8f12-303677eb0d7e" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 555.486764] env[61923]: DEBUG oslo_concurrency.lockutils [None req-844a2215-1038-4a30-8a79-40c1c3117bde tempest-FloatingIPsAssociationNegativeTestJSON-1301210771 tempest-FloatingIPsAssociationNegativeTestJSON-1301210771-project-member] Acquired lock "refresh_cache-5d57303a-ae17-4b92-8f12-303677eb0d7e" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 555.486920] env[61923]: DEBUG nova.network.neutron [None req-844a2215-1038-4a30-8a79-40c1c3117bde tempest-FloatingIPsAssociationNegativeTestJSON-1301210771 tempest-FloatingIPsAssociationNegativeTestJSON-1301210771-project-member] [instance: 5d57303a-ae17-4b92-8f12-303677eb0d7e] Building network info cache for instance {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 555.773672] env[61923]: DEBUG oslo_concurrency.lockutils [None req-fba27a13-474d-4fe3-9187-c31143f40aa5 tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 555.784120] env[61923]: DEBUG oslo_vmware.api [None req-3b10f02e-4025-4174-94c7-8515a7e12175 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] Task: {'id': task-1377433, 'name': Rename_Task, 'duration_secs': 0.142316} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 555.784120] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-3b10f02e-4025-4174-94c7-8515a7e12175 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] [instance: f7cc960c-e06a-4c58-9367-ec8771fe09d5] Powering on the VM {{(pid=61923) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 555.784120] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c7ed7dfc-4d69-43fa-8438-f0f586b04982 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 555.791115] env[61923]: DEBUG oslo_vmware.api [None req-3b10f02e-4025-4174-94c7-8515a7e12175 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] Waiting for the task: (returnval){ [ 555.791115] env[61923]: value = "task-1377434" [ 555.791115] env[61923]: _type = "Task" [ 555.791115] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 555.802031] env[61923]: DEBUG oslo_vmware.api [None req-3b10f02e-4025-4174-94c7-8515a7e12175 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] Task: {'id': task-1377434, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 556.039050] env[61923]: DEBUG nova.network.neutron [None req-844a2215-1038-4a30-8a79-40c1c3117bde tempest-FloatingIPsAssociationNegativeTestJSON-1301210771 tempest-FloatingIPsAssociationNegativeTestJSON-1301210771-project-member] [instance: 5d57303a-ae17-4b92-8f12-303677eb0d7e] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 556.301415] env[61923]: DEBUG oslo_vmware.api [None req-3b10f02e-4025-4174-94c7-8515a7e12175 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] Task: {'id': task-1377434, 'name': PowerOnVM_Task} progress is 87%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 556.444879] env[61923]: DEBUG nova.network.neutron [None req-844a2215-1038-4a30-8a79-40c1c3117bde tempest-FloatingIPsAssociationNegativeTestJSON-1301210771 tempest-FloatingIPsAssociationNegativeTestJSON-1301210771-project-member] [instance: 5d57303a-ae17-4b92-8f12-303677eb0d7e] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 556.806123] env[61923]: DEBUG oslo_vmware.api [None req-3b10f02e-4025-4174-94c7-8515a7e12175 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] Task: {'id': task-1377434, 'name': PowerOnVM_Task, 'duration_secs': 0.619133} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 556.806495] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-3b10f02e-4025-4174-94c7-8515a7e12175 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] [instance: f7cc960c-e06a-4c58-9367-ec8771fe09d5] Powered on the VM {{(pid=61923) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 556.806679] env[61923]: INFO nova.compute.manager [None req-3b10f02e-4025-4174-94c7-8515a7e12175 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] [instance: f7cc960c-e06a-4c58-9367-ec8771fe09d5] Took 10.30 seconds to spawn the instance on the hypervisor. [ 556.806892] env[61923]: DEBUG nova.compute.manager [None req-3b10f02e-4025-4174-94c7-8515a7e12175 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] [instance: f7cc960c-e06a-4c58-9367-ec8771fe09d5] Checking state {{(pid=61923) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 556.807705] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a02a160-d77d-400a-9121-dc446f6f5e83 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 556.950646] env[61923]: DEBUG oslo_concurrency.lockutils [None req-844a2215-1038-4a30-8a79-40c1c3117bde tempest-FloatingIPsAssociationNegativeTestJSON-1301210771 tempest-FloatingIPsAssociationNegativeTestJSON-1301210771-project-member] Releasing lock "refresh_cache-5d57303a-ae17-4b92-8f12-303677eb0d7e" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 556.950943] env[61923]: DEBUG nova.compute.manager [None req-844a2215-1038-4a30-8a79-40c1c3117bde tempest-FloatingIPsAssociationNegativeTestJSON-1301210771 tempest-FloatingIPsAssociationNegativeTestJSON-1301210771-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61923) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 556.953162] env[61923]: DEBUG nova.compute.manager [None req-844a2215-1038-4a30-8a79-40c1c3117bde tempest-FloatingIPsAssociationNegativeTestJSON-1301210771 tempest-FloatingIPsAssociationNegativeTestJSON-1301210771-project-member] [instance: 5d57303a-ae17-4b92-8f12-303677eb0d7e] Deallocating network for instance {{(pid=61923) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 556.953162] env[61923]: DEBUG nova.network.neutron [None req-844a2215-1038-4a30-8a79-40c1c3117bde tempest-FloatingIPsAssociationNegativeTestJSON-1301210771 tempest-FloatingIPsAssociationNegativeTestJSON-1301210771-project-member] [instance: 5d57303a-ae17-4b92-8f12-303677eb0d7e] deallocate_for_instance() {{(pid=61923) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 557.004266] env[61923]: DEBUG nova.network.neutron [None req-844a2215-1038-4a30-8a79-40c1c3117bde tempest-FloatingIPsAssociationNegativeTestJSON-1301210771 tempest-FloatingIPsAssociationNegativeTestJSON-1301210771-project-member] [instance: 5d57303a-ae17-4b92-8f12-303677eb0d7e] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 557.035658] env[61923]: DEBUG nova.compute.resource_tracker [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Instance 5d57303a-ae17-4b92-8f12-303677eb0d7e has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61923) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 557.035658] env[61923]: DEBUG nova.compute.resource_tracker [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Instance 59fc11e8-27be-4553-9de9-3eb5ea612ec4 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61923) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 557.035658] env[61923]: DEBUG nova.compute.resource_tracker [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Instance ca62f959-43b9-4838-8349-11b7d6a7f0a7 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61923) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 557.035658] env[61923]: DEBUG nova.compute.resource_tracker [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Instance f7cc960c-e06a-4c58-9367-ec8771fe09d5 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61923) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 557.035859] env[61923]: DEBUG nova.compute.resource_tracker [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Instance 06bda5ee-c7a8-47d1-9e0c-a5af67d40006 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61923) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 557.035859] env[61923]: DEBUG nova.compute.resource_tracker [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Instance 4cec7bce-0136-43ad-898f-086058a505f2 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61923) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 557.332543] env[61923]: INFO nova.compute.manager [None req-3b10f02e-4025-4174-94c7-8515a7e12175 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] [instance: f7cc960c-e06a-4c58-9367-ec8771fe09d5] Took 24.41 seconds to build instance. [ 557.512961] env[61923]: DEBUG nova.network.neutron [None req-844a2215-1038-4a30-8a79-40c1c3117bde tempest-FloatingIPsAssociationNegativeTestJSON-1301210771 tempest-FloatingIPsAssociationNegativeTestJSON-1301210771-project-member] [instance: 5d57303a-ae17-4b92-8f12-303677eb0d7e] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 557.539976] env[61923]: DEBUG nova.compute.resource_tracker [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Instance f3f872ce-cbbe-4407-b4e0-3341bb598825 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61923) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 557.836840] env[61923]: DEBUG oslo_concurrency.lockutils [None req-3b10f02e-4025-4174-94c7-8515a7e12175 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] Lock "f7cc960c-e06a-4c58-9367-ec8771fe09d5" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 25.932s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 557.945516] env[61923]: INFO nova.compute.manager [None req-1c44cc0c-5505-4b32-99eb-cbff6c4edcc0 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] [instance: f7cc960c-e06a-4c58-9367-ec8771fe09d5] Rebuilding instance [ 558.017170] env[61923]: INFO nova.compute.manager [None req-844a2215-1038-4a30-8a79-40c1c3117bde tempest-FloatingIPsAssociationNegativeTestJSON-1301210771 tempest-FloatingIPsAssociationNegativeTestJSON-1301210771-project-member] [instance: 5d57303a-ae17-4b92-8f12-303677eb0d7e] Took 1.07 seconds to deallocate network for instance. [ 558.020162] env[61923]: DEBUG nova.compute.manager [None req-1c44cc0c-5505-4b32-99eb-cbff6c4edcc0 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] [instance: f7cc960c-e06a-4c58-9367-ec8771fe09d5] Checking state {{(pid=61923) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 558.021257] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c58c44ef-5638-4106-8bd0-3a127afe1c32 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 558.043772] env[61923]: DEBUG nova.compute.resource_tracker [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Instance 6a2d7ede-e0cc-4a36-b1b2-0b36b269ecb5 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61923) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 558.341728] env[61923]: DEBUG nova.compute.manager [None req-543d2693-878a-4ade-aaa2-339016c76eca tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] [instance: 045c0b05-333e-4f95-94b0-8f51d87dca0a] Starting instance... {{(pid=61923) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 558.541802] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-1c44cc0c-5505-4b32-99eb-cbff6c4edcc0 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] [instance: f7cc960c-e06a-4c58-9367-ec8771fe09d5] Powering off the VM {{(pid=61923) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 558.544884] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-026110ae-ada7-4582-a857-e021c09316bd {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 558.546051] env[61923]: DEBUG nova.compute.resource_tracker [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Instance 9c3e2f94-91da-4255-b4d6-1c01e100ff48 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61923) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 558.551764] env[61923]: DEBUG oslo_vmware.api [None req-1c44cc0c-5505-4b32-99eb-cbff6c4edcc0 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] Waiting for the task: (returnval){ [ 558.551764] env[61923]: value = "task-1377435" [ 558.551764] env[61923]: _type = "Task" [ 558.551764] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 558.565256] env[61923]: DEBUG oslo_vmware.api [None req-1c44cc0c-5505-4b32-99eb-cbff6c4edcc0 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] Task: {'id': task-1377435, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 558.872228] env[61923]: DEBUG oslo_concurrency.lockutils [None req-543d2693-878a-4ade-aaa2-339016c76eca tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 559.052611] env[61923]: DEBUG nova.compute.resource_tracker [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Instance 93818ffd-eb85-48f0-aebd-9774467b0a8f has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61923) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 559.066922] env[61923]: DEBUG oslo_vmware.api [None req-1c44cc0c-5505-4b32-99eb-cbff6c4edcc0 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] Task: {'id': task-1377435, 'name': PowerOffVM_Task, 'duration_secs': 0.119825} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 559.068516] env[61923]: INFO nova.scheduler.client.report [None req-844a2215-1038-4a30-8a79-40c1c3117bde tempest-FloatingIPsAssociationNegativeTestJSON-1301210771 tempest-FloatingIPsAssociationNegativeTestJSON-1301210771-project-member] Deleted allocations for instance 5d57303a-ae17-4b92-8f12-303677eb0d7e [ 559.075807] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-1c44cc0c-5505-4b32-99eb-cbff6c4edcc0 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] [instance: f7cc960c-e06a-4c58-9367-ec8771fe09d5] Powered off the VM {{(pid=61923) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 559.076299] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-1c44cc0c-5505-4b32-99eb-cbff6c4edcc0 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] [instance: f7cc960c-e06a-4c58-9367-ec8771fe09d5] Destroying instance {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 559.077829] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7146290-c457-4996-b719-b01b6d0924af {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 559.087632] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-1c44cc0c-5505-4b32-99eb-cbff6c4edcc0 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] [instance: f7cc960c-e06a-4c58-9367-ec8771fe09d5] Unregistering the VM {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 559.087755] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-82754885-fc6b-466b-b106-bea2d3ab7b6d {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 559.113861] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-1c44cc0c-5505-4b32-99eb-cbff6c4edcc0 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] [instance: f7cc960c-e06a-4c58-9367-ec8771fe09d5] Unregistered the VM {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 559.114101] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-1c44cc0c-5505-4b32-99eb-cbff6c4edcc0 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] [instance: f7cc960c-e06a-4c58-9367-ec8771fe09d5] Deleting contents of the VM from datastore datastore1 {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 559.114291] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-1c44cc0c-5505-4b32-99eb-cbff6c4edcc0 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] Deleting the datastore file [datastore1] f7cc960c-e06a-4c58-9367-ec8771fe09d5 {{(pid=61923) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 559.114696] env[61923]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-54c892c2-da96-4112-b4aa-712fbd69c47e {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 559.121950] env[61923]: DEBUG oslo_vmware.api [None req-1c44cc0c-5505-4b32-99eb-cbff6c4edcc0 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] Waiting for the task: (returnval){ [ 559.121950] env[61923]: value = "task-1377437" [ 559.121950] env[61923]: _type = "Task" [ 559.121950] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 559.131040] env[61923]: DEBUG oslo_vmware.api [None req-1c44cc0c-5505-4b32-99eb-cbff6c4edcc0 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] Task: {'id': task-1377437, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 559.563430] env[61923]: DEBUG nova.compute.resource_tracker [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Instance 045c0b05-333e-4f95-94b0-8f51d87dca0a has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61923) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 559.582251] env[61923]: DEBUG oslo_concurrency.lockutils [None req-844a2215-1038-4a30-8a79-40c1c3117bde tempest-FloatingIPsAssociationNegativeTestJSON-1301210771 tempest-FloatingIPsAssociationNegativeTestJSON-1301210771-project-member] Lock "5d57303a-ae17-4b92-8f12-303677eb0d7e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 38.646s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 559.642732] env[61923]: DEBUG oslo_vmware.api [None req-1c44cc0c-5505-4b32-99eb-cbff6c4edcc0 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] Task: {'id': task-1377437, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.100813} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 559.643013] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-1c44cc0c-5505-4b32-99eb-cbff6c4edcc0 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] Deleted the datastore file {{(pid=61923) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 559.643207] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-1c44cc0c-5505-4b32-99eb-cbff6c4edcc0 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] [instance: f7cc960c-e06a-4c58-9367-ec8771fe09d5] Deleted contents of the VM from datastore datastore1 {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 559.643379] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-1c44cc0c-5505-4b32-99eb-cbff6c4edcc0 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] [instance: f7cc960c-e06a-4c58-9367-ec8771fe09d5] Instance destroyed {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 560.068859] env[61923]: DEBUG nova.compute.resource_tracker [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Instance af5feda7-204a-47f1-b7fa-afe44832b9a2 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61923) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 560.087547] env[61923]: DEBUG nova.compute.manager [None req-58d8b214-0b32-4993-bdd8-d652de7f9397 tempest-ServerActionsTestOtherA-1581037736 tempest-ServerActionsTestOtherA-1581037736-project-member] [instance: af5feda7-204a-47f1-b7fa-afe44832b9a2] Starting instance... {{(pid=61923) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 560.572350] env[61923]: DEBUG nova.compute.resource_tracker [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Instance 71d9eccc-3b90-4b0a-a59a-d6a8e08a9417 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61923) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 560.612224] env[61923]: DEBUG oslo_concurrency.lockutils [None req-58d8b214-0b32-4993-bdd8-d652de7f9397 tempest-ServerActionsTestOtherA-1581037736 tempest-ServerActionsTestOtherA-1581037736-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 560.683619] env[61923]: DEBUG nova.compute.manager [req-835b5227-9af1-4b05-b080-ef48053213fc req-55deac0c-8bdd-4517-8016-345b3a5f368f service nova] [instance: 4cec7bce-0136-43ad-898f-086058a505f2] Received event network-changed-fe27e2eb-c2cb-49f1-84e8-03cd81ad3229 {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 560.683619] env[61923]: DEBUG nova.compute.manager [req-835b5227-9af1-4b05-b080-ef48053213fc req-55deac0c-8bdd-4517-8016-345b3a5f368f service nova] [instance: 4cec7bce-0136-43ad-898f-086058a505f2] Refreshing instance network info cache due to event network-changed-fe27e2eb-c2cb-49f1-84e8-03cd81ad3229. {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 560.683619] env[61923]: DEBUG oslo_concurrency.lockutils [req-835b5227-9af1-4b05-b080-ef48053213fc req-55deac0c-8bdd-4517-8016-345b3a5f368f service nova] Acquiring lock "refresh_cache-4cec7bce-0136-43ad-898f-086058a505f2" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 560.683619] env[61923]: DEBUG oslo_concurrency.lockutils [req-835b5227-9af1-4b05-b080-ef48053213fc req-55deac0c-8bdd-4517-8016-345b3a5f368f service nova] Acquired lock "refresh_cache-4cec7bce-0136-43ad-898f-086058a505f2" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 560.684088] env[61923]: DEBUG nova.network.neutron [req-835b5227-9af1-4b05-b080-ef48053213fc req-55deac0c-8bdd-4517-8016-345b3a5f368f service nova] [instance: 4cec7bce-0136-43ad-898f-086058a505f2] Refreshing network info cache for port fe27e2eb-c2cb-49f1-84e8-03cd81ad3229 {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 560.701967] env[61923]: DEBUG nova.virt.hardware [None req-1c44cc0c-5505-4b32-99eb-cbff6c4edcc0 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-29T20:07:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-29T20:07:35Z,direct_url=,disk_format='vmdk',id=0f153f63-ae0a-45b1-b7f5-7f9b673c947f,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='4e7b5e3b3c3443c8bd5c70f8a15739f5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-29T20:07:36Z,virtual_size=,visibility=), allow threads: False {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 560.702664] env[61923]: DEBUG nova.virt.hardware [None req-1c44cc0c-5505-4b32-99eb-cbff6c4edcc0 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] Flavor limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 560.702664] env[61923]: DEBUG nova.virt.hardware [None req-1c44cc0c-5505-4b32-99eb-cbff6c4edcc0 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] Image limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 560.702780] env[61923]: DEBUG nova.virt.hardware [None req-1c44cc0c-5505-4b32-99eb-cbff6c4edcc0 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] Flavor pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 560.702985] env[61923]: DEBUG nova.virt.hardware [None req-1c44cc0c-5505-4b32-99eb-cbff6c4edcc0 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] Image pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 560.703244] env[61923]: DEBUG nova.virt.hardware [None req-1c44cc0c-5505-4b32-99eb-cbff6c4edcc0 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 560.703931] env[61923]: DEBUG nova.virt.hardware [None req-1c44cc0c-5505-4b32-99eb-cbff6c4edcc0 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 560.703931] env[61923]: DEBUG nova.virt.hardware [None req-1c44cc0c-5505-4b32-99eb-cbff6c4edcc0 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 560.704264] env[61923]: DEBUG nova.virt.hardware [None req-1c44cc0c-5505-4b32-99eb-cbff6c4edcc0 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] Got 1 possible topologies {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 560.704518] env[61923]: DEBUG nova.virt.hardware [None req-1c44cc0c-5505-4b32-99eb-cbff6c4edcc0 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 560.704891] env[61923]: DEBUG nova.virt.hardware [None req-1c44cc0c-5505-4b32-99eb-cbff6c4edcc0 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 560.707936] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6083bb7-295f-4c7f-bc07-efc310600fc4 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 560.717424] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74f6081d-7ac7-4b64-ad64-d2d0aa3fa8f1 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 560.733128] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-1c44cc0c-5505-4b32-99eb-cbff6c4edcc0 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] [instance: f7cc960c-e06a-4c58-9367-ec8771fe09d5] Instance VIF info [] {{(pid=61923) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 560.738772] env[61923]: DEBUG oslo.service.loopingcall [None req-1c44cc0c-5505-4b32-99eb-cbff6c4edcc0 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61923) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 560.739235] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f7cc960c-e06a-4c58-9367-ec8771fe09d5] Creating VM on the ESX host {{(pid=61923) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 560.739335] env[61923]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c663d516-c019-415b-8575-9e6d5b1b5648 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 560.757161] env[61923]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 560.757161] env[61923]: value = "task-1377438" [ 560.757161] env[61923]: _type = "Task" [ 560.757161] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 560.766191] env[61923]: DEBUG oslo_vmware.api [-] Task: {'id': task-1377438, 'name': CreateVM_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 560.901767] env[61923]: ERROR nova.compute.manager [None req-15b9e1af-fe57-407e-a058-c08239279239 tempest-ServersWithSpecificFlavorTestJSON-1806421899 tempest-ServersWithSpecificFlavorTestJSON-1806421899-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port fe27e2eb-c2cb-49f1-84e8-03cd81ad3229, please check neutron logs for more information. [ 560.901767] env[61923]: ERROR nova.compute.manager Traceback (most recent call last): [ 560.901767] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 560.901767] env[61923]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 560.901767] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 560.901767] env[61923]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 560.901767] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 560.901767] env[61923]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 560.901767] env[61923]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 560.901767] env[61923]: ERROR nova.compute.manager self.force_reraise() [ 560.901767] env[61923]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 560.901767] env[61923]: ERROR nova.compute.manager raise self.value [ 560.901767] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 560.901767] env[61923]: ERROR nova.compute.manager updated_port = self._update_port( [ 560.901767] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 560.901767] env[61923]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 560.902769] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 560.902769] env[61923]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 560.902769] env[61923]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port fe27e2eb-c2cb-49f1-84e8-03cd81ad3229, please check neutron logs for more information. [ 560.902769] env[61923]: ERROR nova.compute.manager [ 560.902769] env[61923]: Traceback (most recent call last): [ 560.902769] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 560.902769] env[61923]: listener.cb(fileno) [ 560.902769] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 560.902769] env[61923]: result = function(*args, **kwargs) [ 560.902769] env[61923]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 560.902769] env[61923]: return func(*args, **kwargs) [ 560.902769] env[61923]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 560.902769] env[61923]: raise e [ 560.902769] env[61923]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 560.902769] env[61923]: nwinfo = self.network_api.allocate_for_instance( [ 560.902769] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 560.902769] env[61923]: created_port_ids = self._update_ports_for_instance( [ 560.902769] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 560.902769] env[61923]: with excutils.save_and_reraise_exception(): [ 560.902769] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 560.902769] env[61923]: self.force_reraise() [ 560.902769] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 560.902769] env[61923]: raise self.value [ 560.902769] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 560.902769] env[61923]: updated_port = self._update_port( [ 560.902769] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 560.902769] env[61923]: _ensure_no_port_binding_failure(port) [ 560.902769] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 560.902769] env[61923]: raise exception.PortBindingFailed(port_id=port['id']) [ 560.909470] env[61923]: nova.exception.PortBindingFailed: Binding failed for port fe27e2eb-c2cb-49f1-84e8-03cd81ad3229, please check neutron logs for more information. [ 560.909470] env[61923]: Removing descriptor: 15 [ 560.909470] env[61923]: ERROR nova.compute.manager [None req-15b9e1af-fe57-407e-a058-c08239279239 tempest-ServersWithSpecificFlavorTestJSON-1806421899 tempest-ServersWithSpecificFlavorTestJSON-1806421899-project-member] [instance: 4cec7bce-0136-43ad-898f-086058a505f2] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port fe27e2eb-c2cb-49f1-84e8-03cd81ad3229, please check neutron logs for more information. [ 560.909470] env[61923]: ERROR nova.compute.manager [instance: 4cec7bce-0136-43ad-898f-086058a505f2] Traceback (most recent call last): [ 560.909470] env[61923]: ERROR nova.compute.manager [instance: 4cec7bce-0136-43ad-898f-086058a505f2] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 560.909470] env[61923]: ERROR nova.compute.manager [instance: 4cec7bce-0136-43ad-898f-086058a505f2] yield resources [ 560.909470] env[61923]: ERROR nova.compute.manager [instance: 4cec7bce-0136-43ad-898f-086058a505f2] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 560.909470] env[61923]: ERROR nova.compute.manager [instance: 4cec7bce-0136-43ad-898f-086058a505f2] self.driver.spawn(context, instance, image_meta, [ 560.909470] env[61923]: ERROR nova.compute.manager [instance: 4cec7bce-0136-43ad-898f-086058a505f2] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 560.909470] env[61923]: ERROR nova.compute.manager [instance: 4cec7bce-0136-43ad-898f-086058a505f2] self._vmops.spawn(context, instance, image_meta, injected_files, [ 560.909470] env[61923]: ERROR nova.compute.manager [instance: 4cec7bce-0136-43ad-898f-086058a505f2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 560.909470] env[61923]: ERROR nova.compute.manager [instance: 4cec7bce-0136-43ad-898f-086058a505f2] vm_ref = self.build_virtual_machine(instance, [ 560.909852] env[61923]: ERROR nova.compute.manager [instance: 4cec7bce-0136-43ad-898f-086058a505f2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 560.909852] env[61923]: ERROR nova.compute.manager [instance: 4cec7bce-0136-43ad-898f-086058a505f2] vif_infos = vmwarevif.get_vif_info(self._session, [ 560.909852] env[61923]: ERROR nova.compute.manager [instance: 4cec7bce-0136-43ad-898f-086058a505f2] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 560.909852] env[61923]: ERROR nova.compute.manager [instance: 4cec7bce-0136-43ad-898f-086058a505f2] for vif in network_info: [ 560.909852] env[61923]: ERROR nova.compute.manager [instance: 4cec7bce-0136-43ad-898f-086058a505f2] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 560.909852] env[61923]: ERROR nova.compute.manager [instance: 4cec7bce-0136-43ad-898f-086058a505f2] return self._sync_wrapper(fn, *args, **kwargs) [ 560.909852] env[61923]: ERROR nova.compute.manager [instance: 4cec7bce-0136-43ad-898f-086058a505f2] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 560.909852] env[61923]: ERROR nova.compute.manager [instance: 4cec7bce-0136-43ad-898f-086058a505f2] self.wait() [ 560.909852] env[61923]: ERROR nova.compute.manager [instance: 4cec7bce-0136-43ad-898f-086058a505f2] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 560.909852] env[61923]: ERROR nova.compute.manager [instance: 4cec7bce-0136-43ad-898f-086058a505f2] self[:] = self._gt.wait() [ 560.909852] env[61923]: ERROR nova.compute.manager [instance: 4cec7bce-0136-43ad-898f-086058a505f2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 560.909852] env[61923]: ERROR nova.compute.manager [instance: 4cec7bce-0136-43ad-898f-086058a505f2] return self._exit_event.wait() [ 560.909852] env[61923]: ERROR nova.compute.manager [instance: 4cec7bce-0136-43ad-898f-086058a505f2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 560.910370] env[61923]: ERROR nova.compute.manager [instance: 4cec7bce-0136-43ad-898f-086058a505f2] result = hub.switch() [ 560.910370] env[61923]: ERROR nova.compute.manager [instance: 4cec7bce-0136-43ad-898f-086058a505f2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 560.910370] env[61923]: ERROR nova.compute.manager [instance: 4cec7bce-0136-43ad-898f-086058a505f2] return self.greenlet.switch() [ 560.910370] env[61923]: ERROR nova.compute.manager [instance: 4cec7bce-0136-43ad-898f-086058a505f2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 560.910370] env[61923]: ERROR nova.compute.manager [instance: 4cec7bce-0136-43ad-898f-086058a505f2] result = function(*args, **kwargs) [ 560.910370] env[61923]: ERROR nova.compute.manager [instance: 4cec7bce-0136-43ad-898f-086058a505f2] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 560.910370] env[61923]: ERROR nova.compute.manager [instance: 4cec7bce-0136-43ad-898f-086058a505f2] return func(*args, **kwargs) [ 560.910370] env[61923]: ERROR nova.compute.manager [instance: 4cec7bce-0136-43ad-898f-086058a505f2] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 560.910370] env[61923]: ERROR nova.compute.manager [instance: 4cec7bce-0136-43ad-898f-086058a505f2] raise e [ 560.910370] env[61923]: ERROR nova.compute.manager [instance: 4cec7bce-0136-43ad-898f-086058a505f2] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 560.910370] env[61923]: ERROR nova.compute.manager [instance: 4cec7bce-0136-43ad-898f-086058a505f2] nwinfo = self.network_api.allocate_for_instance( [ 560.910370] env[61923]: ERROR nova.compute.manager [instance: 4cec7bce-0136-43ad-898f-086058a505f2] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 560.910370] env[61923]: ERROR nova.compute.manager [instance: 4cec7bce-0136-43ad-898f-086058a505f2] created_port_ids = self._update_ports_for_instance( [ 560.910737] env[61923]: ERROR nova.compute.manager [instance: 4cec7bce-0136-43ad-898f-086058a505f2] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 560.910737] env[61923]: ERROR nova.compute.manager [instance: 4cec7bce-0136-43ad-898f-086058a505f2] with excutils.save_and_reraise_exception(): [ 560.910737] env[61923]: ERROR nova.compute.manager [instance: 4cec7bce-0136-43ad-898f-086058a505f2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 560.910737] env[61923]: ERROR nova.compute.manager [instance: 4cec7bce-0136-43ad-898f-086058a505f2] self.force_reraise() [ 560.910737] env[61923]: ERROR nova.compute.manager [instance: 4cec7bce-0136-43ad-898f-086058a505f2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 560.910737] env[61923]: ERROR nova.compute.manager [instance: 4cec7bce-0136-43ad-898f-086058a505f2] raise self.value [ 560.910737] env[61923]: ERROR nova.compute.manager [instance: 4cec7bce-0136-43ad-898f-086058a505f2] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 560.910737] env[61923]: ERROR nova.compute.manager [instance: 4cec7bce-0136-43ad-898f-086058a505f2] updated_port = self._update_port( [ 560.910737] env[61923]: ERROR nova.compute.manager [instance: 4cec7bce-0136-43ad-898f-086058a505f2] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 560.910737] env[61923]: ERROR nova.compute.manager [instance: 4cec7bce-0136-43ad-898f-086058a505f2] _ensure_no_port_binding_failure(port) [ 560.910737] env[61923]: ERROR nova.compute.manager [instance: 4cec7bce-0136-43ad-898f-086058a505f2] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 560.910737] env[61923]: ERROR nova.compute.manager [instance: 4cec7bce-0136-43ad-898f-086058a505f2] raise exception.PortBindingFailed(port_id=port['id']) [ 560.911103] env[61923]: ERROR nova.compute.manager [instance: 4cec7bce-0136-43ad-898f-086058a505f2] nova.exception.PortBindingFailed: Binding failed for port fe27e2eb-c2cb-49f1-84e8-03cd81ad3229, please check neutron logs for more information. [ 560.911103] env[61923]: ERROR nova.compute.manager [instance: 4cec7bce-0136-43ad-898f-086058a505f2] [ 560.911103] env[61923]: INFO nova.compute.manager [None req-15b9e1af-fe57-407e-a058-c08239279239 tempest-ServersWithSpecificFlavorTestJSON-1806421899 tempest-ServersWithSpecificFlavorTestJSON-1806421899-project-member] [instance: 4cec7bce-0136-43ad-898f-086058a505f2] Terminating instance [ 560.911103] env[61923]: DEBUG oslo_concurrency.lockutils [None req-15b9e1af-fe57-407e-a058-c08239279239 tempest-ServersWithSpecificFlavorTestJSON-1806421899 tempest-ServersWithSpecificFlavorTestJSON-1806421899-project-member] Acquiring lock "refresh_cache-4cec7bce-0136-43ad-898f-086058a505f2" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 561.029923] env[61923]: DEBUG oslo_concurrency.lockutils [None req-76010439-9fb6-4296-9b5a-107842db76db tempest-ServerExternalEventsTest-1479796458 tempest-ServerExternalEventsTest-1479796458-project-member] Acquiring lock "c0f87aa5-447b-4920-8251-354d89a14fbb" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 561.030265] env[61923]: DEBUG oslo_concurrency.lockutils [None req-76010439-9fb6-4296-9b5a-107842db76db tempest-ServerExternalEventsTest-1479796458 tempest-ServerExternalEventsTest-1479796458-project-member] Lock "c0f87aa5-447b-4920-8251-354d89a14fbb" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 561.080577] env[61923]: DEBUG nova.compute.resource_tracker [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Instance cb998da7-8db0-4725-bed2-657d7df748b5 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61923) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 561.254282] env[61923]: DEBUG nova.network.neutron [req-835b5227-9af1-4b05-b080-ef48053213fc req-55deac0c-8bdd-4517-8016-345b3a5f368f service nova] [instance: 4cec7bce-0136-43ad-898f-086058a505f2] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 561.267947] env[61923]: DEBUG oslo_vmware.api [-] Task: {'id': task-1377438, 'name': CreateVM_Task, 'duration_secs': 0.303785} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 561.268595] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f7cc960c-e06a-4c58-9367-ec8771fe09d5] Created VM on the ESX host {{(pid=61923) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 561.268986] env[61923]: DEBUG oslo_vmware.service [None req-1c44cc0c-5505-4b32-99eb-cbff6c4edcc0 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bcbf175f-951b-49e8-8f1f-77280b0c62e8 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 561.276240] env[61923]: DEBUG oslo_concurrency.lockutils [None req-1c44cc0c-5505-4b32-99eb-cbff6c4edcc0 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 561.276438] env[61923]: DEBUG oslo_concurrency.lockutils [None req-1c44cc0c-5505-4b32-99eb-cbff6c4edcc0 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 561.276812] env[61923]: DEBUG oslo_concurrency.lockutils [None req-1c44cc0c-5505-4b32-99eb-cbff6c4edcc0 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 561.277085] env[61923]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-753e1b5d-5948-45d4-9af3-4b1161473729 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 561.283893] env[61923]: DEBUG oslo_vmware.api [None req-1c44cc0c-5505-4b32-99eb-cbff6c4edcc0 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] Waiting for the task: (returnval){ [ 561.283893] env[61923]: value = "session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]5276f770-dc38-4e24-b48c-d79e5ad8a107" [ 561.283893] env[61923]: _type = "Task" [ 561.283893] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 561.290229] env[61923]: DEBUG oslo_vmware.api [None req-1c44cc0c-5505-4b32-99eb-cbff6c4edcc0 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]5276f770-dc38-4e24-b48c-d79e5ad8a107, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 561.586787] env[61923]: DEBUG nova.compute.resource_tracker [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Instance 896dd689-4810-4f23-af0a-d2f557a0796b has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=61923) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 561.743213] env[61923]: DEBUG nova.network.neutron [req-835b5227-9af1-4b05-b080-ef48053213fc req-55deac0c-8bdd-4517-8016-345b3a5f368f service nova] [instance: 4cec7bce-0136-43ad-898f-086058a505f2] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 561.799572] env[61923]: DEBUG oslo_concurrency.lockutils [None req-1c44cc0c-5505-4b32-99eb-cbff6c4edcc0 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 561.799572] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-1c44cc0c-5505-4b32-99eb-cbff6c4edcc0 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] [instance: f7cc960c-e06a-4c58-9367-ec8771fe09d5] Processing image 0f153f63-ae0a-45b1-b7f5-7f9b673c947f {{(pid=61923) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 561.799572] env[61923]: DEBUG oslo_concurrency.lockutils [None req-1c44cc0c-5505-4b32-99eb-cbff6c4edcc0 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 561.799572] env[61923]: DEBUG oslo_concurrency.lockutils [None req-1c44cc0c-5505-4b32-99eb-cbff6c4edcc0 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 561.800179] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-1c44cc0c-5505-4b32-99eb-cbff6c4edcc0 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61923) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 561.800179] env[61923]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-6794820f-3262-44d0-8a34-1fd0274b8e09 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 561.810296] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-1c44cc0c-5505-4b32-99eb-cbff6c4edcc0 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61923) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 561.810481] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-1c44cc0c-5505-4b32-99eb-cbff6c4edcc0 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61923) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 561.816234] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f503bb29-7d38-4bbf-be3e-3cce24d5fa27 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 561.827243] env[61923]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b0a8ffd6-6527-4033-be90-dd2e39a58488 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 561.832366] env[61923]: DEBUG oslo_vmware.api [None req-1c44cc0c-5505-4b32-99eb-cbff6c4edcc0 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] Waiting for the task: (returnval){ [ 561.832366] env[61923]: value = "session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]527e6627-d5f7-dc3d-5473-a20b801c5080" [ 561.832366] env[61923]: _type = "Task" [ 561.832366] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 561.842028] env[61923]: DEBUG oslo_vmware.api [None req-1c44cc0c-5505-4b32-99eb-cbff6c4edcc0 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]527e6627-d5f7-dc3d-5473-a20b801c5080, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 562.092080] env[61923]: DEBUG nova.compute.resource_tracker [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Instance 119bb4a7-2d71-44d7-ae81-9238512937c8 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61923) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 562.248294] env[61923]: DEBUG oslo_concurrency.lockutils [req-835b5227-9af1-4b05-b080-ef48053213fc req-55deac0c-8bdd-4517-8016-345b3a5f368f service nova] Releasing lock "refresh_cache-4cec7bce-0136-43ad-898f-086058a505f2" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 562.248294] env[61923]: DEBUG oslo_concurrency.lockutils [None req-15b9e1af-fe57-407e-a058-c08239279239 tempest-ServersWithSpecificFlavorTestJSON-1806421899 tempest-ServersWithSpecificFlavorTestJSON-1806421899-project-member] Acquired lock "refresh_cache-4cec7bce-0136-43ad-898f-086058a505f2" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 562.248294] env[61923]: DEBUG nova.network.neutron [None req-15b9e1af-fe57-407e-a058-c08239279239 tempest-ServersWithSpecificFlavorTestJSON-1806421899 tempest-ServersWithSpecificFlavorTestJSON-1806421899-project-member] [instance: 4cec7bce-0136-43ad-898f-086058a505f2] Building network info cache for instance {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 562.345999] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-1c44cc0c-5505-4b32-99eb-cbff6c4edcc0 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] [instance: f7cc960c-e06a-4c58-9367-ec8771fe09d5] Preparing fetch location {{(pid=61923) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 562.346405] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-1c44cc0c-5505-4b32-99eb-cbff6c4edcc0 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] Creating directory with path [datastore2] vmware_temp/3bb8b4cb-269d-4553-9cb6-269c9a249735/0f153f63-ae0a-45b1-b7f5-7f9b673c947f {{(pid=61923) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 562.346924] env[61923]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ac5fe34c-49eb-4697-8cfa-920b8f0097a1 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 562.368397] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-1c44cc0c-5505-4b32-99eb-cbff6c4edcc0 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] Created directory with path [datastore2] vmware_temp/3bb8b4cb-269d-4553-9cb6-269c9a249735/0f153f63-ae0a-45b1-b7f5-7f9b673c947f {{(pid=61923) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 562.368617] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-1c44cc0c-5505-4b32-99eb-cbff6c4edcc0 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] [instance: f7cc960c-e06a-4c58-9367-ec8771fe09d5] Fetch image to [datastore2] vmware_temp/3bb8b4cb-269d-4553-9cb6-269c9a249735/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/tmp-sparse.vmdk {{(pid=61923) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 562.368795] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-1c44cc0c-5505-4b32-99eb-cbff6c4edcc0 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] [instance: f7cc960c-e06a-4c58-9367-ec8771fe09d5] Downloading image file data 0f153f63-ae0a-45b1-b7f5-7f9b673c947f to [datastore2] vmware_temp/3bb8b4cb-269d-4553-9cb6-269c9a249735/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/tmp-sparse.vmdk on the data store datastore2 {{(pid=61923) _fetch_image_as_file /opt/stack/nova/nova/virt/vmwareapi/vmops.py:399}} [ 562.369857] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c45e128e-76e2-4817-8fdd-3982eff306d1 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 562.377267] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7566199a-b440-4355-8f0a-a9e84d34ece5 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 562.386633] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c43be4f-c5b7-4dcb-849c-05a9f82b8658 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 562.417249] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1b6989e-9c24-4ffa-8270-bd8c07a34439 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 562.423052] env[61923]: DEBUG oslo_vmware.service [-] Invoking SessionManager.AcquireGenericServiceTicket with opID=oslo.vmware-f2136960-4475-448a-8518-b779be7eb2f3 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 562.451564] env[61923]: DEBUG nova.virt.vmwareapi.images [None req-1c44cc0c-5505-4b32-99eb-cbff6c4edcc0 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] [instance: f7cc960c-e06a-4c58-9367-ec8771fe09d5] Downloading image file data 0f153f63-ae0a-45b1-b7f5-7f9b673c947f to the data store datastore2 {{(pid=61923) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:245}} [ 562.518998] env[61923]: DEBUG oslo_vmware.rw_handles [None req-1c44cc0c-5505-4b32-99eb-cbff6c4edcc0 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] Creating HTTP connection to write to file with size = 21318656 and URL = https://esx7c1n3.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/3bb8b4cb-269d-4553-9cb6-269c9a249735/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore2. {{(pid=61923) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 562.593115] env[61923]: DEBUG nova.compute.resource_tracker [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Instance 0c1b8117-199f-4101-ab81-4ae7ef0d1251 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61923) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 562.820088] env[61923]: DEBUG nova.network.neutron [None req-15b9e1af-fe57-407e-a058-c08239279239 tempest-ServersWithSpecificFlavorTestJSON-1806421899 tempest-ServersWithSpecificFlavorTestJSON-1806421899-project-member] [instance: 4cec7bce-0136-43ad-898f-086058a505f2] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 563.002096] env[61923]: DEBUG nova.compute.manager [req-2360ca81-4b7f-4f2b-aa6d-e09f4663fbe5 req-772bf5f8-ac83-411c-b89e-88498326bc12 service nova] [instance: 4cec7bce-0136-43ad-898f-086058a505f2] Received event network-vif-deleted-fe27e2eb-c2cb-49f1-84e8-03cd81ad3229 {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 563.096805] env[61923]: DEBUG nova.compute.resource_tracker [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Instance 1b509288-2df3-4927-afcf-cb76918d3da3 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61923) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 563.115356] env[61923]: DEBUG nova.network.neutron [None req-15b9e1af-fe57-407e-a058-c08239279239 tempest-ServersWithSpecificFlavorTestJSON-1806421899 tempest-ServersWithSpecificFlavorTestJSON-1806421899-project-member] [instance: 4cec7bce-0136-43ad-898f-086058a505f2] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 563.266276] env[61923]: DEBUG oslo_vmware.rw_handles [None req-1c44cc0c-5505-4b32-99eb-cbff6c4edcc0 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] Completed reading data from the image iterator. {{(pid=61923) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 563.266276] env[61923]: DEBUG oslo_vmware.rw_handles [None req-1c44cc0c-5505-4b32-99eb-cbff6c4edcc0 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] Closing write handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/3bb8b4cb-269d-4553-9cb6-269c9a249735/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore2. {{(pid=61923) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:281}} [ 563.327958] env[61923]: DEBUG nova.virt.vmwareapi.images [None req-1c44cc0c-5505-4b32-99eb-cbff6c4edcc0 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] [instance: f7cc960c-e06a-4c58-9367-ec8771fe09d5] Downloaded image file data 0f153f63-ae0a-45b1-b7f5-7f9b673c947f to vmware_temp/3bb8b4cb-269d-4553-9cb6-269c9a249735/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/tmp-sparse.vmdk on the data store datastore2 {{(pid=61923) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:258}} [ 563.329971] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-1c44cc0c-5505-4b32-99eb-cbff6c4edcc0 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] [instance: f7cc960c-e06a-4c58-9367-ec8771fe09d5] Caching image {{(pid=61923) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 563.330239] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-1c44cc0c-5505-4b32-99eb-cbff6c4edcc0 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] Copying Virtual Disk [datastore2] vmware_temp/3bb8b4cb-269d-4553-9cb6-269c9a249735/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/tmp-sparse.vmdk to [datastore2] vmware_temp/3bb8b4cb-269d-4553-9cb6-269c9a249735/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk {{(pid=61923) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 563.333581] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-49c1216a-727b-408d-befa-45d54f3a5d20 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 563.350222] env[61923]: DEBUG oslo_vmware.api [None req-1c44cc0c-5505-4b32-99eb-cbff6c4edcc0 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] Waiting for the task: (returnval){ [ 563.350222] env[61923]: value = "task-1377443" [ 563.350222] env[61923]: _type = "Task" [ 563.350222] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 563.360322] env[61923]: DEBUG oslo_vmware.api [None req-1c44cc0c-5505-4b32-99eb-cbff6c4edcc0 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] Task: {'id': task-1377443, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 563.602508] env[61923]: DEBUG nova.compute.resource_tracker [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Instance a4b54a7a-3f3d-49ae-ba43-5a10c262d4e5 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61923) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 563.619724] env[61923]: DEBUG oslo_concurrency.lockutils [None req-15b9e1af-fe57-407e-a058-c08239279239 tempest-ServersWithSpecificFlavorTestJSON-1806421899 tempest-ServersWithSpecificFlavorTestJSON-1806421899-project-member] Releasing lock "refresh_cache-4cec7bce-0136-43ad-898f-086058a505f2" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 563.620270] env[61923]: DEBUG nova.compute.manager [None req-15b9e1af-fe57-407e-a058-c08239279239 tempest-ServersWithSpecificFlavorTestJSON-1806421899 tempest-ServersWithSpecificFlavorTestJSON-1806421899-project-member] [instance: 4cec7bce-0136-43ad-898f-086058a505f2] Start destroying the instance on the hypervisor. {{(pid=61923) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 563.620434] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-15b9e1af-fe57-407e-a058-c08239279239 tempest-ServersWithSpecificFlavorTestJSON-1806421899 tempest-ServersWithSpecificFlavorTestJSON-1806421899-project-member] [instance: 4cec7bce-0136-43ad-898f-086058a505f2] Destroying instance {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 563.620851] env[61923]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-34c7bf91-d3dc-4954-b6de-23170d3bce53 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 563.635659] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3785036-8261-4130-a646-adfe27a52d9c {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 563.663283] env[61923]: WARNING nova.virt.vmwareapi.vmops [None req-15b9e1af-fe57-407e-a058-c08239279239 tempest-ServersWithSpecificFlavorTestJSON-1806421899 tempest-ServersWithSpecificFlavorTestJSON-1806421899-project-member] [instance: 4cec7bce-0136-43ad-898f-086058a505f2] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 4cec7bce-0136-43ad-898f-086058a505f2 could not be found. [ 563.663530] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-15b9e1af-fe57-407e-a058-c08239279239 tempest-ServersWithSpecificFlavorTestJSON-1806421899 tempest-ServersWithSpecificFlavorTestJSON-1806421899-project-member] [instance: 4cec7bce-0136-43ad-898f-086058a505f2] Instance destroyed {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 563.663713] env[61923]: INFO nova.compute.manager [None req-15b9e1af-fe57-407e-a058-c08239279239 tempest-ServersWithSpecificFlavorTestJSON-1806421899 tempest-ServersWithSpecificFlavorTestJSON-1806421899-project-member] [instance: 4cec7bce-0136-43ad-898f-086058a505f2] Took 0.04 seconds to destroy the instance on the hypervisor. [ 563.663960] env[61923]: DEBUG oslo.service.loopingcall [None req-15b9e1af-fe57-407e-a058-c08239279239 tempest-ServersWithSpecificFlavorTestJSON-1806421899 tempest-ServersWithSpecificFlavorTestJSON-1806421899-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61923) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 563.664273] env[61923]: DEBUG nova.compute.manager [-] [instance: 4cec7bce-0136-43ad-898f-086058a505f2] Deallocating network for instance {{(pid=61923) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 563.664377] env[61923]: DEBUG nova.network.neutron [-] [instance: 4cec7bce-0136-43ad-898f-086058a505f2] deallocate_for_instance() {{(pid=61923) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 563.730854] env[61923]: DEBUG nova.network.neutron [-] [instance: 4cec7bce-0136-43ad-898f-086058a505f2] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 563.860012] env[61923]: DEBUG oslo_vmware.api [None req-1c44cc0c-5505-4b32-99eb-cbff6c4edcc0 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] Task: {'id': task-1377443, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 564.106543] env[61923]: DEBUG nova.compute.resource_tracker [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Instance dc8cc22d-aba9-4dba-b456-1b323013d234 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61923) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 564.234047] env[61923]: DEBUG nova.network.neutron [-] [instance: 4cec7bce-0136-43ad-898f-086058a505f2] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 564.365788] env[61923]: DEBUG oslo_vmware.api [None req-1c44cc0c-5505-4b32-99eb-cbff6c4edcc0 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] Task: {'id': task-1377443, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.744107} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 564.366143] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-1c44cc0c-5505-4b32-99eb-cbff6c4edcc0 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] Copied Virtual Disk [datastore2] vmware_temp/3bb8b4cb-269d-4553-9cb6-269c9a249735/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/tmp-sparse.vmdk to [datastore2] vmware_temp/3bb8b4cb-269d-4553-9cb6-269c9a249735/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk {{(pid=61923) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 564.366358] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-1c44cc0c-5505-4b32-99eb-cbff6c4edcc0 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] Deleting the datastore file [datastore2] vmware_temp/3bb8b4cb-269d-4553-9cb6-269c9a249735/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/tmp-sparse.vmdk {{(pid=61923) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 564.366631] env[61923]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-3388c880-12df-49dc-86a9-436caf4578c9 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 564.377844] env[61923]: DEBUG oslo_vmware.api [None req-1c44cc0c-5505-4b32-99eb-cbff6c4edcc0 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] Waiting for the task: (returnval){ [ 564.377844] env[61923]: value = "task-1377445" [ 564.377844] env[61923]: _type = "Task" [ 564.377844] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 564.391316] env[61923]: DEBUG oslo_vmware.api [None req-1c44cc0c-5505-4b32-99eb-cbff6c4edcc0 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] Task: {'id': task-1377445, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 564.609752] env[61923]: DEBUG nova.compute.resource_tracker [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Instance 03100310-86ce-4739-8636-cd2aa1b542ff has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61923) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 564.615578] env[61923]: DEBUG nova.compute.resource_tracker [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Total usable vcpus: 48, total allocated vcpus: 5 {{(pid=61923) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 564.615578] env[61923]: DEBUG nova.compute.resource_tracker [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1472MB phys_disk=200GB used_disk=5GB total_vcpus=48 used_vcpus=5 pci_stats=[] {{(pid=61923) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 564.740578] env[61923]: INFO nova.compute.manager [-] [instance: 4cec7bce-0136-43ad-898f-086058a505f2] Took 1.08 seconds to deallocate network for instance. [ 564.745657] env[61923]: DEBUG nova.compute.claims [None req-15b9e1af-fe57-407e-a058-c08239279239 tempest-ServersWithSpecificFlavorTestJSON-1806421899 tempest-ServersWithSpecificFlavorTestJSON-1806421899-project-member] [instance: 4cec7bce-0136-43ad-898f-086058a505f2] Aborting claim: {{(pid=61923) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 564.745840] env[61923]: DEBUG oslo_concurrency.lockutils [None req-15b9e1af-fe57-407e-a058-c08239279239 tempest-ServersWithSpecificFlavorTestJSON-1806421899 tempest-ServersWithSpecificFlavorTestJSON-1806421899-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 564.896949] env[61923]: DEBUG oslo_vmware.api [None req-1c44cc0c-5505-4b32-99eb-cbff6c4edcc0 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] Task: {'id': task-1377445, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.082746} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 564.897630] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-1c44cc0c-5505-4b32-99eb-cbff6c4edcc0 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] Deleted the datastore file {{(pid=61923) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 564.898591] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-1c44cc0c-5505-4b32-99eb-cbff6c4edcc0 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] Moving file from [datastore2] vmware_temp/3bb8b4cb-269d-4553-9cb6-269c9a249735/0f153f63-ae0a-45b1-b7f5-7f9b673c947f to [datastore2] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f. {{(pid=61923) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:334}} [ 564.898591] env[61923]: DEBUG oslo_vmware.service [-] Invoking FileManager.MoveDatastoreFile_Task with opID=oslo.vmware-395bb0d3-abcd-4a05-b2d9-04537745b82e {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 564.911563] env[61923]: DEBUG oslo_vmware.api [None req-1c44cc0c-5505-4b32-99eb-cbff6c4edcc0 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] Waiting for the task: (returnval){ [ 564.911563] env[61923]: value = "task-1377446" [ 564.911563] env[61923]: _type = "Task" [ 564.911563] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 564.920883] env[61923]: DEBUG oslo_vmware.api [None req-1c44cc0c-5505-4b32-99eb-cbff6c4edcc0 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] Task: {'id': task-1377446, 'name': MoveDatastoreFile_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 565.009672] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4191bd5-db8e-43e5-b860-ef9eee0fd23e {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 565.017890] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a60bdb7a-ffe8-4fcf-83a5-33bbc654d4e6 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 565.050671] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b50d373-131d-498b-9f39-d0af55c0bb3f {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 565.058567] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ccb2a719-77d3-47a6-a8e9-b743825c7a43 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 565.072301] env[61923]: DEBUG nova.compute.provider_tree [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 565.428028] env[61923]: DEBUG oslo_vmware.api [None req-1c44cc0c-5505-4b32-99eb-cbff6c4edcc0 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] Task: {'id': task-1377446, 'name': MoveDatastoreFile_Task, 'duration_secs': 0.030882} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 565.428607] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-1c44cc0c-5505-4b32-99eb-cbff6c4edcc0 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] File moved {{(pid=61923) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:346}} [ 565.429294] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-1c44cc0c-5505-4b32-99eb-cbff6c4edcc0 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] [instance: f7cc960c-e06a-4c58-9367-ec8771fe09d5] Cleaning up location [datastore2] vmware_temp/3bb8b4cb-269d-4553-9cb6-269c9a249735 {{(pid=61923) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 565.429591] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-1c44cc0c-5505-4b32-99eb-cbff6c4edcc0 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] Deleting the datastore file [datastore2] vmware_temp/3bb8b4cb-269d-4553-9cb6-269c9a249735 {{(pid=61923) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 565.429781] env[61923]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-56d8b19a-592f-401d-a69c-965149f65c4e {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 565.437170] env[61923]: DEBUG oslo_vmware.api [None req-1c44cc0c-5505-4b32-99eb-cbff6c4edcc0 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] Waiting for the task: (returnval){ [ 565.437170] env[61923]: value = "task-1377447" [ 565.437170] env[61923]: _type = "Task" [ 565.437170] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 565.447919] env[61923]: DEBUG oslo_vmware.api [None req-1c44cc0c-5505-4b32-99eb-cbff6c4edcc0 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] Task: {'id': task-1377447, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 565.575294] env[61923]: DEBUG nova.scheduler.client.report [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 565.948475] env[61923]: DEBUG oslo_vmware.api [None req-1c44cc0c-5505-4b32-99eb-cbff6c4edcc0 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] Task: {'id': task-1377447, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.024726} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 565.948746] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-1c44cc0c-5505-4b32-99eb-cbff6c4edcc0 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] Deleted the datastore file {{(pid=61923) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 565.950185] env[61923]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8246f2a6-a05e-482c-bc1f-087026db1651 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 565.955790] env[61923]: DEBUG oslo_vmware.api [None req-1c44cc0c-5505-4b32-99eb-cbff6c4edcc0 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] Waiting for the task: (returnval){ [ 565.955790] env[61923]: value = "session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52b7c378-91d9-a7f2-2c83-fad71a00d41e" [ 565.955790] env[61923]: _type = "Task" [ 565.955790] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 565.969697] env[61923]: DEBUG oslo_vmware.api [None req-1c44cc0c-5505-4b32-99eb-cbff6c4edcc0 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52b7c378-91d9-a7f2-2c83-fad71a00d41e, 'name': SearchDatastore_Task, 'duration_secs': 0.008849} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 565.970010] env[61923]: DEBUG oslo_concurrency.lockutils [None req-1c44cc0c-5505-4b32-99eb-cbff6c4edcc0 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 565.970280] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-1c44cc0c-5505-4b32-99eb-cbff6c4edcc0 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk to [datastore2] f7cc960c-e06a-4c58-9367-ec8771fe09d5/f7cc960c-e06a-4c58-9367-ec8771fe09d5.vmdk {{(pid=61923) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 565.970537] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f0e5a34d-561b-4686-876d-bd43a8214aa0 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 565.983260] env[61923]: DEBUG oslo_vmware.api [None req-1c44cc0c-5505-4b32-99eb-cbff6c4edcc0 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] Waiting for the task: (returnval){ [ 565.983260] env[61923]: value = "task-1377449" [ 565.983260] env[61923]: _type = "Task" [ 565.983260] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 565.996397] env[61923]: DEBUG oslo_vmware.api [None req-1c44cc0c-5505-4b32-99eb-cbff6c4edcc0 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] Task: {'id': task-1377449, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 566.087401] env[61923]: DEBUG nova.compute.resource_tracker [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61923) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 566.087401] env[61923]: DEBUG oslo_concurrency.lockutils [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 10.601s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 566.087401] env[61923]: DEBUG oslo_concurrency.lockutils [None req-f30f4038-8fd0-4bbd-8b35-7fb8e0dea3a4 tempest-SecurityGroupsTestJSON-509490211 tempest-SecurityGroupsTestJSON-509490211-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 24.362s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 566.500607] env[61923]: DEBUG oslo_vmware.api [None req-1c44cc0c-5505-4b32-99eb-cbff6c4edcc0 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] Task: {'id': task-1377449, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 567.004776] env[61923]: DEBUG oslo_vmware.api [None req-1c44cc0c-5505-4b32-99eb-cbff6c4edcc0 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] Task: {'id': task-1377449, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.526211} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 567.005140] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-1c44cc0c-5505-4b32-99eb-cbff6c4edcc0 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk to [datastore2] f7cc960c-e06a-4c58-9367-ec8771fe09d5/f7cc960c-e06a-4c58-9367-ec8771fe09d5.vmdk {{(pid=61923) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 567.005140] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-1c44cc0c-5505-4b32-99eb-cbff6c4edcc0 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] [instance: f7cc960c-e06a-4c58-9367-ec8771fe09d5] Extending root virtual disk to 1048576 {{(pid=61923) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 567.008543] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-d59bcc79-683f-4421-be52-a2c545ddefcd {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 567.021256] env[61923]: DEBUG oslo_vmware.api [None req-1c44cc0c-5505-4b32-99eb-cbff6c4edcc0 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] Waiting for the task: (returnval){ [ 567.021256] env[61923]: value = "task-1377450" [ 567.021256] env[61923]: _type = "Task" [ 567.021256] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 567.029663] env[61923]: DEBUG oslo_vmware.api [None req-1c44cc0c-5505-4b32-99eb-cbff6c4edcc0 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] Task: {'id': task-1377450, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 567.116052] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-baf00195-d923-4164-b81f-730012cfbbfd {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 567.123394] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1620affe-7c6f-4b1d-9b91-6ab4e62d1a30 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 567.160284] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ce1df7d-6fb6-4aee-ae49-807c3956bc6b {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 567.168907] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bbabcfc3-30bd-47f6-b391-f640d385a753 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 567.186863] env[61923]: DEBUG nova.compute.provider_tree [None req-f30f4038-8fd0-4bbd-8b35-7fb8e0dea3a4 tempest-SecurityGroupsTestJSON-509490211 tempest-SecurityGroupsTestJSON-509490211-project-member] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 567.531963] env[61923]: DEBUG oslo_vmware.api [None req-1c44cc0c-5505-4b32-99eb-cbff6c4edcc0 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] Task: {'id': task-1377450, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.067321} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 567.531963] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-1c44cc0c-5505-4b32-99eb-cbff6c4edcc0 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] [instance: f7cc960c-e06a-4c58-9367-ec8771fe09d5] Extended root virtual disk {{(pid=61923) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 567.535753] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c679eb6f-7704-4acc-8a34-a4a37b4b2783 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 567.563908] env[61923]: DEBUG nova.virt.vmwareapi.volumeops [None req-1c44cc0c-5505-4b32-99eb-cbff6c4edcc0 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] [instance: f7cc960c-e06a-4c58-9367-ec8771fe09d5] Reconfiguring VM instance instance-0000000a to attach disk [datastore2] f7cc960c-e06a-4c58-9367-ec8771fe09d5/f7cc960c-e06a-4c58-9367-ec8771fe09d5.vmdk or device None with type sparse {{(pid=61923) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 567.567014] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-604572b4-54ec-4a84-96bd-3b87e4e9e792 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 567.590134] env[61923]: DEBUG oslo_vmware.api [None req-1c44cc0c-5505-4b32-99eb-cbff6c4edcc0 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] Waiting for the task: (returnval){ [ 567.590134] env[61923]: value = "task-1377451" [ 567.590134] env[61923]: _type = "Task" [ 567.590134] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 567.603269] env[61923]: DEBUG oslo_vmware.api [None req-1c44cc0c-5505-4b32-99eb-cbff6c4edcc0 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] Task: {'id': task-1377451, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 567.684568] env[61923]: DEBUG nova.scheduler.client.report [None req-f30f4038-8fd0-4bbd-8b35-7fb8e0dea3a4 tempest-SecurityGroupsTestJSON-509490211 tempest-SecurityGroupsTestJSON-509490211-project-member] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 568.103727] env[61923]: DEBUG oslo_vmware.api [None req-1c44cc0c-5505-4b32-99eb-cbff6c4edcc0 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] Task: {'id': task-1377451, 'name': ReconfigVM_Task, 'duration_secs': 0.318221} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 568.103727] env[61923]: DEBUG nova.virt.vmwareapi.volumeops [None req-1c44cc0c-5505-4b32-99eb-cbff6c4edcc0 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] [instance: f7cc960c-e06a-4c58-9367-ec8771fe09d5] Reconfigured VM instance instance-0000000a to attach disk [datastore2] f7cc960c-e06a-4c58-9367-ec8771fe09d5/f7cc960c-e06a-4c58-9367-ec8771fe09d5.vmdk or device None with type sparse {{(pid=61923) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 568.103727] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-63c0177c-80cd-4ce8-bfd6-e80237a9f30f {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 568.115253] env[61923]: DEBUG oslo_vmware.api [None req-1c44cc0c-5505-4b32-99eb-cbff6c4edcc0 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] Waiting for the task: (returnval){ [ 568.115253] env[61923]: value = "task-1377453" [ 568.115253] env[61923]: _type = "Task" [ 568.115253] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 568.132170] env[61923]: DEBUG oslo_vmware.api [None req-1c44cc0c-5505-4b32-99eb-cbff6c4edcc0 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] Task: {'id': task-1377453, 'name': Rename_Task} progress is 14%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 568.193629] env[61923]: DEBUG oslo_concurrency.lockutils [None req-f30f4038-8fd0-4bbd-8b35-7fb8e0dea3a4 tempest-SecurityGroupsTestJSON-509490211 tempest-SecurityGroupsTestJSON-509490211-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.105s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 568.193629] env[61923]: ERROR nova.compute.manager [None req-f30f4038-8fd0-4bbd-8b35-7fb8e0dea3a4 tempest-SecurityGroupsTestJSON-509490211 tempest-SecurityGroupsTestJSON-509490211-project-member] [instance: 59fc11e8-27be-4553-9de9-3eb5ea612ec4] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 238265b6-446c-493c-b510-5f51005c1787, please check neutron logs for more information. [ 568.193629] env[61923]: ERROR nova.compute.manager [instance: 59fc11e8-27be-4553-9de9-3eb5ea612ec4] Traceback (most recent call last): [ 568.193629] env[61923]: ERROR nova.compute.manager [instance: 59fc11e8-27be-4553-9de9-3eb5ea612ec4] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 568.193629] env[61923]: ERROR nova.compute.manager [instance: 59fc11e8-27be-4553-9de9-3eb5ea612ec4] self.driver.spawn(context, instance, image_meta, [ 568.193629] env[61923]: ERROR nova.compute.manager [instance: 59fc11e8-27be-4553-9de9-3eb5ea612ec4] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 568.193629] env[61923]: ERROR nova.compute.manager [instance: 59fc11e8-27be-4553-9de9-3eb5ea612ec4] self._vmops.spawn(context, instance, image_meta, injected_files, [ 568.193629] env[61923]: ERROR nova.compute.manager [instance: 59fc11e8-27be-4553-9de9-3eb5ea612ec4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 568.193629] env[61923]: ERROR nova.compute.manager [instance: 59fc11e8-27be-4553-9de9-3eb5ea612ec4] vm_ref = self.build_virtual_machine(instance, [ 568.194039] env[61923]: ERROR nova.compute.manager [instance: 59fc11e8-27be-4553-9de9-3eb5ea612ec4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 568.194039] env[61923]: ERROR nova.compute.manager [instance: 59fc11e8-27be-4553-9de9-3eb5ea612ec4] vif_infos = vmwarevif.get_vif_info(self._session, [ 568.194039] env[61923]: ERROR nova.compute.manager [instance: 59fc11e8-27be-4553-9de9-3eb5ea612ec4] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 568.194039] env[61923]: ERROR nova.compute.manager [instance: 59fc11e8-27be-4553-9de9-3eb5ea612ec4] for vif in network_info: [ 568.194039] env[61923]: ERROR nova.compute.manager [instance: 59fc11e8-27be-4553-9de9-3eb5ea612ec4] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 568.194039] env[61923]: ERROR nova.compute.manager [instance: 59fc11e8-27be-4553-9de9-3eb5ea612ec4] return self._sync_wrapper(fn, *args, **kwargs) [ 568.194039] env[61923]: ERROR nova.compute.manager [instance: 59fc11e8-27be-4553-9de9-3eb5ea612ec4] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 568.194039] env[61923]: ERROR nova.compute.manager [instance: 59fc11e8-27be-4553-9de9-3eb5ea612ec4] self.wait() [ 568.194039] env[61923]: ERROR nova.compute.manager [instance: 59fc11e8-27be-4553-9de9-3eb5ea612ec4] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 568.194039] env[61923]: ERROR nova.compute.manager [instance: 59fc11e8-27be-4553-9de9-3eb5ea612ec4] self[:] = self._gt.wait() [ 568.194039] env[61923]: ERROR nova.compute.manager [instance: 59fc11e8-27be-4553-9de9-3eb5ea612ec4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 568.194039] env[61923]: ERROR nova.compute.manager [instance: 59fc11e8-27be-4553-9de9-3eb5ea612ec4] return self._exit_event.wait() [ 568.194039] env[61923]: ERROR nova.compute.manager [instance: 59fc11e8-27be-4553-9de9-3eb5ea612ec4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 568.194705] env[61923]: ERROR nova.compute.manager [instance: 59fc11e8-27be-4553-9de9-3eb5ea612ec4] result = hub.switch() [ 568.194705] env[61923]: ERROR nova.compute.manager [instance: 59fc11e8-27be-4553-9de9-3eb5ea612ec4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 568.194705] env[61923]: ERROR nova.compute.manager [instance: 59fc11e8-27be-4553-9de9-3eb5ea612ec4] return self.greenlet.switch() [ 568.194705] env[61923]: ERROR nova.compute.manager [instance: 59fc11e8-27be-4553-9de9-3eb5ea612ec4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 568.194705] env[61923]: ERROR nova.compute.manager [instance: 59fc11e8-27be-4553-9de9-3eb5ea612ec4] result = function(*args, **kwargs) [ 568.194705] env[61923]: ERROR nova.compute.manager [instance: 59fc11e8-27be-4553-9de9-3eb5ea612ec4] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 568.194705] env[61923]: ERROR nova.compute.manager [instance: 59fc11e8-27be-4553-9de9-3eb5ea612ec4] return func(*args, **kwargs) [ 568.194705] env[61923]: ERROR nova.compute.manager [instance: 59fc11e8-27be-4553-9de9-3eb5ea612ec4] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 568.194705] env[61923]: ERROR nova.compute.manager [instance: 59fc11e8-27be-4553-9de9-3eb5ea612ec4] raise e [ 568.194705] env[61923]: ERROR nova.compute.manager [instance: 59fc11e8-27be-4553-9de9-3eb5ea612ec4] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 568.194705] env[61923]: ERROR nova.compute.manager [instance: 59fc11e8-27be-4553-9de9-3eb5ea612ec4] nwinfo = self.network_api.allocate_for_instance( [ 568.194705] env[61923]: ERROR nova.compute.manager [instance: 59fc11e8-27be-4553-9de9-3eb5ea612ec4] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 568.194705] env[61923]: ERROR nova.compute.manager [instance: 59fc11e8-27be-4553-9de9-3eb5ea612ec4] created_port_ids = self._update_ports_for_instance( [ 568.195209] env[61923]: ERROR nova.compute.manager [instance: 59fc11e8-27be-4553-9de9-3eb5ea612ec4] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 568.195209] env[61923]: ERROR nova.compute.manager [instance: 59fc11e8-27be-4553-9de9-3eb5ea612ec4] with excutils.save_and_reraise_exception(): [ 568.195209] env[61923]: ERROR nova.compute.manager [instance: 59fc11e8-27be-4553-9de9-3eb5ea612ec4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 568.195209] env[61923]: ERROR nova.compute.manager [instance: 59fc11e8-27be-4553-9de9-3eb5ea612ec4] self.force_reraise() [ 568.195209] env[61923]: ERROR nova.compute.manager [instance: 59fc11e8-27be-4553-9de9-3eb5ea612ec4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 568.195209] env[61923]: ERROR nova.compute.manager [instance: 59fc11e8-27be-4553-9de9-3eb5ea612ec4] raise self.value [ 568.195209] env[61923]: ERROR nova.compute.manager [instance: 59fc11e8-27be-4553-9de9-3eb5ea612ec4] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 568.195209] env[61923]: ERROR nova.compute.manager [instance: 59fc11e8-27be-4553-9de9-3eb5ea612ec4] updated_port = self._update_port( [ 568.195209] env[61923]: ERROR nova.compute.manager [instance: 59fc11e8-27be-4553-9de9-3eb5ea612ec4] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 568.195209] env[61923]: ERROR nova.compute.manager [instance: 59fc11e8-27be-4553-9de9-3eb5ea612ec4] _ensure_no_port_binding_failure(port) [ 568.195209] env[61923]: ERROR nova.compute.manager [instance: 59fc11e8-27be-4553-9de9-3eb5ea612ec4] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 568.195209] env[61923]: ERROR nova.compute.manager [instance: 59fc11e8-27be-4553-9de9-3eb5ea612ec4] raise exception.PortBindingFailed(port_id=port['id']) [ 568.195631] env[61923]: ERROR nova.compute.manager [instance: 59fc11e8-27be-4553-9de9-3eb5ea612ec4] nova.exception.PortBindingFailed: Binding failed for port 238265b6-446c-493c-b510-5f51005c1787, please check neutron logs for more information. [ 568.195631] env[61923]: ERROR nova.compute.manager [instance: 59fc11e8-27be-4553-9de9-3eb5ea612ec4] [ 568.195631] env[61923]: DEBUG nova.compute.utils [None req-f30f4038-8fd0-4bbd-8b35-7fb8e0dea3a4 tempest-SecurityGroupsTestJSON-509490211 tempest-SecurityGroupsTestJSON-509490211-project-member] [instance: 59fc11e8-27be-4553-9de9-3eb5ea612ec4] Binding failed for port 238265b6-446c-493c-b510-5f51005c1787, please check neutron logs for more information. {{(pid=61923) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 568.195631] env[61923]: DEBUG oslo_concurrency.lockutils [None req-36da9865-5d98-4c2a-a964-964bee74523c tempest-ListServerFiltersTestJSON-483030130 tempest-ListServerFiltersTestJSON-483030130-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 25.632s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 568.197207] env[61923]: INFO nova.compute.claims [None req-36da9865-5d98-4c2a-a964-964bee74523c tempest-ListServerFiltersTestJSON-483030130 tempest-ListServerFiltersTestJSON-483030130-project-member] [instance: f3f872ce-cbbe-4407-b4e0-3341bb598825] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 568.200378] env[61923]: DEBUG nova.compute.manager [None req-f30f4038-8fd0-4bbd-8b35-7fb8e0dea3a4 tempest-SecurityGroupsTestJSON-509490211 tempest-SecurityGroupsTestJSON-509490211-project-member] [instance: 59fc11e8-27be-4553-9de9-3eb5ea612ec4] Build of instance 59fc11e8-27be-4553-9de9-3eb5ea612ec4 was re-scheduled: Binding failed for port 238265b6-446c-493c-b510-5f51005c1787, please check neutron logs for more information. {{(pid=61923) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 568.200998] env[61923]: DEBUG nova.compute.manager [None req-f30f4038-8fd0-4bbd-8b35-7fb8e0dea3a4 tempest-SecurityGroupsTestJSON-509490211 tempest-SecurityGroupsTestJSON-509490211-project-member] [instance: 59fc11e8-27be-4553-9de9-3eb5ea612ec4] Unplugging VIFs for instance {{(pid=61923) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 568.201249] env[61923]: DEBUG oslo_concurrency.lockutils [None req-f30f4038-8fd0-4bbd-8b35-7fb8e0dea3a4 tempest-SecurityGroupsTestJSON-509490211 tempest-SecurityGroupsTestJSON-509490211-project-member] Acquiring lock "refresh_cache-59fc11e8-27be-4553-9de9-3eb5ea612ec4" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 568.205159] env[61923]: DEBUG oslo_concurrency.lockutils [None req-f30f4038-8fd0-4bbd-8b35-7fb8e0dea3a4 tempest-SecurityGroupsTestJSON-509490211 tempest-SecurityGroupsTestJSON-509490211-project-member] Acquired lock "refresh_cache-59fc11e8-27be-4553-9de9-3eb5ea612ec4" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 568.205412] env[61923]: DEBUG nova.network.neutron [None req-f30f4038-8fd0-4bbd-8b35-7fb8e0dea3a4 tempest-SecurityGroupsTestJSON-509490211 tempest-SecurityGroupsTestJSON-509490211-project-member] [instance: 59fc11e8-27be-4553-9de9-3eb5ea612ec4] Building network info cache for instance {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 568.629635] env[61923]: DEBUG oslo_vmware.api [None req-1c44cc0c-5505-4b32-99eb-cbff6c4edcc0 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] Task: {'id': task-1377453, 'name': Rename_Task, 'duration_secs': 0.143361} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 568.629635] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-1c44cc0c-5505-4b32-99eb-cbff6c4edcc0 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] [instance: f7cc960c-e06a-4c58-9367-ec8771fe09d5] Powering on the VM {{(pid=61923) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 568.629635] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ca1cddaf-be2f-4213-bea5-445ea751adc8 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 568.638032] env[61923]: DEBUG oslo_vmware.api [None req-1c44cc0c-5505-4b32-99eb-cbff6c4edcc0 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] Waiting for the task: (returnval){ [ 568.638032] env[61923]: value = "task-1377454" [ 568.638032] env[61923]: _type = "Task" [ 568.638032] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 568.651874] env[61923]: DEBUG oslo_vmware.api [None req-1c44cc0c-5505-4b32-99eb-cbff6c4edcc0 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] Task: {'id': task-1377454, 'name': PowerOnVM_Task} progress is 33%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 568.744712] env[61923]: DEBUG nova.network.neutron [None req-f30f4038-8fd0-4bbd-8b35-7fb8e0dea3a4 tempest-SecurityGroupsTestJSON-509490211 tempest-SecurityGroupsTestJSON-509490211-project-member] [instance: 59fc11e8-27be-4553-9de9-3eb5ea612ec4] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 569.126075] env[61923]: DEBUG nova.network.neutron [None req-f30f4038-8fd0-4bbd-8b35-7fb8e0dea3a4 tempest-SecurityGroupsTestJSON-509490211 tempest-SecurityGroupsTestJSON-509490211-project-member] [instance: 59fc11e8-27be-4553-9de9-3eb5ea612ec4] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 569.152485] env[61923]: DEBUG oslo_vmware.api [None req-1c44cc0c-5505-4b32-99eb-cbff6c4edcc0 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] Task: {'id': task-1377454, 'name': PowerOnVM_Task, 'duration_secs': 0.431919} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 569.153716] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-1c44cc0c-5505-4b32-99eb-cbff6c4edcc0 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] [instance: f7cc960c-e06a-4c58-9367-ec8771fe09d5] Powered on the VM {{(pid=61923) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 569.154028] env[61923]: DEBUG nova.compute.manager [None req-1c44cc0c-5505-4b32-99eb-cbff6c4edcc0 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] [instance: f7cc960c-e06a-4c58-9367-ec8771fe09d5] Checking state {{(pid=61923) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 569.155015] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea6bb326-c7d4-4792-9b79-521dccc1f72a {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 569.615789] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d2fcbeb-ec3b-4d6e-85ea-359c54227355 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 569.624305] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-405a24d1-a0a4-485d-873d-ebc06f807023 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 569.630188] env[61923]: DEBUG oslo_concurrency.lockutils [None req-f30f4038-8fd0-4bbd-8b35-7fb8e0dea3a4 tempest-SecurityGroupsTestJSON-509490211 tempest-SecurityGroupsTestJSON-509490211-project-member] Releasing lock "refresh_cache-59fc11e8-27be-4553-9de9-3eb5ea612ec4" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 569.630188] env[61923]: DEBUG nova.compute.manager [None req-f30f4038-8fd0-4bbd-8b35-7fb8e0dea3a4 tempest-SecurityGroupsTestJSON-509490211 tempest-SecurityGroupsTestJSON-509490211-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61923) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 569.630188] env[61923]: DEBUG nova.compute.manager [None req-f30f4038-8fd0-4bbd-8b35-7fb8e0dea3a4 tempest-SecurityGroupsTestJSON-509490211 tempest-SecurityGroupsTestJSON-509490211-project-member] [instance: 59fc11e8-27be-4553-9de9-3eb5ea612ec4] Deallocating network for instance {{(pid=61923) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 569.630553] env[61923]: DEBUG nova.network.neutron [None req-f30f4038-8fd0-4bbd-8b35-7fb8e0dea3a4 tempest-SecurityGroupsTestJSON-509490211 tempest-SecurityGroupsTestJSON-509490211-project-member] [instance: 59fc11e8-27be-4553-9de9-3eb5ea612ec4] deallocate_for_instance() {{(pid=61923) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 569.663486] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad1d846a-3373-41b7-a49b-fe25eb3735b5 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 569.674321] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c645bbb-8f34-471f-b008-abd0dd7a47da {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 569.695829] env[61923]: DEBUG nova.compute.provider_tree [None req-36da9865-5d98-4c2a-a964-964bee74523c tempest-ListServerFiltersTestJSON-483030130 tempest-ListServerFiltersTestJSON-483030130-project-member] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 569.699440] env[61923]: DEBUG oslo_concurrency.lockutils [None req-1c44cc0c-5505-4b32-99eb-cbff6c4edcc0 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 569.718597] env[61923]: DEBUG nova.network.neutron [None req-f30f4038-8fd0-4bbd-8b35-7fb8e0dea3a4 tempest-SecurityGroupsTestJSON-509490211 tempest-SecurityGroupsTestJSON-509490211-project-member] [instance: 59fc11e8-27be-4553-9de9-3eb5ea612ec4] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 570.202042] env[61923]: DEBUG nova.scheduler.client.report [None req-36da9865-5d98-4c2a-a964-964bee74523c tempest-ListServerFiltersTestJSON-483030130 tempest-ListServerFiltersTestJSON-483030130-project-member] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 570.222186] env[61923]: DEBUG nova.network.neutron [None req-f30f4038-8fd0-4bbd-8b35-7fb8e0dea3a4 tempest-SecurityGroupsTestJSON-509490211 tempest-SecurityGroupsTestJSON-509490211-project-member] [instance: 59fc11e8-27be-4553-9de9-3eb5ea612ec4] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 570.403355] env[61923]: DEBUG oslo_concurrency.lockutils [None req-85ac5bbf-a6d0-481b-b0bf-1428265b44a9 tempest-AttachInterfacesV270Test-411696101 tempest-AttachInterfacesV270Test-411696101-project-member] Acquiring lock "525f37d5-2262-40c8-a339-fc262c53ce6e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 570.405027] env[61923]: DEBUG oslo_concurrency.lockutils [None req-85ac5bbf-a6d0-481b-b0bf-1428265b44a9 tempest-AttachInterfacesV270Test-411696101 tempest-AttachInterfacesV270Test-411696101-project-member] Lock "525f37d5-2262-40c8-a339-fc262c53ce6e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 570.708469] env[61923]: DEBUG oslo_concurrency.lockutils [None req-36da9865-5d98-4c2a-a964-964bee74523c tempest-ListServerFiltersTestJSON-483030130 tempest-ListServerFiltersTestJSON-483030130-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.514s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 570.708469] env[61923]: DEBUG nova.compute.manager [None req-36da9865-5d98-4c2a-a964-964bee74523c tempest-ListServerFiltersTestJSON-483030130 tempest-ListServerFiltersTestJSON-483030130-project-member] [instance: f3f872ce-cbbe-4407-b4e0-3341bb598825] Start building networks asynchronously for instance. {{(pid=61923) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 570.711934] env[61923]: DEBUG oslo_concurrency.lockutils [None req-dcb477a6-85d7-4838-be85-b9b2ecf96877 tempest-ImagesOneServerNegativeTestJSON-1560548781 tempest-ImagesOneServerNegativeTestJSON-1560548781-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 24.993s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 570.731757] env[61923]: INFO nova.compute.manager [None req-f30f4038-8fd0-4bbd-8b35-7fb8e0dea3a4 tempest-SecurityGroupsTestJSON-509490211 tempest-SecurityGroupsTestJSON-509490211-project-member] [instance: 59fc11e8-27be-4553-9de9-3eb5ea612ec4] Took 1.10 seconds to deallocate network for instance. [ 571.220095] env[61923]: DEBUG nova.compute.utils [None req-36da9865-5d98-4c2a-a964-964bee74523c tempest-ListServerFiltersTestJSON-483030130 tempest-ListServerFiltersTestJSON-483030130-project-member] Using /dev/sd instead of None {{(pid=61923) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 571.226933] env[61923]: DEBUG nova.compute.manager [None req-36da9865-5d98-4c2a-a964-964bee74523c tempest-ListServerFiltersTestJSON-483030130 tempest-ListServerFiltersTestJSON-483030130-project-member] [instance: f3f872ce-cbbe-4407-b4e0-3341bb598825] Allocating IP information in the background. {{(pid=61923) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 571.238370] env[61923]: DEBUG nova.network.neutron [None req-36da9865-5d98-4c2a-a964-964bee74523c tempest-ListServerFiltersTestJSON-483030130 tempest-ListServerFiltersTestJSON-483030130-project-member] [instance: f3f872ce-cbbe-4407-b4e0-3341bb598825] allocate_for_instance() {{(pid=61923) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 571.495075] env[61923]: DEBUG nova.policy [None req-36da9865-5d98-4c2a-a964-964bee74523c tempest-ListServerFiltersTestJSON-483030130 tempest-ListServerFiltersTestJSON-483030130-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '16836760c008401586e5b67b19a53cc0', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '6c5ff14dc18843cc875b21b949eca824', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61923) authorize /opt/stack/nova/nova/policy.py:201}} [ 571.676526] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3389b91-565d-4b8c-8a94-799dfd81783f {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 571.686355] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce371562-add4-49a0-be73-099278c2ab21 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 571.728685] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1bd11e33-6e59-4d7a-8675-30db6613f8c8 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 571.728685] env[61923]: DEBUG nova.compute.manager [None req-36da9865-5d98-4c2a-a964-964bee74523c tempest-ListServerFiltersTestJSON-483030130 tempest-ListServerFiltersTestJSON-483030130-project-member] [instance: f3f872ce-cbbe-4407-b4e0-3341bb598825] Start building block device mappings for instance. {{(pid=61923) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 571.741020] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2f3d45a-fb3b-469d-92c1-49ff4155c075 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 571.761849] env[61923]: DEBUG nova.compute.provider_tree [None req-dcb477a6-85d7-4838-be85-b9b2ecf96877 tempest-ImagesOneServerNegativeTestJSON-1560548781 tempest-ImagesOneServerNegativeTestJSON-1560548781-project-member] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 571.781539] env[61923]: INFO nova.scheduler.client.report [None req-f30f4038-8fd0-4bbd-8b35-7fb8e0dea3a4 tempest-SecurityGroupsTestJSON-509490211 tempest-SecurityGroupsTestJSON-509490211-project-member] Deleted allocations for instance 59fc11e8-27be-4553-9de9-3eb5ea612ec4 [ 572.237664] env[61923]: INFO nova.compute.manager [None req-ae7ff058-a25b-42a3-af4b-0715677fedd4 tempest-ServersAdmin275Test-1076075580 tempest-ServersAdmin275Test-1076075580-project-admin] [instance: f7cc960c-e06a-4c58-9367-ec8771fe09d5] Rebuilding instance [ 572.264933] env[61923]: DEBUG nova.scheduler.client.report [None req-dcb477a6-85d7-4838-be85-b9b2ecf96877 tempest-ImagesOneServerNegativeTestJSON-1560548781 tempest-ImagesOneServerNegativeTestJSON-1560548781-project-member] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 572.298924] env[61923]: DEBUG oslo_concurrency.lockutils [None req-f30f4038-8fd0-4bbd-8b35-7fb8e0dea3a4 tempest-SecurityGroupsTestJSON-509490211 tempest-SecurityGroupsTestJSON-509490211-project-member] Lock "59fc11e8-27be-4553-9de9-3eb5ea612ec4" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 50.342s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 572.322520] env[61923]: DEBUG nova.compute.manager [None req-ae7ff058-a25b-42a3-af4b-0715677fedd4 tempest-ServersAdmin275Test-1076075580 tempest-ServersAdmin275Test-1076075580-project-admin] [instance: f7cc960c-e06a-4c58-9367-ec8771fe09d5] Checking state {{(pid=61923) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 572.323088] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a79ed569-8050-47b6-91c9-444721196d0f {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 572.743186] env[61923]: DEBUG nova.compute.manager [None req-36da9865-5d98-4c2a-a964-964bee74523c tempest-ListServerFiltersTestJSON-483030130 tempest-ListServerFiltersTestJSON-483030130-project-member] [instance: f3f872ce-cbbe-4407-b4e0-3341bb598825] Start spawning the instance on the hypervisor. {{(pid=61923) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 572.773544] env[61923]: DEBUG nova.network.neutron [None req-36da9865-5d98-4c2a-a964-964bee74523c tempest-ListServerFiltersTestJSON-483030130 tempest-ListServerFiltersTestJSON-483030130-project-member] [instance: f3f872ce-cbbe-4407-b4e0-3341bb598825] Successfully created port: 793a91d7-2744-40de-8e41-428eacfdaee8 {{(pid=61923) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 572.780937] env[61923]: DEBUG oslo_concurrency.lockutils [None req-dcb477a6-85d7-4838-be85-b9b2ecf96877 tempest-ImagesOneServerNegativeTestJSON-1560548781 tempest-ImagesOneServerNegativeTestJSON-1560548781-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.066s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 572.782113] env[61923]: ERROR nova.compute.manager [None req-dcb477a6-85d7-4838-be85-b9b2ecf96877 tempest-ImagesOneServerNegativeTestJSON-1560548781 tempest-ImagesOneServerNegativeTestJSON-1560548781-project-member] [instance: ca62f959-43b9-4838-8349-11b7d6a7f0a7] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 63e8ed78-99f3-41e0-b0a6-e6bc6fe00fd4, please check neutron logs for more information. [ 572.782113] env[61923]: ERROR nova.compute.manager [instance: ca62f959-43b9-4838-8349-11b7d6a7f0a7] Traceback (most recent call last): [ 572.782113] env[61923]: ERROR nova.compute.manager [instance: ca62f959-43b9-4838-8349-11b7d6a7f0a7] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 572.782113] env[61923]: ERROR nova.compute.manager [instance: ca62f959-43b9-4838-8349-11b7d6a7f0a7] self.driver.spawn(context, instance, image_meta, [ 572.782113] env[61923]: ERROR nova.compute.manager [instance: ca62f959-43b9-4838-8349-11b7d6a7f0a7] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 572.782113] env[61923]: ERROR nova.compute.manager [instance: ca62f959-43b9-4838-8349-11b7d6a7f0a7] self._vmops.spawn(context, instance, image_meta, injected_files, [ 572.782113] env[61923]: ERROR nova.compute.manager [instance: ca62f959-43b9-4838-8349-11b7d6a7f0a7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 572.782113] env[61923]: ERROR nova.compute.manager [instance: ca62f959-43b9-4838-8349-11b7d6a7f0a7] vm_ref = self.build_virtual_machine(instance, [ 572.782113] env[61923]: ERROR nova.compute.manager [instance: ca62f959-43b9-4838-8349-11b7d6a7f0a7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 572.782113] env[61923]: ERROR nova.compute.manager [instance: ca62f959-43b9-4838-8349-11b7d6a7f0a7] vif_infos = vmwarevif.get_vif_info(self._session, [ 572.782113] env[61923]: ERROR nova.compute.manager [instance: ca62f959-43b9-4838-8349-11b7d6a7f0a7] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 572.782469] env[61923]: ERROR nova.compute.manager [instance: ca62f959-43b9-4838-8349-11b7d6a7f0a7] for vif in network_info: [ 572.782469] env[61923]: ERROR nova.compute.manager [instance: ca62f959-43b9-4838-8349-11b7d6a7f0a7] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 572.782469] env[61923]: ERROR nova.compute.manager [instance: ca62f959-43b9-4838-8349-11b7d6a7f0a7] return self._sync_wrapper(fn, *args, **kwargs) [ 572.782469] env[61923]: ERROR nova.compute.manager [instance: ca62f959-43b9-4838-8349-11b7d6a7f0a7] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 572.782469] env[61923]: ERROR nova.compute.manager [instance: ca62f959-43b9-4838-8349-11b7d6a7f0a7] self.wait() [ 572.782469] env[61923]: ERROR nova.compute.manager [instance: ca62f959-43b9-4838-8349-11b7d6a7f0a7] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 572.782469] env[61923]: ERROR nova.compute.manager [instance: ca62f959-43b9-4838-8349-11b7d6a7f0a7] self[:] = self._gt.wait() [ 572.782469] env[61923]: ERROR nova.compute.manager [instance: ca62f959-43b9-4838-8349-11b7d6a7f0a7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 572.782469] env[61923]: ERROR nova.compute.manager [instance: ca62f959-43b9-4838-8349-11b7d6a7f0a7] return self._exit_event.wait() [ 572.782469] env[61923]: ERROR nova.compute.manager [instance: ca62f959-43b9-4838-8349-11b7d6a7f0a7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 572.782469] env[61923]: ERROR nova.compute.manager [instance: ca62f959-43b9-4838-8349-11b7d6a7f0a7] result = hub.switch() [ 572.782469] env[61923]: ERROR nova.compute.manager [instance: ca62f959-43b9-4838-8349-11b7d6a7f0a7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 572.782469] env[61923]: ERROR nova.compute.manager [instance: ca62f959-43b9-4838-8349-11b7d6a7f0a7] return self.greenlet.switch() [ 572.782861] env[61923]: ERROR nova.compute.manager [instance: ca62f959-43b9-4838-8349-11b7d6a7f0a7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 572.782861] env[61923]: ERROR nova.compute.manager [instance: ca62f959-43b9-4838-8349-11b7d6a7f0a7] result = function(*args, **kwargs) [ 572.782861] env[61923]: ERROR nova.compute.manager [instance: ca62f959-43b9-4838-8349-11b7d6a7f0a7] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 572.782861] env[61923]: ERROR nova.compute.manager [instance: ca62f959-43b9-4838-8349-11b7d6a7f0a7] return func(*args, **kwargs) [ 572.782861] env[61923]: ERROR nova.compute.manager [instance: ca62f959-43b9-4838-8349-11b7d6a7f0a7] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 572.782861] env[61923]: ERROR nova.compute.manager [instance: ca62f959-43b9-4838-8349-11b7d6a7f0a7] raise e [ 572.782861] env[61923]: ERROR nova.compute.manager [instance: ca62f959-43b9-4838-8349-11b7d6a7f0a7] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 572.782861] env[61923]: ERROR nova.compute.manager [instance: ca62f959-43b9-4838-8349-11b7d6a7f0a7] nwinfo = self.network_api.allocate_for_instance( [ 572.782861] env[61923]: ERROR nova.compute.manager [instance: ca62f959-43b9-4838-8349-11b7d6a7f0a7] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 572.782861] env[61923]: ERROR nova.compute.manager [instance: ca62f959-43b9-4838-8349-11b7d6a7f0a7] created_port_ids = self._update_ports_for_instance( [ 572.782861] env[61923]: ERROR nova.compute.manager [instance: ca62f959-43b9-4838-8349-11b7d6a7f0a7] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 572.782861] env[61923]: ERROR nova.compute.manager [instance: ca62f959-43b9-4838-8349-11b7d6a7f0a7] with excutils.save_and_reraise_exception(): [ 572.782861] env[61923]: ERROR nova.compute.manager [instance: ca62f959-43b9-4838-8349-11b7d6a7f0a7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 572.783257] env[61923]: ERROR nova.compute.manager [instance: ca62f959-43b9-4838-8349-11b7d6a7f0a7] self.force_reraise() [ 572.783257] env[61923]: ERROR nova.compute.manager [instance: ca62f959-43b9-4838-8349-11b7d6a7f0a7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 572.783257] env[61923]: ERROR nova.compute.manager [instance: ca62f959-43b9-4838-8349-11b7d6a7f0a7] raise self.value [ 572.783257] env[61923]: ERROR nova.compute.manager [instance: ca62f959-43b9-4838-8349-11b7d6a7f0a7] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 572.783257] env[61923]: ERROR nova.compute.manager [instance: ca62f959-43b9-4838-8349-11b7d6a7f0a7] updated_port = self._update_port( [ 572.783257] env[61923]: ERROR nova.compute.manager [instance: ca62f959-43b9-4838-8349-11b7d6a7f0a7] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 572.783257] env[61923]: ERROR nova.compute.manager [instance: ca62f959-43b9-4838-8349-11b7d6a7f0a7] _ensure_no_port_binding_failure(port) [ 572.783257] env[61923]: ERROR nova.compute.manager [instance: ca62f959-43b9-4838-8349-11b7d6a7f0a7] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 572.783257] env[61923]: ERROR nova.compute.manager [instance: ca62f959-43b9-4838-8349-11b7d6a7f0a7] raise exception.PortBindingFailed(port_id=port['id']) [ 572.783257] env[61923]: ERROR nova.compute.manager [instance: ca62f959-43b9-4838-8349-11b7d6a7f0a7] nova.exception.PortBindingFailed: Binding failed for port 63e8ed78-99f3-41e0-b0a6-e6bc6fe00fd4, please check neutron logs for more information. [ 572.783257] env[61923]: ERROR nova.compute.manager [instance: ca62f959-43b9-4838-8349-11b7d6a7f0a7] [ 572.783602] env[61923]: DEBUG nova.compute.utils [None req-dcb477a6-85d7-4838-be85-b9b2ecf96877 tempest-ImagesOneServerNegativeTestJSON-1560548781 tempest-ImagesOneServerNegativeTestJSON-1560548781-project-member] [instance: ca62f959-43b9-4838-8349-11b7d6a7f0a7] Binding failed for port 63e8ed78-99f3-41e0-b0a6-e6bc6fe00fd4, please check neutron logs for more information. {{(pid=61923) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 572.787934] env[61923]: DEBUG oslo_concurrency.lockutils [None req-fba27a13-474d-4fe3-9187-c31143f40aa5 tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 25.955s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 572.789232] env[61923]: INFO nova.compute.claims [None req-fba27a13-474d-4fe3-9187-c31143f40aa5 tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] [instance: 6a2d7ede-e0cc-4a36-b1b2-0b36b269ecb5] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 572.793281] env[61923]: DEBUG nova.compute.manager [None req-dcb477a6-85d7-4838-be85-b9b2ecf96877 tempest-ImagesOneServerNegativeTestJSON-1560548781 tempest-ImagesOneServerNegativeTestJSON-1560548781-project-member] [instance: ca62f959-43b9-4838-8349-11b7d6a7f0a7] Build of instance ca62f959-43b9-4838-8349-11b7d6a7f0a7 was re-scheduled: Binding failed for port 63e8ed78-99f3-41e0-b0a6-e6bc6fe00fd4, please check neutron logs for more information. {{(pid=61923) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 572.793724] env[61923]: DEBUG nova.compute.manager [None req-dcb477a6-85d7-4838-be85-b9b2ecf96877 tempest-ImagesOneServerNegativeTestJSON-1560548781 tempest-ImagesOneServerNegativeTestJSON-1560548781-project-member] [instance: ca62f959-43b9-4838-8349-11b7d6a7f0a7] Unplugging VIFs for instance {{(pid=61923) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 572.793980] env[61923]: DEBUG oslo_concurrency.lockutils [None req-dcb477a6-85d7-4838-be85-b9b2ecf96877 tempest-ImagesOneServerNegativeTestJSON-1560548781 tempest-ImagesOneServerNegativeTestJSON-1560548781-project-member] Acquiring lock "refresh_cache-ca62f959-43b9-4838-8349-11b7d6a7f0a7" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 572.794089] env[61923]: DEBUG oslo_concurrency.lockutils [None req-dcb477a6-85d7-4838-be85-b9b2ecf96877 tempest-ImagesOneServerNegativeTestJSON-1560548781 tempest-ImagesOneServerNegativeTestJSON-1560548781-project-member] Acquired lock "refresh_cache-ca62f959-43b9-4838-8349-11b7d6a7f0a7" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 572.794244] env[61923]: DEBUG nova.network.neutron [None req-dcb477a6-85d7-4838-be85-b9b2ecf96877 tempest-ImagesOneServerNegativeTestJSON-1560548781 tempest-ImagesOneServerNegativeTestJSON-1560548781-project-member] [instance: ca62f959-43b9-4838-8349-11b7d6a7f0a7] Building network info cache for instance {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 572.805889] env[61923]: DEBUG nova.virt.hardware [None req-36da9865-5d98-4c2a-a964-964bee74523c tempest-ListServerFiltersTestJSON-483030130 tempest-ListServerFiltersTestJSON-483030130-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-29T20:07:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-29T20:07:35Z,direct_url=,disk_format='vmdk',id=0f153f63-ae0a-45b1-b7f5-7f9b673c947f,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='4e7b5e3b3c3443c8bd5c70f8a15739f5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-29T20:07:36Z,virtual_size=,visibility=), allow threads: False {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 572.806172] env[61923]: DEBUG nova.virt.hardware [None req-36da9865-5d98-4c2a-a964-964bee74523c tempest-ListServerFiltersTestJSON-483030130 tempest-ListServerFiltersTestJSON-483030130-project-member] Flavor limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 572.806304] env[61923]: DEBUG nova.virt.hardware [None req-36da9865-5d98-4c2a-a964-964bee74523c tempest-ListServerFiltersTestJSON-483030130 tempest-ListServerFiltersTestJSON-483030130-project-member] Image limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 572.806503] env[61923]: DEBUG nova.virt.hardware [None req-36da9865-5d98-4c2a-a964-964bee74523c tempest-ListServerFiltersTestJSON-483030130 tempest-ListServerFiltersTestJSON-483030130-project-member] Flavor pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 572.806676] env[61923]: DEBUG nova.virt.hardware [None req-36da9865-5d98-4c2a-a964-964bee74523c tempest-ListServerFiltersTestJSON-483030130 tempest-ListServerFiltersTestJSON-483030130-project-member] Image pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 572.806821] env[61923]: DEBUG nova.virt.hardware [None req-36da9865-5d98-4c2a-a964-964bee74523c tempest-ListServerFiltersTestJSON-483030130 tempest-ListServerFiltersTestJSON-483030130-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 572.808435] env[61923]: DEBUG nova.virt.hardware [None req-36da9865-5d98-4c2a-a964-964bee74523c tempest-ListServerFiltersTestJSON-483030130 tempest-ListServerFiltersTestJSON-483030130-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 572.808818] env[61923]: DEBUG nova.virt.hardware [None req-36da9865-5d98-4c2a-a964-964bee74523c tempest-ListServerFiltersTestJSON-483030130 tempest-ListServerFiltersTestJSON-483030130-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 572.809011] env[61923]: DEBUG nova.virt.hardware [None req-36da9865-5d98-4c2a-a964-964bee74523c tempest-ListServerFiltersTestJSON-483030130 tempest-ListServerFiltersTestJSON-483030130-project-member] Got 1 possible topologies {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 572.809444] env[61923]: DEBUG nova.virt.hardware [None req-36da9865-5d98-4c2a-a964-964bee74523c tempest-ListServerFiltersTestJSON-483030130 tempest-ListServerFiltersTestJSON-483030130-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 572.809444] env[61923]: DEBUG nova.virt.hardware [None req-36da9865-5d98-4c2a-a964-964bee74523c tempest-ListServerFiltersTestJSON-483030130 tempest-ListServerFiltersTestJSON-483030130-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 572.809937] env[61923]: DEBUG nova.compute.manager [None req-595ebb22-536b-4375-9cbe-6fc472a5a17d tempest-ListServerFiltersTestJSON-483030130 tempest-ListServerFiltersTestJSON-483030130-project-member] [instance: 71d9eccc-3b90-4b0a-a59a-d6a8e08a9417] Starting instance... {{(pid=61923) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 572.814403] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3647fe1c-40b1-4381-b187-4369cd2720a2 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 572.827970] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03b7d61f-d4a5-4174-b78e-7955f8c0c757 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 572.848698] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-ae7ff058-a25b-42a3-af4b-0715677fedd4 tempest-ServersAdmin275Test-1076075580 tempest-ServersAdmin275Test-1076075580-project-admin] [instance: f7cc960c-e06a-4c58-9367-ec8771fe09d5] Powering off the VM {{(pid=61923) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 572.848698] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-584dd16e-2e94-4bf9-a299-623b042dcf1e {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 572.861293] env[61923]: DEBUG oslo_vmware.api [None req-ae7ff058-a25b-42a3-af4b-0715677fedd4 tempest-ServersAdmin275Test-1076075580 tempest-ServersAdmin275Test-1076075580-project-admin] Waiting for the task: (returnval){ [ 572.861293] env[61923]: value = "task-1377459" [ 572.861293] env[61923]: _type = "Task" [ 572.861293] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 572.877045] env[61923]: DEBUG oslo_vmware.api [None req-ae7ff058-a25b-42a3-af4b-0715677fedd4 tempest-ServersAdmin275Test-1076075580 tempest-ServersAdmin275Test-1076075580-project-admin] Task: {'id': task-1377459, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 572.986744] env[61923]: DEBUG oslo_concurrency.lockutils [None req-4bd5c12b-d52a-4c54-93f7-87894f09013a tempest-ServersTestJSON-483293076 tempest-ServersTestJSON-483293076-project-member] Acquiring lock "55b60fc2-6484-4fad-aa40-3e73e020841a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 572.987035] env[61923]: DEBUG oslo_concurrency.lockutils [None req-4bd5c12b-d52a-4c54-93f7-87894f09013a tempest-ServersTestJSON-483293076 tempest-ServersTestJSON-483293076-project-member] Lock "55b60fc2-6484-4fad-aa40-3e73e020841a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 573.334234] env[61923]: DEBUG oslo_concurrency.lockutils [None req-595ebb22-536b-4375-9cbe-6fc472a5a17d tempest-ListServerFiltersTestJSON-483030130 tempest-ListServerFiltersTestJSON-483030130-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 573.355495] env[61923]: DEBUG nova.network.neutron [None req-dcb477a6-85d7-4838-be85-b9b2ecf96877 tempest-ImagesOneServerNegativeTestJSON-1560548781 tempest-ImagesOneServerNegativeTestJSON-1560548781-project-member] [instance: ca62f959-43b9-4838-8349-11b7d6a7f0a7] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 573.371386] env[61923]: DEBUG oslo_vmware.api [None req-ae7ff058-a25b-42a3-af4b-0715677fedd4 tempest-ServersAdmin275Test-1076075580 tempest-ServersAdmin275Test-1076075580-project-admin] Task: {'id': task-1377459, 'name': PowerOffVM_Task, 'duration_secs': 0.162558} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 573.371914] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-ae7ff058-a25b-42a3-af4b-0715677fedd4 tempest-ServersAdmin275Test-1076075580 tempest-ServersAdmin275Test-1076075580-project-admin] [instance: f7cc960c-e06a-4c58-9367-ec8771fe09d5] Powered off the VM {{(pid=61923) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 573.372296] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-ae7ff058-a25b-42a3-af4b-0715677fedd4 tempest-ServersAdmin275Test-1076075580 tempest-ServersAdmin275Test-1076075580-project-admin] [instance: f7cc960c-e06a-4c58-9367-ec8771fe09d5] Destroying instance {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 573.373350] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d93cec4d-f64a-4f6f-ab7d-23d90c25d62d {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 573.381180] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-ae7ff058-a25b-42a3-af4b-0715677fedd4 tempest-ServersAdmin275Test-1076075580 tempest-ServersAdmin275Test-1076075580-project-admin] [instance: f7cc960c-e06a-4c58-9367-ec8771fe09d5] Unregistering the VM {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 573.383022] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c63e405d-d8eb-4be7-bf73-f0692c678ba8 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 573.406397] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-ae7ff058-a25b-42a3-af4b-0715677fedd4 tempest-ServersAdmin275Test-1076075580 tempest-ServersAdmin275Test-1076075580-project-admin] [instance: f7cc960c-e06a-4c58-9367-ec8771fe09d5] Unregistered the VM {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 573.406710] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-ae7ff058-a25b-42a3-af4b-0715677fedd4 tempest-ServersAdmin275Test-1076075580 tempest-ServersAdmin275Test-1076075580-project-admin] [instance: f7cc960c-e06a-4c58-9367-ec8771fe09d5] Deleting contents of the VM from datastore datastore2 {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 573.406969] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-ae7ff058-a25b-42a3-af4b-0715677fedd4 tempest-ServersAdmin275Test-1076075580 tempest-ServersAdmin275Test-1076075580-project-admin] Deleting the datastore file [datastore2] f7cc960c-e06a-4c58-9367-ec8771fe09d5 {{(pid=61923) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 573.407324] env[61923]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-8484cfc1-9d4f-41d1-a797-2dbcedc55176 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 573.414167] env[61923]: DEBUG oslo_vmware.api [None req-ae7ff058-a25b-42a3-af4b-0715677fedd4 tempest-ServersAdmin275Test-1076075580 tempest-ServersAdmin275Test-1076075580-project-admin] Waiting for the task: (returnval){ [ 573.414167] env[61923]: value = "task-1377461" [ 573.414167] env[61923]: _type = "Task" [ 573.414167] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 573.425068] env[61923]: DEBUG oslo_vmware.api [None req-ae7ff058-a25b-42a3-af4b-0715677fedd4 tempest-ServersAdmin275Test-1076075580 tempest-ServersAdmin275Test-1076075580-project-admin] Task: {'id': task-1377461, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 573.737545] env[61923]: DEBUG nova.network.neutron [None req-dcb477a6-85d7-4838-be85-b9b2ecf96877 tempest-ImagesOneServerNegativeTestJSON-1560548781 tempest-ImagesOneServerNegativeTestJSON-1560548781-project-member] [instance: ca62f959-43b9-4838-8349-11b7d6a7f0a7] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 573.807358] env[61923]: DEBUG oslo_concurrency.lockutils [None req-56950b6c-0b87-48bf-a575-f699fcd1a5b0 tempest-FloatingIPsAssociationTestJSON-1125124803 tempest-FloatingIPsAssociationTestJSON-1125124803-project-member] Acquiring lock "aeca6a94-478d-435a-9d93-59936f26654e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 573.807830] env[61923]: DEBUG oslo_concurrency.lockutils [None req-56950b6c-0b87-48bf-a575-f699fcd1a5b0 tempest-FloatingIPsAssociationTestJSON-1125124803 tempest-FloatingIPsAssociationTestJSON-1125124803-project-member] Lock "aeca6a94-478d-435a-9d93-59936f26654e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 573.935595] env[61923]: DEBUG oslo_vmware.api [None req-ae7ff058-a25b-42a3-af4b-0715677fedd4 tempest-ServersAdmin275Test-1076075580 tempest-ServersAdmin275Test-1076075580-project-admin] Task: {'id': task-1377461, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.181827} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 573.935595] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-ae7ff058-a25b-42a3-af4b-0715677fedd4 tempest-ServersAdmin275Test-1076075580 tempest-ServersAdmin275Test-1076075580-project-admin] Deleted the datastore file {{(pid=61923) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 573.935811] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-ae7ff058-a25b-42a3-af4b-0715677fedd4 tempest-ServersAdmin275Test-1076075580 tempest-ServersAdmin275Test-1076075580-project-admin] [instance: f7cc960c-e06a-4c58-9367-ec8771fe09d5] Deleted contents of the VM from datastore datastore2 {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 573.935967] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-ae7ff058-a25b-42a3-af4b-0715677fedd4 tempest-ServersAdmin275Test-1076075580 tempest-ServersAdmin275Test-1076075580-project-admin] [instance: f7cc960c-e06a-4c58-9367-ec8771fe09d5] Instance destroyed {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 574.245214] env[61923]: DEBUG oslo_concurrency.lockutils [None req-dcb477a6-85d7-4838-be85-b9b2ecf96877 tempest-ImagesOneServerNegativeTestJSON-1560548781 tempest-ImagesOneServerNegativeTestJSON-1560548781-project-member] Releasing lock "refresh_cache-ca62f959-43b9-4838-8349-11b7d6a7f0a7" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 574.245445] env[61923]: DEBUG nova.compute.manager [None req-dcb477a6-85d7-4838-be85-b9b2ecf96877 tempest-ImagesOneServerNegativeTestJSON-1560548781 tempest-ImagesOneServerNegativeTestJSON-1560548781-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61923) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 574.246200] env[61923]: DEBUG nova.compute.manager [None req-dcb477a6-85d7-4838-be85-b9b2ecf96877 tempest-ImagesOneServerNegativeTestJSON-1560548781 tempest-ImagesOneServerNegativeTestJSON-1560548781-project-member] [instance: ca62f959-43b9-4838-8349-11b7d6a7f0a7] Deallocating network for instance {{(pid=61923) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 574.246200] env[61923]: DEBUG nova.network.neutron [None req-dcb477a6-85d7-4838-be85-b9b2ecf96877 tempest-ImagesOneServerNegativeTestJSON-1560548781 tempest-ImagesOneServerNegativeTestJSON-1560548781-project-member] [instance: ca62f959-43b9-4838-8349-11b7d6a7f0a7] deallocate_for_instance() {{(pid=61923) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 574.282546] env[61923]: DEBUG nova.network.neutron [None req-dcb477a6-85d7-4838-be85-b9b2ecf96877 tempest-ImagesOneServerNegativeTestJSON-1560548781 tempest-ImagesOneServerNegativeTestJSON-1560548781-project-member] [instance: ca62f959-43b9-4838-8349-11b7d6a7f0a7] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 574.300355] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a03d7bbd-b954-4db3-9e56-d788443885c2 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 574.309390] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6438d91-a9ff-4e86-8049-d3ee5b01fb96 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 574.343807] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45eb20b0-32c2-4aa4-9ba8-f2c4a3ce9542 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 574.351501] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bfbc3974-5df2-4f25-b9fa-c3c6e3a80b24 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 574.370276] env[61923]: DEBUG nova.compute.provider_tree [None req-fba27a13-474d-4fe3-9187-c31143f40aa5 tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 574.785434] env[61923]: DEBUG nova.network.neutron [None req-dcb477a6-85d7-4838-be85-b9b2ecf96877 tempest-ImagesOneServerNegativeTestJSON-1560548781 tempest-ImagesOneServerNegativeTestJSON-1560548781-project-member] [instance: ca62f959-43b9-4838-8349-11b7d6a7f0a7] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 574.875024] env[61923]: DEBUG nova.scheduler.client.report [None req-fba27a13-474d-4fe3-9187-c31143f40aa5 tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 574.988013] env[61923]: DEBUG nova.virt.hardware [None req-ae7ff058-a25b-42a3-af4b-0715677fedd4 tempest-ServersAdmin275Test-1076075580 tempest-ServersAdmin275Test-1076075580-project-admin] Getting desirable topologies for flavor Flavor(created_at=2024-10-29T20:07:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-29T20:07:35Z,direct_url=,disk_format='vmdk',id=0f153f63-ae0a-45b1-b7f5-7f9b673c947f,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='4e7b5e3b3c3443c8bd5c70f8a15739f5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-29T20:07:36Z,virtual_size=,visibility=), allow threads: False {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 574.988290] env[61923]: DEBUG nova.virt.hardware [None req-ae7ff058-a25b-42a3-af4b-0715677fedd4 tempest-ServersAdmin275Test-1076075580 tempest-ServersAdmin275Test-1076075580-project-admin] Flavor limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 574.988639] env[61923]: DEBUG nova.virt.hardware [None req-ae7ff058-a25b-42a3-af4b-0715677fedd4 tempest-ServersAdmin275Test-1076075580 tempest-ServersAdmin275Test-1076075580-project-admin] Image limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 574.988686] env[61923]: DEBUG nova.virt.hardware [None req-ae7ff058-a25b-42a3-af4b-0715677fedd4 tempest-ServersAdmin275Test-1076075580 tempest-ServersAdmin275Test-1076075580-project-admin] Flavor pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 574.991890] env[61923]: DEBUG nova.virt.hardware [None req-ae7ff058-a25b-42a3-af4b-0715677fedd4 tempest-ServersAdmin275Test-1076075580 tempest-ServersAdmin275Test-1076075580-project-admin] Image pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 574.991890] env[61923]: DEBUG nova.virt.hardware [None req-ae7ff058-a25b-42a3-af4b-0715677fedd4 tempest-ServersAdmin275Test-1076075580 tempest-ServersAdmin275Test-1076075580-project-admin] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 574.991890] env[61923]: DEBUG nova.virt.hardware [None req-ae7ff058-a25b-42a3-af4b-0715677fedd4 tempest-ServersAdmin275Test-1076075580 tempest-ServersAdmin275Test-1076075580-project-admin] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 574.991890] env[61923]: DEBUG nova.virt.hardware [None req-ae7ff058-a25b-42a3-af4b-0715677fedd4 tempest-ServersAdmin275Test-1076075580 tempest-ServersAdmin275Test-1076075580-project-admin] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 574.991890] env[61923]: DEBUG nova.virt.hardware [None req-ae7ff058-a25b-42a3-af4b-0715677fedd4 tempest-ServersAdmin275Test-1076075580 tempest-ServersAdmin275Test-1076075580-project-admin] Got 1 possible topologies {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 574.992538] env[61923]: DEBUG nova.virt.hardware [None req-ae7ff058-a25b-42a3-af4b-0715677fedd4 tempest-ServersAdmin275Test-1076075580 tempest-ServersAdmin275Test-1076075580-project-admin] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 574.992789] env[61923]: DEBUG nova.virt.hardware [None req-ae7ff058-a25b-42a3-af4b-0715677fedd4 tempest-ServersAdmin275Test-1076075580 tempest-ServersAdmin275Test-1076075580-project-admin] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 574.993669] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf19cda6-50be-4f54-8038-f4d6b259d469 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 575.003210] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a06af2f3-ba4c-4ff2-bae1-ec6079f8d6ec {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 575.025068] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-ae7ff058-a25b-42a3-af4b-0715677fedd4 tempest-ServersAdmin275Test-1076075580 tempest-ServersAdmin275Test-1076075580-project-admin] [instance: f7cc960c-e06a-4c58-9367-ec8771fe09d5] Instance VIF info [] {{(pid=61923) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 575.031445] env[61923]: DEBUG oslo.service.loopingcall [None req-ae7ff058-a25b-42a3-af4b-0715677fedd4 tempest-ServersAdmin275Test-1076075580 tempest-ServersAdmin275Test-1076075580-project-admin] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61923) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 575.031770] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f7cc960c-e06a-4c58-9367-ec8771fe09d5] Creating VM on the ESX host {{(pid=61923) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 575.032966] env[61923]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-2622b555-38f9-4080-9935-ba1fbad71561 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 575.049331] env[61923]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 575.049331] env[61923]: value = "task-1377464" [ 575.049331] env[61923]: _type = "Task" [ 575.049331] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 575.058320] env[61923]: DEBUG oslo_vmware.api [-] Task: {'id': task-1377464, 'name': CreateVM_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 575.289070] env[61923]: INFO nova.compute.manager [None req-dcb477a6-85d7-4838-be85-b9b2ecf96877 tempest-ImagesOneServerNegativeTestJSON-1560548781 tempest-ImagesOneServerNegativeTestJSON-1560548781-project-member] [instance: ca62f959-43b9-4838-8349-11b7d6a7f0a7] Took 1.04 seconds to deallocate network for instance. [ 575.387383] env[61923]: DEBUG oslo_concurrency.lockutils [None req-fba27a13-474d-4fe3-9187-c31143f40aa5 tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.597s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 575.387383] env[61923]: DEBUG nova.compute.manager [None req-fba27a13-474d-4fe3-9187-c31143f40aa5 tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] [instance: 6a2d7ede-e0cc-4a36-b1b2-0b36b269ecb5] Start building networks asynchronously for instance. {{(pid=61923) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 575.389404] env[61923]: DEBUG oslo_concurrency.lockutils [None req-d9e2f878-40e6-4b41-8b07-da611f2f81f1 tempest-VolumesAssistedSnapshotsTest-2043366068 tempest-VolumesAssistedSnapshotsTest-2043366068-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 23.384s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 575.564522] env[61923]: DEBUG oslo_vmware.api [-] Task: {'id': task-1377464, 'name': CreateVM_Task} progress is 25%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 575.664561] env[61923]: DEBUG oslo_concurrency.lockutils [None req-859ffdff-aaee-46b4-bdd6-0a15cce9792f tempest-ServerRescueNegativeTestJSON-1435265126 tempest-ServerRescueNegativeTestJSON-1435265126-project-member] Acquiring lock "41b6744e-d548-442c-b6cc-e12bd403f229" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 575.664895] env[61923]: DEBUG oslo_concurrency.lockutils [None req-859ffdff-aaee-46b4-bdd6-0a15cce9792f tempest-ServerRescueNegativeTestJSON-1435265126 tempest-ServerRescueNegativeTestJSON-1435265126-project-member] Lock "41b6744e-d548-442c-b6cc-e12bd403f229" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 575.891941] env[61923]: DEBUG nova.compute.utils [None req-fba27a13-474d-4fe3-9187-c31143f40aa5 tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] Using /dev/sd instead of None {{(pid=61923) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 575.893370] env[61923]: DEBUG nova.compute.manager [None req-fba27a13-474d-4fe3-9187-c31143f40aa5 tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] [instance: 6a2d7ede-e0cc-4a36-b1b2-0b36b269ecb5] Allocating IP information in the background. {{(pid=61923) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 575.893692] env[61923]: DEBUG nova.network.neutron [None req-fba27a13-474d-4fe3-9187-c31143f40aa5 tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] [instance: 6a2d7ede-e0cc-4a36-b1b2-0b36b269ecb5] allocate_for_instance() {{(pid=61923) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 576.014524] env[61923]: DEBUG nova.policy [None req-fba27a13-474d-4fe3-9187-c31143f40aa5 tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '74859dc3474543f0829d2d7bea84bebc', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '609c22f73ddf4b9b9e764d428891f2a8', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61923) authorize /opt/stack/nova/nova/policy.py:201}} [ 576.078919] env[61923]: DEBUG oslo_vmware.api [-] Task: {'id': task-1377464, 'name': CreateVM_Task, 'duration_secs': 0.667807} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 576.079274] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f7cc960c-e06a-4c58-9367-ec8771fe09d5] Created VM on the ESX host {{(pid=61923) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 576.079539] env[61923]: DEBUG oslo_concurrency.lockutils [None req-ae7ff058-a25b-42a3-af4b-0715677fedd4 tempest-ServersAdmin275Test-1076075580 tempest-ServersAdmin275Test-1076075580-project-admin] Acquiring lock "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 576.079767] env[61923]: DEBUG oslo_concurrency.lockutils [None req-ae7ff058-a25b-42a3-af4b-0715677fedd4 tempest-ServersAdmin275Test-1076075580 tempest-ServersAdmin275Test-1076075580-project-admin] Acquired lock "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 576.079998] env[61923]: DEBUG oslo_concurrency.lockutils [None req-ae7ff058-a25b-42a3-af4b-0715677fedd4 tempest-ServersAdmin275Test-1076075580 tempest-ServersAdmin275Test-1076075580-project-admin] Acquired external semaphore "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 576.080259] env[61923]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-edc0f3e0-2208-445d-97b4-864ec460f9bc {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 576.086260] env[61923]: DEBUG oslo_vmware.api [None req-ae7ff058-a25b-42a3-af4b-0715677fedd4 tempest-ServersAdmin275Test-1076075580 tempest-ServersAdmin275Test-1076075580-project-admin] Waiting for the task: (returnval){ [ 576.086260] env[61923]: value = "session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52f58043-f83c-102a-d6e5-5af7f41a6a7d" [ 576.086260] env[61923]: _type = "Task" [ 576.086260] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 576.096778] env[61923]: DEBUG oslo_vmware.api [None req-ae7ff058-a25b-42a3-af4b-0715677fedd4 tempest-ServersAdmin275Test-1076075580 tempest-ServersAdmin275Test-1076075580-project-admin] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52f58043-f83c-102a-d6e5-5af7f41a6a7d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 576.109206] env[61923]: ERROR nova.compute.manager [None req-36da9865-5d98-4c2a-a964-964bee74523c tempest-ListServerFiltersTestJSON-483030130 tempest-ListServerFiltersTestJSON-483030130-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 793a91d7-2744-40de-8e41-428eacfdaee8, please check neutron logs for more information. [ 576.109206] env[61923]: ERROR nova.compute.manager Traceback (most recent call last): [ 576.109206] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 576.109206] env[61923]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 576.109206] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 576.109206] env[61923]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 576.109206] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 576.109206] env[61923]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 576.109206] env[61923]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 576.109206] env[61923]: ERROR nova.compute.manager self.force_reraise() [ 576.109206] env[61923]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 576.109206] env[61923]: ERROR nova.compute.manager raise self.value [ 576.109206] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 576.109206] env[61923]: ERROR nova.compute.manager updated_port = self._update_port( [ 576.109206] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 576.109206] env[61923]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 576.109750] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 576.109750] env[61923]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 576.109750] env[61923]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 793a91d7-2744-40de-8e41-428eacfdaee8, please check neutron logs for more information. [ 576.109750] env[61923]: ERROR nova.compute.manager [ 576.109750] env[61923]: Traceback (most recent call last): [ 576.109750] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 576.109750] env[61923]: listener.cb(fileno) [ 576.109750] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 576.109750] env[61923]: result = function(*args, **kwargs) [ 576.109750] env[61923]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 576.109750] env[61923]: return func(*args, **kwargs) [ 576.109750] env[61923]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 576.109750] env[61923]: raise e [ 576.109750] env[61923]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 576.109750] env[61923]: nwinfo = self.network_api.allocate_for_instance( [ 576.109750] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 576.109750] env[61923]: created_port_ids = self._update_ports_for_instance( [ 576.109750] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 576.109750] env[61923]: with excutils.save_and_reraise_exception(): [ 576.109750] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 576.109750] env[61923]: self.force_reraise() [ 576.109750] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 576.109750] env[61923]: raise self.value [ 576.109750] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 576.109750] env[61923]: updated_port = self._update_port( [ 576.109750] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 576.109750] env[61923]: _ensure_no_port_binding_failure(port) [ 576.109750] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 576.109750] env[61923]: raise exception.PortBindingFailed(port_id=port['id']) [ 576.110721] env[61923]: nova.exception.PortBindingFailed: Binding failed for port 793a91d7-2744-40de-8e41-428eacfdaee8, please check neutron logs for more information. [ 576.110721] env[61923]: Removing descriptor: 15 [ 576.110721] env[61923]: ERROR nova.compute.manager [None req-36da9865-5d98-4c2a-a964-964bee74523c tempest-ListServerFiltersTestJSON-483030130 tempest-ListServerFiltersTestJSON-483030130-project-member] [instance: f3f872ce-cbbe-4407-b4e0-3341bb598825] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 793a91d7-2744-40de-8e41-428eacfdaee8, please check neutron logs for more information. [ 576.110721] env[61923]: ERROR nova.compute.manager [instance: f3f872ce-cbbe-4407-b4e0-3341bb598825] Traceback (most recent call last): [ 576.110721] env[61923]: ERROR nova.compute.manager [instance: f3f872ce-cbbe-4407-b4e0-3341bb598825] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 576.110721] env[61923]: ERROR nova.compute.manager [instance: f3f872ce-cbbe-4407-b4e0-3341bb598825] yield resources [ 576.110721] env[61923]: ERROR nova.compute.manager [instance: f3f872ce-cbbe-4407-b4e0-3341bb598825] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 576.110721] env[61923]: ERROR nova.compute.manager [instance: f3f872ce-cbbe-4407-b4e0-3341bb598825] self.driver.spawn(context, instance, image_meta, [ 576.110721] env[61923]: ERROR nova.compute.manager [instance: f3f872ce-cbbe-4407-b4e0-3341bb598825] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 576.110721] env[61923]: ERROR nova.compute.manager [instance: f3f872ce-cbbe-4407-b4e0-3341bb598825] self._vmops.spawn(context, instance, image_meta, injected_files, [ 576.110721] env[61923]: ERROR nova.compute.manager [instance: f3f872ce-cbbe-4407-b4e0-3341bb598825] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 576.110721] env[61923]: ERROR nova.compute.manager [instance: f3f872ce-cbbe-4407-b4e0-3341bb598825] vm_ref = self.build_virtual_machine(instance, [ 576.111172] env[61923]: ERROR nova.compute.manager [instance: f3f872ce-cbbe-4407-b4e0-3341bb598825] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 576.111172] env[61923]: ERROR nova.compute.manager [instance: f3f872ce-cbbe-4407-b4e0-3341bb598825] vif_infos = vmwarevif.get_vif_info(self._session, [ 576.111172] env[61923]: ERROR nova.compute.manager [instance: f3f872ce-cbbe-4407-b4e0-3341bb598825] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 576.111172] env[61923]: ERROR nova.compute.manager [instance: f3f872ce-cbbe-4407-b4e0-3341bb598825] for vif in network_info: [ 576.111172] env[61923]: ERROR nova.compute.manager [instance: f3f872ce-cbbe-4407-b4e0-3341bb598825] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 576.111172] env[61923]: ERROR nova.compute.manager [instance: f3f872ce-cbbe-4407-b4e0-3341bb598825] return self._sync_wrapper(fn, *args, **kwargs) [ 576.111172] env[61923]: ERROR nova.compute.manager [instance: f3f872ce-cbbe-4407-b4e0-3341bb598825] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 576.111172] env[61923]: ERROR nova.compute.manager [instance: f3f872ce-cbbe-4407-b4e0-3341bb598825] self.wait() [ 576.111172] env[61923]: ERROR nova.compute.manager [instance: f3f872ce-cbbe-4407-b4e0-3341bb598825] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 576.111172] env[61923]: ERROR nova.compute.manager [instance: f3f872ce-cbbe-4407-b4e0-3341bb598825] self[:] = self._gt.wait() [ 576.111172] env[61923]: ERROR nova.compute.manager [instance: f3f872ce-cbbe-4407-b4e0-3341bb598825] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 576.111172] env[61923]: ERROR nova.compute.manager [instance: f3f872ce-cbbe-4407-b4e0-3341bb598825] return self._exit_event.wait() [ 576.111172] env[61923]: ERROR nova.compute.manager [instance: f3f872ce-cbbe-4407-b4e0-3341bb598825] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 576.111639] env[61923]: ERROR nova.compute.manager [instance: f3f872ce-cbbe-4407-b4e0-3341bb598825] result = hub.switch() [ 576.111639] env[61923]: ERROR nova.compute.manager [instance: f3f872ce-cbbe-4407-b4e0-3341bb598825] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 576.111639] env[61923]: ERROR nova.compute.manager [instance: f3f872ce-cbbe-4407-b4e0-3341bb598825] return self.greenlet.switch() [ 576.111639] env[61923]: ERROR nova.compute.manager [instance: f3f872ce-cbbe-4407-b4e0-3341bb598825] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 576.111639] env[61923]: ERROR nova.compute.manager [instance: f3f872ce-cbbe-4407-b4e0-3341bb598825] result = function(*args, **kwargs) [ 576.111639] env[61923]: ERROR nova.compute.manager [instance: f3f872ce-cbbe-4407-b4e0-3341bb598825] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 576.111639] env[61923]: ERROR nova.compute.manager [instance: f3f872ce-cbbe-4407-b4e0-3341bb598825] return func(*args, **kwargs) [ 576.111639] env[61923]: ERROR nova.compute.manager [instance: f3f872ce-cbbe-4407-b4e0-3341bb598825] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 576.111639] env[61923]: ERROR nova.compute.manager [instance: f3f872ce-cbbe-4407-b4e0-3341bb598825] raise e [ 576.111639] env[61923]: ERROR nova.compute.manager [instance: f3f872ce-cbbe-4407-b4e0-3341bb598825] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 576.111639] env[61923]: ERROR nova.compute.manager [instance: f3f872ce-cbbe-4407-b4e0-3341bb598825] nwinfo = self.network_api.allocate_for_instance( [ 576.111639] env[61923]: ERROR nova.compute.manager [instance: f3f872ce-cbbe-4407-b4e0-3341bb598825] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 576.111639] env[61923]: ERROR nova.compute.manager [instance: f3f872ce-cbbe-4407-b4e0-3341bb598825] created_port_ids = self._update_ports_for_instance( [ 576.112085] env[61923]: ERROR nova.compute.manager [instance: f3f872ce-cbbe-4407-b4e0-3341bb598825] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 576.112085] env[61923]: ERROR nova.compute.manager [instance: f3f872ce-cbbe-4407-b4e0-3341bb598825] with excutils.save_and_reraise_exception(): [ 576.112085] env[61923]: ERROR nova.compute.manager [instance: f3f872ce-cbbe-4407-b4e0-3341bb598825] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 576.112085] env[61923]: ERROR nova.compute.manager [instance: f3f872ce-cbbe-4407-b4e0-3341bb598825] self.force_reraise() [ 576.112085] env[61923]: ERROR nova.compute.manager [instance: f3f872ce-cbbe-4407-b4e0-3341bb598825] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 576.112085] env[61923]: ERROR nova.compute.manager [instance: f3f872ce-cbbe-4407-b4e0-3341bb598825] raise self.value [ 576.112085] env[61923]: ERROR nova.compute.manager [instance: f3f872ce-cbbe-4407-b4e0-3341bb598825] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 576.112085] env[61923]: ERROR nova.compute.manager [instance: f3f872ce-cbbe-4407-b4e0-3341bb598825] updated_port = self._update_port( [ 576.112085] env[61923]: ERROR nova.compute.manager [instance: f3f872ce-cbbe-4407-b4e0-3341bb598825] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 576.112085] env[61923]: ERROR nova.compute.manager [instance: f3f872ce-cbbe-4407-b4e0-3341bb598825] _ensure_no_port_binding_failure(port) [ 576.112085] env[61923]: ERROR nova.compute.manager [instance: f3f872ce-cbbe-4407-b4e0-3341bb598825] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 576.112085] env[61923]: ERROR nova.compute.manager [instance: f3f872ce-cbbe-4407-b4e0-3341bb598825] raise exception.PortBindingFailed(port_id=port['id']) [ 576.112536] env[61923]: ERROR nova.compute.manager [instance: f3f872ce-cbbe-4407-b4e0-3341bb598825] nova.exception.PortBindingFailed: Binding failed for port 793a91d7-2744-40de-8e41-428eacfdaee8, please check neutron logs for more information. [ 576.112536] env[61923]: ERROR nova.compute.manager [instance: f3f872ce-cbbe-4407-b4e0-3341bb598825] [ 576.112536] env[61923]: INFO nova.compute.manager [None req-36da9865-5d98-4c2a-a964-964bee74523c tempest-ListServerFiltersTestJSON-483030130 tempest-ListServerFiltersTestJSON-483030130-project-member] [instance: f3f872ce-cbbe-4407-b4e0-3341bb598825] Terminating instance [ 576.114648] env[61923]: DEBUG oslo_concurrency.lockutils [None req-36da9865-5d98-4c2a-a964-964bee74523c tempest-ListServerFiltersTestJSON-483030130 tempest-ListServerFiltersTestJSON-483030130-project-member] Acquiring lock "refresh_cache-f3f872ce-cbbe-4407-b4e0-3341bb598825" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 576.114809] env[61923]: DEBUG oslo_concurrency.lockutils [None req-36da9865-5d98-4c2a-a964-964bee74523c tempest-ListServerFiltersTestJSON-483030130 tempest-ListServerFiltersTestJSON-483030130-project-member] Acquired lock "refresh_cache-f3f872ce-cbbe-4407-b4e0-3341bb598825" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 576.114971] env[61923]: DEBUG nova.network.neutron [None req-36da9865-5d98-4c2a-a964-964bee74523c tempest-ListServerFiltersTestJSON-483030130 tempest-ListServerFiltersTestJSON-483030130-project-member] [instance: f3f872ce-cbbe-4407-b4e0-3341bb598825] Building network info cache for instance {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 576.328149] env[61923]: INFO nova.scheduler.client.report [None req-dcb477a6-85d7-4838-be85-b9b2ecf96877 tempest-ImagesOneServerNegativeTestJSON-1560548781 tempest-ImagesOneServerNegativeTestJSON-1560548781-project-member] Deleted allocations for instance ca62f959-43b9-4838-8349-11b7d6a7f0a7 [ 576.357020] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5ab550d-0845-49ae-8080-238d854ee3b1 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 576.364534] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2467038f-e1e7-4b55-b591-771d18e7ba1a {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 576.397927] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e24d6bfc-e071-4d2b-a61f-06dde0e6aa49 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 576.400294] env[61923]: DEBUG nova.compute.manager [None req-fba27a13-474d-4fe3-9187-c31143f40aa5 tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] [instance: 6a2d7ede-e0cc-4a36-b1b2-0b36b269ecb5] Start building block device mappings for instance. {{(pid=61923) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 576.407942] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a6785bc-2569-40f2-be70-6396deb4325e {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 576.430718] env[61923]: DEBUG nova.compute.provider_tree [None req-d9e2f878-40e6-4b41-8b07-da611f2f81f1 tempest-VolumesAssistedSnapshotsTest-2043366068 tempest-VolumesAssistedSnapshotsTest-2043366068-project-member] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 576.576049] env[61923]: DEBUG oslo_concurrency.lockutils [None req-1f24f04c-cf4b-4940-9447-eacc11fa5e66 tempest-ServersTestFqdnHostnames-345291565 tempest-ServersTestFqdnHostnames-345291565-project-member] Acquiring lock "c94e8e46-7697-426c-ae2f-aece493fa8f4" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 576.576471] env[61923]: DEBUG oslo_concurrency.lockutils [None req-1f24f04c-cf4b-4940-9447-eacc11fa5e66 tempest-ServersTestFqdnHostnames-345291565 tempest-ServersTestFqdnHostnames-345291565-project-member] Lock "c94e8e46-7697-426c-ae2f-aece493fa8f4" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 576.598263] env[61923]: DEBUG oslo_vmware.api [None req-ae7ff058-a25b-42a3-af4b-0715677fedd4 tempest-ServersAdmin275Test-1076075580 tempest-ServersAdmin275Test-1076075580-project-admin] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52f58043-f83c-102a-d6e5-5af7f41a6a7d, 'name': SearchDatastore_Task, 'duration_secs': 0.009758} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 576.598574] env[61923]: DEBUG oslo_concurrency.lockutils [None req-ae7ff058-a25b-42a3-af4b-0715677fedd4 tempest-ServersAdmin275Test-1076075580 tempest-ServersAdmin275Test-1076075580-project-admin] Releasing lock "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 576.599167] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-ae7ff058-a25b-42a3-af4b-0715677fedd4 tempest-ServersAdmin275Test-1076075580 tempest-ServersAdmin275Test-1076075580-project-admin] [instance: f7cc960c-e06a-4c58-9367-ec8771fe09d5] Processing image 0f153f63-ae0a-45b1-b7f5-7f9b673c947f {{(pid=61923) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 576.599167] env[61923]: DEBUG oslo_concurrency.lockutils [None req-ae7ff058-a25b-42a3-af4b-0715677fedd4 tempest-ServersAdmin275Test-1076075580 tempest-ServersAdmin275Test-1076075580-project-admin] Acquiring lock "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 576.599253] env[61923]: DEBUG oslo_concurrency.lockutils [None req-ae7ff058-a25b-42a3-af4b-0715677fedd4 tempest-ServersAdmin275Test-1076075580 tempest-ServersAdmin275Test-1076075580-project-admin] Acquired lock "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 576.599355] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-ae7ff058-a25b-42a3-af4b-0715677fedd4 tempest-ServersAdmin275Test-1076075580 tempest-ServersAdmin275Test-1076075580-project-admin] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61923) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 576.599614] env[61923]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-3c3c0053-e89b-4249-aee4-105d0e883340 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 576.612218] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-ae7ff058-a25b-42a3-af4b-0715677fedd4 tempest-ServersAdmin275Test-1076075580 tempest-ServersAdmin275Test-1076075580-project-admin] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61923) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 576.612411] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-ae7ff058-a25b-42a3-af4b-0715677fedd4 tempest-ServersAdmin275Test-1076075580 tempest-ServersAdmin275Test-1076075580-project-admin] Folder [datastore1] devstack-image-cache_base created. {{(pid=61923) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 576.613193] env[61923]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a75ef498-f149-4015-ab87-9b33a366ac11 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 576.620496] env[61923]: DEBUG oslo_vmware.api [None req-ae7ff058-a25b-42a3-af4b-0715677fedd4 tempest-ServersAdmin275Test-1076075580 tempest-ServersAdmin275Test-1076075580-project-admin] Waiting for the task: (returnval){ [ 576.620496] env[61923]: value = "session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52c17e01-9bd7-bd1d-9002-8a7621ccfd79" [ 576.620496] env[61923]: _type = "Task" [ 576.620496] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 576.632125] env[61923]: DEBUG oslo_vmware.api [None req-ae7ff058-a25b-42a3-af4b-0715677fedd4 tempest-ServersAdmin275Test-1076075580 tempest-ServersAdmin275Test-1076075580-project-admin] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52c17e01-9bd7-bd1d-9002-8a7621ccfd79, 'name': SearchDatastore_Task, 'duration_secs': 0.008119} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 576.632954] env[61923]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5f34d072-14e1-4190-86df-88d3d7339307 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 576.639548] env[61923]: DEBUG oslo_vmware.api [None req-ae7ff058-a25b-42a3-af4b-0715677fedd4 tempest-ServersAdmin275Test-1076075580 tempest-ServersAdmin275Test-1076075580-project-admin] Waiting for the task: (returnval){ [ 576.639548] env[61923]: value = "session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52d63292-39f4-3c53-319a-a912248557ba" [ 576.639548] env[61923]: _type = "Task" [ 576.639548] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 576.644291] env[61923]: DEBUG nova.network.neutron [None req-36da9865-5d98-4c2a-a964-964bee74523c tempest-ListServerFiltersTestJSON-483030130 tempest-ListServerFiltersTestJSON-483030130-project-member] [instance: f3f872ce-cbbe-4407-b4e0-3341bb598825] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 576.650254] env[61923]: DEBUG oslo_vmware.api [None req-ae7ff058-a25b-42a3-af4b-0715677fedd4 tempest-ServersAdmin275Test-1076075580 tempest-ServersAdmin275Test-1076075580-project-admin] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52d63292-39f4-3c53-319a-a912248557ba, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 576.786232] env[61923]: DEBUG nova.network.neutron [None req-36da9865-5d98-4c2a-a964-964bee74523c tempest-ListServerFiltersTestJSON-483030130 tempest-ListServerFiltersTestJSON-483030130-project-member] [instance: f3f872ce-cbbe-4407-b4e0-3341bb598825] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 576.820177] env[61923]: DEBUG nova.compute.manager [req-7843786b-2b0d-499e-80ff-bfe9e36a190b req-6bd788bb-868d-4aca-8fcf-67f7bb92c5bb service nova] [instance: f3f872ce-cbbe-4407-b4e0-3341bb598825] Received event network-changed-793a91d7-2744-40de-8e41-428eacfdaee8 {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 576.820177] env[61923]: DEBUG nova.compute.manager [req-7843786b-2b0d-499e-80ff-bfe9e36a190b req-6bd788bb-868d-4aca-8fcf-67f7bb92c5bb service nova] [instance: f3f872ce-cbbe-4407-b4e0-3341bb598825] Refreshing instance network info cache due to event network-changed-793a91d7-2744-40de-8e41-428eacfdaee8. {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 576.820177] env[61923]: DEBUG oslo_concurrency.lockutils [req-7843786b-2b0d-499e-80ff-bfe9e36a190b req-6bd788bb-868d-4aca-8fcf-67f7bb92c5bb service nova] Acquiring lock "refresh_cache-f3f872ce-cbbe-4407-b4e0-3341bb598825" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 576.823617] env[61923]: DEBUG nova.network.neutron [None req-fba27a13-474d-4fe3-9187-c31143f40aa5 tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] [instance: 6a2d7ede-e0cc-4a36-b1b2-0b36b269ecb5] Successfully created port: 3b3effc4-2768-46ad-8e11-d2719c57d383 {{(pid=61923) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 576.841758] env[61923]: DEBUG oslo_concurrency.lockutils [None req-dcb477a6-85d7-4838-be85-b9b2ecf96877 tempest-ImagesOneServerNegativeTestJSON-1560548781 tempest-ImagesOneServerNegativeTestJSON-1560548781-project-member] Lock "ca62f959-43b9-4838-8349-11b7d6a7f0a7" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 51.036s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 576.863662] env[61923]: DEBUG oslo_concurrency.lockutils [None req-aadda1bb-f770-4742-8c34-cd25e3f7e7cb tempest-InstanceActionsNegativeTestJSON-1700332362 tempest-InstanceActionsNegativeTestJSON-1700332362-project-member] Acquiring lock "64584976-b3f3-4da5-a76b-a05cf2ed6aa4" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 576.864025] env[61923]: DEBUG oslo_concurrency.lockutils [None req-aadda1bb-f770-4742-8c34-cd25e3f7e7cb tempest-InstanceActionsNegativeTestJSON-1700332362 tempest-InstanceActionsNegativeTestJSON-1700332362-project-member] Lock "64584976-b3f3-4da5-a76b-a05cf2ed6aa4" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 576.934767] env[61923]: DEBUG nova.scheduler.client.report [None req-d9e2f878-40e6-4b41-8b07-da611f2f81f1 tempest-VolumesAssistedSnapshotsTest-2043366068 tempest-VolumesAssistedSnapshotsTest-2043366068-project-member] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 577.149840] env[61923]: DEBUG oslo_vmware.api [None req-ae7ff058-a25b-42a3-af4b-0715677fedd4 tempest-ServersAdmin275Test-1076075580 tempest-ServersAdmin275Test-1076075580-project-admin] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52d63292-39f4-3c53-319a-a912248557ba, 'name': SearchDatastore_Task, 'duration_secs': 0.008685} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 577.150119] env[61923]: DEBUG oslo_concurrency.lockutils [None req-ae7ff058-a25b-42a3-af4b-0715677fedd4 tempest-ServersAdmin275Test-1076075580 tempest-ServersAdmin275Test-1076075580-project-admin] Releasing lock "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 577.150428] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-ae7ff058-a25b-42a3-af4b-0715677fedd4 tempest-ServersAdmin275Test-1076075580 tempest-ServersAdmin275Test-1076075580-project-admin] Copying Virtual Disk [datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk to [datastore1] f7cc960c-e06a-4c58-9367-ec8771fe09d5/f7cc960c-e06a-4c58-9367-ec8771fe09d5.vmdk {{(pid=61923) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 577.150688] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-9402756f-0519-4076-9f17-2d246ca46b44 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 577.157249] env[61923]: DEBUG oslo_vmware.api [None req-ae7ff058-a25b-42a3-af4b-0715677fedd4 tempest-ServersAdmin275Test-1076075580 tempest-ServersAdmin275Test-1076075580-project-admin] Waiting for the task: (returnval){ [ 577.157249] env[61923]: value = "task-1377467" [ 577.157249] env[61923]: _type = "Task" [ 577.157249] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 577.164673] env[61923]: DEBUG oslo_vmware.api [None req-ae7ff058-a25b-42a3-af4b-0715677fedd4 tempest-ServersAdmin275Test-1076075580 tempest-ServersAdmin275Test-1076075580-project-admin] Task: {'id': task-1377467, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 577.289886] env[61923]: DEBUG oslo_concurrency.lockutils [None req-36da9865-5d98-4c2a-a964-964bee74523c tempest-ListServerFiltersTestJSON-483030130 tempest-ListServerFiltersTestJSON-483030130-project-member] Releasing lock "refresh_cache-f3f872ce-cbbe-4407-b4e0-3341bb598825" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 577.290353] env[61923]: DEBUG nova.compute.manager [None req-36da9865-5d98-4c2a-a964-964bee74523c tempest-ListServerFiltersTestJSON-483030130 tempest-ListServerFiltersTestJSON-483030130-project-member] [instance: f3f872ce-cbbe-4407-b4e0-3341bb598825] Start destroying the instance on the hypervisor. {{(pid=61923) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 577.290550] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-36da9865-5d98-4c2a-a964-964bee74523c tempest-ListServerFiltersTestJSON-483030130 tempest-ListServerFiltersTestJSON-483030130-project-member] [instance: f3f872ce-cbbe-4407-b4e0-3341bb598825] Destroying instance {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 577.290862] env[61923]: DEBUG oslo_concurrency.lockutils [req-7843786b-2b0d-499e-80ff-bfe9e36a190b req-6bd788bb-868d-4aca-8fcf-67f7bb92c5bb service nova] Acquired lock "refresh_cache-f3f872ce-cbbe-4407-b4e0-3341bb598825" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 577.291045] env[61923]: DEBUG nova.network.neutron [req-7843786b-2b0d-499e-80ff-bfe9e36a190b req-6bd788bb-868d-4aca-8fcf-67f7bb92c5bb service nova] [instance: f3f872ce-cbbe-4407-b4e0-3341bb598825] Refreshing network info cache for port 793a91d7-2744-40de-8e41-428eacfdaee8 {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 577.292126] env[61923]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-0c3ad6ed-5991-49f8-91cc-2e8a4b8c780d {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 577.300962] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4c31bf0-417b-4d32-9703-0f78c67d2dc3 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 577.329377] env[61923]: WARNING nova.virt.vmwareapi.vmops [None req-36da9865-5d98-4c2a-a964-964bee74523c tempest-ListServerFiltersTestJSON-483030130 tempest-ListServerFiltersTestJSON-483030130-project-member] [instance: f3f872ce-cbbe-4407-b4e0-3341bb598825] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance f3f872ce-cbbe-4407-b4e0-3341bb598825 could not be found. [ 577.329619] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-36da9865-5d98-4c2a-a964-964bee74523c tempest-ListServerFiltersTestJSON-483030130 tempest-ListServerFiltersTestJSON-483030130-project-member] [instance: f3f872ce-cbbe-4407-b4e0-3341bb598825] Instance destroyed {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 577.329798] env[61923]: INFO nova.compute.manager [None req-36da9865-5d98-4c2a-a964-964bee74523c tempest-ListServerFiltersTestJSON-483030130 tempest-ListServerFiltersTestJSON-483030130-project-member] [instance: f3f872ce-cbbe-4407-b4e0-3341bb598825] Took 0.04 seconds to destroy the instance on the hypervisor. [ 577.330052] env[61923]: DEBUG oslo.service.loopingcall [None req-36da9865-5d98-4c2a-a964-964bee74523c tempest-ListServerFiltersTestJSON-483030130 tempest-ListServerFiltersTestJSON-483030130-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61923) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 577.330300] env[61923]: DEBUG nova.compute.manager [-] [instance: f3f872ce-cbbe-4407-b4e0-3341bb598825] Deallocating network for instance {{(pid=61923) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 577.330393] env[61923]: DEBUG nova.network.neutron [-] [instance: f3f872ce-cbbe-4407-b4e0-3341bb598825] deallocate_for_instance() {{(pid=61923) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 577.344984] env[61923]: DEBUG nova.compute.manager [None req-03cb7270-244b-47b9-9dd5-e358ebdfa946 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: cb998da7-8db0-4725-bed2-657d7df748b5] Starting instance... {{(pid=61923) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 577.349310] env[61923]: DEBUG nova.network.neutron [-] [instance: f3f872ce-cbbe-4407-b4e0-3341bb598825] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 577.416252] env[61923]: DEBUG nova.compute.manager [None req-fba27a13-474d-4fe3-9187-c31143f40aa5 tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] [instance: 6a2d7ede-e0cc-4a36-b1b2-0b36b269ecb5] Start spawning the instance on the hypervisor. {{(pid=61923) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 577.443586] env[61923]: DEBUG nova.virt.hardware [None req-fba27a13-474d-4fe3-9187-c31143f40aa5 tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-29T20:07:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-29T20:07:35Z,direct_url=,disk_format='vmdk',id=0f153f63-ae0a-45b1-b7f5-7f9b673c947f,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='4e7b5e3b3c3443c8bd5c70f8a15739f5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-29T20:07:36Z,virtual_size=,visibility=), allow threads: False {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 577.443928] env[61923]: DEBUG nova.virt.hardware [None req-fba27a13-474d-4fe3-9187-c31143f40aa5 tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] Flavor limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 577.443995] env[61923]: DEBUG nova.virt.hardware [None req-fba27a13-474d-4fe3-9187-c31143f40aa5 tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] Image limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 577.444151] env[61923]: DEBUG nova.virt.hardware [None req-fba27a13-474d-4fe3-9187-c31143f40aa5 tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] Flavor pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 577.444398] env[61923]: DEBUG nova.virt.hardware [None req-fba27a13-474d-4fe3-9187-c31143f40aa5 tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] Image pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 577.444496] env[61923]: DEBUG nova.virt.hardware [None req-fba27a13-474d-4fe3-9187-c31143f40aa5 tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 577.444647] env[61923]: DEBUG nova.virt.hardware [None req-fba27a13-474d-4fe3-9187-c31143f40aa5 tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 577.445301] env[61923]: DEBUG nova.virt.hardware [None req-fba27a13-474d-4fe3-9187-c31143f40aa5 tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 577.445301] env[61923]: DEBUG nova.virt.hardware [None req-fba27a13-474d-4fe3-9187-c31143f40aa5 tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] Got 1 possible topologies {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 577.445301] env[61923]: DEBUG nova.virt.hardware [None req-fba27a13-474d-4fe3-9187-c31143f40aa5 tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 577.445450] env[61923]: DEBUG nova.virt.hardware [None req-fba27a13-474d-4fe3-9187-c31143f40aa5 tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 577.446547] env[61923]: DEBUG oslo_concurrency.lockutils [None req-d9e2f878-40e6-4b41-8b07-da611f2f81f1 tempest-VolumesAssistedSnapshotsTest-2043366068 tempest-VolumesAssistedSnapshotsTest-2043366068-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.057s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 577.446855] env[61923]: ERROR nova.compute.manager [None req-d9e2f878-40e6-4b41-8b07-da611f2f81f1 tempest-VolumesAssistedSnapshotsTest-2043366068 tempest-VolumesAssistedSnapshotsTest-2043366068-project-member] [instance: 06bda5ee-c7a8-47d1-9e0c-a5af67d40006] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 48788634-af14-4531-8cff-07f9a6f55f4c, please check neutron logs for more information. [ 577.446855] env[61923]: ERROR nova.compute.manager [instance: 06bda5ee-c7a8-47d1-9e0c-a5af67d40006] Traceback (most recent call last): [ 577.446855] env[61923]: ERROR nova.compute.manager [instance: 06bda5ee-c7a8-47d1-9e0c-a5af67d40006] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 577.446855] env[61923]: ERROR nova.compute.manager [instance: 06bda5ee-c7a8-47d1-9e0c-a5af67d40006] self.driver.spawn(context, instance, image_meta, [ 577.446855] env[61923]: ERROR nova.compute.manager [instance: 06bda5ee-c7a8-47d1-9e0c-a5af67d40006] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 577.446855] env[61923]: ERROR nova.compute.manager [instance: 06bda5ee-c7a8-47d1-9e0c-a5af67d40006] self._vmops.spawn(context, instance, image_meta, injected_files, [ 577.446855] env[61923]: ERROR nova.compute.manager [instance: 06bda5ee-c7a8-47d1-9e0c-a5af67d40006] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 577.446855] env[61923]: ERROR nova.compute.manager [instance: 06bda5ee-c7a8-47d1-9e0c-a5af67d40006] vm_ref = self.build_virtual_machine(instance, [ 577.446855] env[61923]: ERROR nova.compute.manager [instance: 06bda5ee-c7a8-47d1-9e0c-a5af67d40006] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 577.446855] env[61923]: ERROR nova.compute.manager [instance: 06bda5ee-c7a8-47d1-9e0c-a5af67d40006] vif_infos = vmwarevif.get_vif_info(self._session, [ 577.446855] env[61923]: ERROR nova.compute.manager [instance: 06bda5ee-c7a8-47d1-9e0c-a5af67d40006] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 577.447407] env[61923]: ERROR nova.compute.manager [instance: 06bda5ee-c7a8-47d1-9e0c-a5af67d40006] for vif in network_info: [ 577.447407] env[61923]: ERROR nova.compute.manager [instance: 06bda5ee-c7a8-47d1-9e0c-a5af67d40006] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 577.447407] env[61923]: ERROR nova.compute.manager [instance: 06bda5ee-c7a8-47d1-9e0c-a5af67d40006] return self._sync_wrapper(fn, *args, **kwargs) [ 577.447407] env[61923]: ERROR nova.compute.manager [instance: 06bda5ee-c7a8-47d1-9e0c-a5af67d40006] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 577.447407] env[61923]: ERROR nova.compute.manager [instance: 06bda5ee-c7a8-47d1-9e0c-a5af67d40006] self.wait() [ 577.447407] env[61923]: ERROR nova.compute.manager [instance: 06bda5ee-c7a8-47d1-9e0c-a5af67d40006] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 577.447407] env[61923]: ERROR nova.compute.manager [instance: 06bda5ee-c7a8-47d1-9e0c-a5af67d40006] self[:] = self._gt.wait() [ 577.447407] env[61923]: ERROR nova.compute.manager [instance: 06bda5ee-c7a8-47d1-9e0c-a5af67d40006] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 577.447407] env[61923]: ERROR nova.compute.manager [instance: 06bda5ee-c7a8-47d1-9e0c-a5af67d40006] return self._exit_event.wait() [ 577.447407] env[61923]: ERROR nova.compute.manager [instance: 06bda5ee-c7a8-47d1-9e0c-a5af67d40006] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 577.447407] env[61923]: ERROR nova.compute.manager [instance: 06bda5ee-c7a8-47d1-9e0c-a5af67d40006] result = hub.switch() [ 577.447407] env[61923]: ERROR nova.compute.manager [instance: 06bda5ee-c7a8-47d1-9e0c-a5af67d40006] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 577.447407] env[61923]: ERROR nova.compute.manager [instance: 06bda5ee-c7a8-47d1-9e0c-a5af67d40006] return self.greenlet.switch() [ 577.447899] env[61923]: ERROR nova.compute.manager [instance: 06bda5ee-c7a8-47d1-9e0c-a5af67d40006] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 577.447899] env[61923]: ERROR nova.compute.manager [instance: 06bda5ee-c7a8-47d1-9e0c-a5af67d40006] result = function(*args, **kwargs) [ 577.447899] env[61923]: ERROR nova.compute.manager [instance: 06bda5ee-c7a8-47d1-9e0c-a5af67d40006] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 577.447899] env[61923]: ERROR nova.compute.manager [instance: 06bda5ee-c7a8-47d1-9e0c-a5af67d40006] return func(*args, **kwargs) [ 577.447899] env[61923]: ERROR nova.compute.manager [instance: 06bda5ee-c7a8-47d1-9e0c-a5af67d40006] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 577.447899] env[61923]: ERROR nova.compute.manager [instance: 06bda5ee-c7a8-47d1-9e0c-a5af67d40006] raise e [ 577.447899] env[61923]: ERROR nova.compute.manager [instance: 06bda5ee-c7a8-47d1-9e0c-a5af67d40006] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 577.447899] env[61923]: ERROR nova.compute.manager [instance: 06bda5ee-c7a8-47d1-9e0c-a5af67d40006] nwinfo = self.network_api.allocate_for_instance( [ 577.447899] env[61923]: ERROR nova.compute.manager [instance: 06bda5ee-c7a8-47d1-9e0c-a5af67d40006] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 577.447899] env[61923]: ERROR nova.compute.manager [instance: 06bda5ee-c7a8-47d1-9e0c-a5af67d40006] created_port_ids = self._update_ports_for_instance( [ 577.447899] env[61923]: ERROR nova.compute.manager [instance: 06bda5ee-c7a8-47d1-9e0c-a5af67d40006] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 577.447899] env[61923]: ERROR nova.compute.manager [instance: 06bda5ee-c7a8-47d1-9e0c-a5af67d40006] with excutils.save_and_reraise_exception(): [ 577.447899] env[61923]: ERROR nova.compute.manager [instance: 06bda5ee-c7a8-47d1-9e0c-a5af67d40006] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 577.448341] env[61923]: ERROR nova.compute.manager [instance: 06bda5ee-c7a8-47d1-9e0c-a5af67d40006] self.force_reraise() [ 577.448341] env[61923]: ERROR nova.compute.manager [instance: 06bda5ee-c7a8-47d1-9e0c-a5af67d40006] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 577.448341] env[61923]: ERROR nova.compute.manager [instance: 06bda5ee-c7a8-47d1-9e0c-a5af67d40006] raise self.value [ 577.448341] env[61923]: ERROR nova.compute.manager [instance: 06bda5ee-c7a8-47d1-9e0c-a5af67d40006] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 577.448341] env[61923]: ERROR nova.compute.manager [instance: 06bda5ee-c7a8-47d1-9e0c-a5af67d40006] updated_port = self._update_port( [ 577.448341] env[61923]: ERROR nova.compute.manager [instance: 06bda5ee-c7a8-47d1-9e0c-a5af67d40006] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 577.448341] env[61923]: ERROR nova.compute.manager [instance: 06bda5ee-c7a8-47d1-9e0c-a5af67d40006] _ensure_no_port_binding_failure(port) [ 577.448341] env[61923]: ERROR nova.compute.manager [instance: 06bda5ee-c7a8-47d1-9e0c-a5af67d40006] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 577.448341] env[61923]: ERROR nova.compute.manager [instance: 06bda5ee-c7a8-47d1-9e0c-a5af67d40006] raise exception.PortBindingFailed(port_id=port['id']) [ 577.448341] env[61923]: ERROR nova.compute.manager [instance: 06bda5ee-c7a8-47d1-9e0c-a5af67d40006] nova.exception.PortBindingFailed: Binding failed for port 48788634-af14-4531-8cff-07f9a6f55f4c, please check neutron logs for more information. [ 577.448341] env[61923]: ERROR nova.compute.manager [instance: 06bda5ee-c7a8-47d1-9e0c-a5af67d40006] [ 577.448810] env[61923]: DEBUG nova.compute.utils [None req-d9e2f878-40e6-4b41-8b07-da611f2f81f1 tempest-VolumesAssistedSnapshotsTest-2043366068 tempest-VolumesAssistedSnapshotsTest-2043366068-project-member] [instance: 06bda5ee-c7a8-47d1-9e0c-a5af67d40006] Binding failed for port 48788634-af14-4531-8cff-07f9a6f55f4c, please check neutron logs for more information. {{(pid=61923) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 577.449387] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93f9d3da-982e-4903-be6f-a4209d0ac2de {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 577.452689] env[61923]: DEBUG nova.compute.manager [None req-d9e2f878-40e6-4b41-8b07-da611f2f81f1 tempest-VolumesAssistedSnapshotsTest-2043366068 tempest-VolumesAssistedSnapshotsTest-2043366068-project-member] [instance: 06bda5ee-c7a8-47d1-9e0c-a5af67d40006] Build of instance 06bda5ee-c7a8-47d1-9e0c-a5af67d40006 was re-scheduled: Binding failed for port 48788634-af14-4531-8cff-07f9a6f55f4c, please check neutron logs for more information. {{(pid=61923) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 577.453098] env[61923]: DEBUG nova.compute.manager [None req-d9e2f878-40e6-4b41-8b07-da611f2f81f1 tempest-VolumesAssistedSnapshotsTest-2043366068 tempest-VolumesAssistedSnapshotsTest-2043366068-project-member] [instance: 06bda5ee-c7a8-47d1-9e0c-a5af67d40006] Unplugging VIFs for instance {{(pid=61923) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 577.453293] env[61923]: DEBUG oslo_concurrency.lockutils [None req-d9e2f878-40e6-4b41-8b07-da611f2f81f1 tempest-VolumesAssistedSnapshotsTest-2043366068 tempest-VolumesAssistedSnapshotsTest-2043366068-project-member] Acquiring lock "refresh_cache-06bda5ee-c7a8-47d1-9e0c-a5af67d40006" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 577.453494] env[61923]: DEBUG oslo_concurrency.lockutils [None req-d9e2f878-40e6-4b41-8b07-da611f2f81f1 tempest-VolumesAssistedSnapshotsTest-2043366068 tempest-VolumesAssistedSnapshotsTest-2043366068-project-member] Acquired lock "refresh_cache-06bda5ee-c7a8-47d1-9e0c-a5af67d40006" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 577.454136] env[61923]: DEBUG nova.network.neutron [None req-d9e2f878-40e6-4b41-8b07-da611f2f81f1 tempest-VolumesAssistedSnapshotsTest-2043366068 tempest-VolumesAssistedSnapshotsTest-2043366068-project-member] [instance: 06bda5ee-c7a8-47d1-9e0c-a5af67d40006] Building network info cache for instance {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 577.454738] env[61923]: DEBUG oslo_concurrency.lockutils [None req-fba27a13-474d-4fe3-9187-c31143f40aa5 tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 23.857s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 577.457917] env[61923]: INFO nova.compute.claims [None req-fba27a13-474d-4fe3-9187-c31143f40aa5 tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] [instance: 9c3e2f94-91da-4255-b4d6-1c01e100ff48] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 577.466829] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fea8f471-b575-40f7-a906-ff8022bd7a09 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 577.667929] env[61923]: DEBUG oslo_vmware.api [None req-ae7ff058-a25b-42a3-af4b-0715677fedd4 tempest-ServersAdmin275Test-1076075580 tempest-ServersAdmin275Test-1076075580-project-admin] Task: {'id': task-1377467, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 577.830189] env[61923]: DEBUG nova.network.neutron [req-7843786b-2b0d-499e-80ff-bfe9e36a190b req-6bd788bb-868d-4aca-8fcf-67f7bb92c5bb service nova] [instance: f3f872ce-cbbe-4407-b4e0-3341bb598825] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 577.855408] env[61923]: DEBUG nova.network.neutron [-] [instance: f3f872ce-cbbe-4407-b4e0-3341bb598825] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 577.889826] env[61923]: DEBUG oslo_concurrency.lockutils [None req-03cb7270-244b-47b9-9dd5-e358ebdfa946 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 577.945313] env[61923]: DEBUG nova.network.neutron [req-7843786b-2b0d-499e-80ff-bfe9e36a190b req-6bd788bb-868d-4aca-8fcf-67f7bb92c5bb service nova] [instance: f3f872ce-cbbe-4407-b4e0-3341bb598825] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 577.992992] env[61923]: DEBUG nova.network.neutron [None req-d9e2f878-40e6-4b41-8b07-da611f2f81f1 tempest-VolumesAssistedSnapshotsTest-2043366068 tempest-VolumesAssistedSnapshotsTest-2043366068-project-member] [instance: 06bda5ee-c7a8-47d1-9e0c-a5af67d40006] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 578.087520] env[61923]: DEBUG nova.network.neutron [None req-d9e2f878-40e6-4b41-8b07-da611f2f81f1 tempest-VolumesAssistedSnapshotsTest-2043366068 tempest-VolumesAssistedSnapshotsTest-2043366068-project-member] [instance: 06bda5ee-c7a8-47d1-9e0c-a5af67d40006] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 578.173829] env[61923]: DEBUG oslo_vmware.api [None req-ae7ff058-a25b-42a3-af4b-0715677fedd4 tempest-ServersAdmin275Test-1076075580 tempest-ServersAdmin275Test-1076075580-project-admin] Task: {'id': task-1377467, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.971184} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 578.174128] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-ae7ff058-a25b-42a3-af4b-0715677fedd4 tempest-ServersAdmin275Test-1076075580 tempest-ServersAdmin275Test-1076075580-project-admin] Copied Virtual Disk [datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk to [datastore1] f7cc960c-e06a-4c58-9367-ec8771fe09d5/f7cc960c-e06a-4c58-9367-ec8771fe09d5.vmdk {{(pid=61923) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 578.174352] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-ae7ff058-a25b-42a3-af4b-0715677fedd4 tempest-ServersAdmin275Test-1076075580 tempest-ServersAdmin275Test-1076075580-project-admin] [instance: f7cc960c-e06a-4c58-9367-ec8771fe09d5] Extending root virtual disk to 1048576 {{(pid=61923) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 578.174610] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-21476743-bf8f-4ef1-bb35-d53974663b05 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 578.183615] env[61923]: DEBUG oslo_vmware.api [None req-ae7ff058-a25b-42a3-af4b-0715677fedd4 tempest-ServersAdmin275Test-1076075580 tempest-ServersAdmin275Test-1076075580-project-admin] Waiting for the task: (returnval){ [ 578.183615] env[61923]: value = "task-1377468" [ 578.183615] env[61923]: _type = "Task" [ 578.183615] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 578.192731] env[61923]: DEBUG oslo_vmware.api [None req-ae7ff058-a25b-42a3-af4b-0715677fedd4 tempest-ServersAdmin275Test-1076075580 tempest-ServersAdmin275Test-1076075580-project-admin] Task: {'id': task-1377468, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 578.215912] env[61923]: DEBUG oslo_concurrency.lockutils [None req-5f08bfba-63c5-4161-be1b-a884c0ae71af tempest-VolumesAdminNegativeTest-1422120048 tempest-VolumesAdminNegativeTest-1422120048-project-member] Acquiring lock "590907a7-b6cc-48fc-a4bf-e4a2e48b05b8" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 578.215912] env[61923]: DEBUG oslo_concurrency.lockutils [None req-5f08bfba-63c5-4161-be1b-a884c0ae71af tempest-VolumesAdminNegativeTest-1422120048 tempest-VolumesAdminNegativeTest-1422120048-project-member] Lock "590907a7-b6cc-48fc-a4bf-e4a2e48b05b8" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 578.361780] env[61923]: INFO nova.compute.manager [-] [instance: f3f872ce-cbbe-4407-b4e0-3341bb598825] Took 1.03 seconds to deallocate network for instance. [ 578.364421] env[61923]: DEBUG nova.compute.claims [None req-36da9865-5d98-4c2a-a964-964bee74523c tempest-ListServerFiltersTestJSON-483030130 tempest-ListServerFiltersTestJSON-483030130-project-member] [instance: f3f872ce-cbbe-4407-b4e0-3341bb598825] Aborting claim: {{(pid=61923) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 578.364651] env[61923]: DEBUG oslo_concurrency.lockutils [None req-36da9865-5d98-4c2a-a964-964bee74523c tempest-ListServerFiltersTestJSON-483030130 tempest-ListServerFiltersTestJSON-483030130-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 578.448841] env[61923]: DEBUG oslo_concurrency.lockutils [req-7843786b-2b0d-499e-80ff-bfe9e36a190b req-6bd788bb-868d-4aca-8fcf-67f7bb92c5bb service nova] Releasing lock "refresh_cache-f3f872ce-cbbe-4407-b4e0-3341bb598825" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 578.592277] env[61923]: DEBUG oslo_concurrency.lockutils [None req-d9e2f878-40e6-4b41-8b07-da611f2f81f1 tempest-VolumesAssistedSnapshotsTest-2043366068 tempest-VolumesAssistedSnapshotsTest-2043366068-project-member] Releasing lock "refresh_cache-06bda5ee-c7a8-47d1-9e0c-a5af67d40006" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 578.592277] env[61923]: DEBUG nova.compute.manager [None req-d9e2f878-40e6-4b41-8b07-da611f2f81f1 tempest-VolumesAssistedSnapshotsTest-2043366068 tempest-VolumesAssistedSnapshotsTest-2043366068-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61923) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 578.592501] env[61923]: DEBUG nova.compute.manager [None req-d9e2f878-40e6-4b41-8b07-da611f2f81f1 tempest-VolumesAssistedSnapshotsTest-2043366068 tempest-VolumesAssistedSnapshotsTest-2043366068-project-member] [instance: 06bda5ee-c7a8-47d1-9e0c-a5af67d40006] Deallocating network for instance {{(pid=61923) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 578.592563] env[61923]: DEBUG nova.network.neutron [None req-d9e2f878-40e6-4b41-8b07-da611f2f81f1 tempest-VolumesAssistedSnapshotsTest-2043366068 tempest-VolumesAssistedSnapshotsTest-2043366068-project-member] [instance: 06bda5ee-c7a8-47d1-9e0c-a5af67d40006] deallocate_for_instance() {{(pid=61923) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 578.620654] env[61923]: DEBUG nova.network.neutron [None req-d9e2f878-40e6-4b41-8b07-da611f2f81f1 tempest-VolumesAssistedSnapshotsTest-2043366068 tempest-VolumesAssistedSnapshotsTest-2043366068-project-member] [instance: 06bda5ee-c7a8-47d1-9e0c-a5af67d40006] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 578.698018] env[61923]: DEBUG oslo_vmware.api [None req-ae7ff058-a25b-42a3-af4b-0715677fedd4 tempest-ServersAdmin275Test-1076075580 tempest-ServersAdmin275Test-1076075580-project-admin] Task: {'id': task-1377468, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.065022} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 578.698018] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-ae7ff058-a25b-42a3-af4b-0715677fedd4 tempest-ServersAdmin275Test-1076075580 tempest-ServersAdmin275Test-1076075580-project-admin] [instance: f7cc960c-e06a-4c58-9367-ec8771fe09d5] Extended root virtual disk {{(pid=61923) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 578.698018] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a3d37d6-bb85-49c9-85fe-e34c927e5012 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 578.720316] env[61923]: DEBUG nova.virt.vmwareapi.volumeops [None req-ae7ff058-a25b-42a3-af4b-0715677fedd4 tempest-ServersAdmin275Test-1076075580 tempest-ServersAdmin275Test-1076075580-project-admin] [instance: f7cc960c-e06a-4c58-9367-ec8771fe09d5] Reconfiguring VM instance instance-0000000a to attach disk [datastore1] f7cc960c-e06a-4c58-9367-ec8771fe09d5/f7cc960c-e06a-4c58-9367-ec8771fe09d5.vmdk or device None with type sparse {{(pid=61923) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 578.723217] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e874c065-7291-4e96-aa48-275d00791ffd {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 578.748976] env[61923]: DEBUG oslo_vmware.api [None req-ae7ff058-a25b-42a3-af4b-0715677fedd4 tempest-ServersAdmin275Test-1076075580 tempest-ServersAdmin275Test-1076075580-project-admin] Waiting for the task: (returnval){ [ 578.748976] env[61923]: value = "task-1377470" [ 578.748976] env[61923]: _type = "Task" [ 578.748976] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 578.757271] env[61923]: DEBUG oslo_vmware.api [None req-ae7ff058-a25b-42a3-af4b-0715677fedd4 tempest-ServersAdmin275Test-1076075580 tempest-ServersAdmin275Test-1076075580-project-admin] Task: {'id': task-1377470, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 578.954813] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41434fb4-08e7-4282-a80e-6344f8e3f117 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 578.963268] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4b66c27-3a94-4806-9659-21cb7bf3655c {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 578.995160] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0eaf3486-ccfc-479e-8e65-1bc5acc146e5 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 579.006814] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d87b4e0b-0f8a-45f7-821e-62f498d56347 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 579.026919] env[61923]: DEBUG nova.compute.provider_tree [None req-fba27a13-474d-4fe3-9187-c31143f40aa5 tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 579.035563] env[61923]: DEBUG oslo_concurrency.lockutils [None req-08382130-d70f-42e8-a5b5-93e86f5fa49a tempest-AttachVolumeShelveTestJSON-1252849694 tempest-AttachVolumeShelveTestJSON-1252849694-project-member] Acquiring lock "60ffc8b0-4a7b-4e2a-8774-d16e502d8a06" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 579.035863] env[61923]: DEBUG oslo_concurrency.lockutils [None req-08382130-d70f-42e8-a5b5-93e86f5fa49a tempest-AttachVolumeShelveTestJSON-1252849694 tempest-AttachVolumeShelveTestJSON-1252849694-project-member] Lock "60ffc8b0-4a7b-4e2a-8774-d16e502d8a06" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.003s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 579.076355] env[61923]: ERROR nova.compute.manager [None req-fba27a13-474d-4fe3-9187-c31143f40aa5 tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 3b3effc4-2768-46ad-8e11-d2719c57d383, please check neutron logs for more information. [ 579.076355] env[61923]: ERROR nova.compute.manager Traceback (most recent call last): [ 579.076355] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 579.076355] env[61923]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 579.076355] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 579.076355] env[61923]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 579.076355] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 579.076355] env[61923]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 579.076355] env[61923]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 579.076355] env[61923]: ERROR nova.compute.manager self.force_reraise() [ 579.076355] env[61923]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 579.076355] env[61923]: ERROR nova.compute.manager raise self.value [ 579.076355] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 579.076355] env[61923]: ERROR nova.compute.manager updated_port = self._update_port( [ 579.076355] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 579.076355] env[61923]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 579.076870] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 579.076870] env[61923]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 579.076870] env[61923]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 3b3effc4-2768-46ad-8e11-d2719c57d383, please check neutron logs for more information. [ 579.076870] env[61923]: ERROR nova.compute.manager [ 579.076870] env[61923]: Traceback (most recent call last): [ 579.076870] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 579.076870] env[61923]: listener.cb(fileno) [ 579.076870] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 579.076870] env[61923]: result = function(*args, **kwargs) [ 579.076870] env[61923]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 579.076870] env[61923]: return func(*args, **kwargs) [ 579.076870] env[61923]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 579.076870] env[61923]: raise e [ 579.076870] env[61923]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 579.076870] env[61923]: nwinfo = self.network_api.allocate_for_instance( [ 579.076870] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 579.076870] env[61923]: created_port_ids = self._update_ports_for_instance( [ 579.076870] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 579.076870] env[61923]: with excutils.save_and_reraise_exception(): [ 579.076870] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 579.076870] env[61923]: self.force_reraise() [ 579.076870] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 579.076870] env[61923]: raise self.value [ 579.076870] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 579.076870] env[61923]: updated_port = self._update_port( [ 579.076870] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 579.076870] env[61923]: _ensure_no_port_binding_failure(port) [ 579.076870] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 579.076870] env[61923]: raise exception.PortBindingFailed(port_id=port['id']) [ 579.077801] env[61923]: nova.exception.PortBindingFailed: Binding failed for port 3b3effc4-2768-46ad-8e11-d2719c57d383, please check neutron logs for more information. [ 579.077801] env[61923]: Removing descriptor: 14 [ 579.077801] env[61923]: ERROR nova.compute.manager [None req-fba27a13-474d-4fe3-9187-c31143f40aa5 tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] [instance: 6a2d7ede-e0cc-4a36-b1b2-0b36b269ecb5] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 3b3effc4-2768-46ad-8e11-d2719c57d383, please check neutron logs for more information. [ 579.077801] env[61923]: ERROR nova.compute.manager [instance: 6a2d7ede-e0cc-4a36-b1b2-0b36b269ecb5] Traceback (most recent call last): [ 579.077801] env[61923]: ERROR nova.compute.manager [instance: 6a2d7ede-e0cc-4a36-b1b2-0b36b269ecb5] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 579.077801] env[61923]: ERROR nova.compute.manager [instance: 6a2d7ede-e0cc-4a36-b1b2-0b36b269ecb5] yield resources [ 579.077801] env[61923]: ERROR nova.compute.manager [instance: 6a2d7ede-e0cc-4a36-b1b2-0b36b269ecb5] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 579.077801] env[61923]: ERROR nova.compute.manager [instance: 6a2d7ede-e0cc-4a36-b1b2-0b36b269ecb5] self.driver.spawn(context, instance, image_meta, [ 579.077801] env[61923]: ERROR nova.compute.manager [instance: 6a2d7ede-e0cc-4a36-b1b2-0b36b269ecb5] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 579.077801] env[61923]: ERROR nova.compute.manager [instance: 6a2d7ede-e0cc-4a36-b1b2-0b36b269ecb5] self._vmops.spawn(context, instance, image_meta, injected_files, [ 579.077801] env[61923]: ERROR nova.compute.manager [instance: 6a2d7ede-e0cc-4a36-b1b2-0b36b269ecb5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 579.077801] env[61923]: ERROR nova.compute.manager [instance: 6a2d7ede-e0cc-4a36-b1b2-0b36b269ecb5] vm_ref = self.build_virtual_machine(instance, [ 579.078456] env[61923]: ERROR nova.compute.manager [instance: 6a2d7ede-e0cc-4a36-b1b2-0b36b269ecb5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 579.078456] env[61923]: ERROR nova.compute.manager [instance: 6a2d7ede-e0cc-4a36-b1b2-0b36b269ecb5] vif_infos = vmwarevif.get_vif_info(self._session, [ 579.078456] env[61923]: ERROR nova.compute.manager [instance: 6a2d7ede-e0cc-4a36-b1b2-0b36b269ecb5] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 579.078456] env[61923]: ERROR nova.compute.manager [instance: 6a2d7ede-e0cc-4a36-b1b2-0b36b269ecb5] for vif in network_info: [ 579.078456] env[61923]: ERROR nova.compute.manager [instance: 6a2d7ede-e0cc-4a36-b1b2-0b36b269ecb5] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 579.078456] env[61923]: ERROR nova.compute.manager [instance: 6a2d7ede-e0cc-4a36-b1b2-0b36b269ecb5] return self._sync_wrapper(fn, *args, **kwargs) [ 579.078456] env[61923]: ERROR nova.compute.manager [instance: 6a2d7ede-e0cc-4a36-b1b2-0b36b269ecb5] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 579.078456] env[61923]: ERROR nova.compute.manager [instance: 6a2d7ede-e0cc-4a36-b1b2-0b36b269ecb5] self.wait() [ 579.078456] env[61923]: ERROR nova.compute.manager [instance: 6a2d7ede-e0cc-4a36-b1b2-0b36b269ecb5] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 579.078456] env[61923]: ERROR nova.compute.manager [instance: 6a2d7ede-e0cc-4a36-b1b2-0b36b269ecb5] self[:] = self._gt.wait() [ 579.078456] env[61923]: ERROR nova.compute.manager [instance: 6a2d7ede-e0cc-4a36-b1b2-0b36b269ecb5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 579.078456] env[61923]: ERROR nova.compute.manager [instance: 6a2d7ede-e0cc-4a36-b1b2-0b36b269ecb5] return self._exit_event.wait() [ 579.078456] env[61923]: ERROR nova.compute.manager [instance: 6a2d7ede-e0cc-4a36-b1b2-0b36b269ecb5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 579.078966] env[61923]: ERROR nova.compute.manager [instance: 6a2d7ede-e0cc-4a36-b1b2-0b36b269ecb5] result = hub.switch() [ 579.078966] env[61923]: ERROR nova.compute.manager [instance: 6a2d7ede-e0cc-4a36-b1b2-0b36b269ecb5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 579.078966] env[61923]: ERROR nova.compute.manager [instance: 6a2d7ede-e0cc-4a36-b1b2-0b36b269ecb5] return self.greenlet.switch() [ 579.078966] env[61923]: ERROR nova.compute.manager [instance: 6a2d7ede-e0cc-4a36-b1b2-0b36b269ecb5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 579.078966] env[61923]: ERROR nova.compute.manager [instance: 6a2d7ede-e0cc-4a36-b1b2-0b36b269ecb5] result = function(*args, **kwargs) [ 579.078966] env[61923]: ERROR nova.compute.manager [instance: 6a2d7ede-e0cc-4a36-b1b2-0b36b269ecb5] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 579.078966] env[61923]: ERROR nova.compute.manager [instance: 6a2d7ede-e0cc-4a36-b1b2-0b36b269ecb5] return func(*args, **kwargs) [ 579.078966] env[61923]: ERROR nova.compute.manager [instance: 6a2d7ede-e0cc-4a36-b1b2-0b36b269ecb5] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 579.078966] env[61923]: ERROR nova.compute.manager [instance: 6a2d7ede-e0cc-4a36-b1b2-0b36b269ecb5] raise e [ 579.078966] env[61923]: ERROR nova.compute.manager [instance: 6a2d7ede-e0cc-4a36-b1b2-0b36b269ecb5] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 579.078966] env[61923]: ERROR nova.compute.manager [instance: 6a2d7ede-e0cc-4a36-b1b2-0b36b269ecb5] nwinfo = self.network_api.allocate_for_instance( [ 579.078966] env[61923]: ERROR nova.compute.manager [instance: 6a2d7ede-e0cc-4a36-b1b2-0b36b269ecb5] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 579.078966] env[61923]: ERROR nova.compute.manager [instance: 6a2d7ede-e0cc-4a36-b1b2-0b36b269ecb5] created_port_ids = self._update_ports_for_instance( [ 579.079407] env[61923]: ERROR nova.compute.manager [instance: 6a2d7ede-e0cc-4a36-b1b2-0b36b269ecb5] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 579.079407] env[61923]: ERROR nova.compute.manager [instance: 6a2d7ede-e0cc-4a36-b1b2-0b36b269ecb5] with excutils.save_and_reraise_exception(): [ 579.079407] env[61923]: ERROR nova.compute.manager [instance: 6a2d7ede-e0cc-4a36-b1b2-0b36b269ecb5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 579.079407] env[61923]: ERROR nova.compute.manager [instance: 6a2d7ede-e0cc-4a36-b1b2-0b36b269ecb5] self.force_reraise() [ 579.079407] env[61923]: ERROR nova.compute.manager [instance: 6a2d7ede-e0cc-4a36-b1b2-0b36b269ecb5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 579.079407] env[61923]: ERROR nova.compute.manager [instance: 6a2d7ede-e0cc-4a36-b1b2-0b36b269ecb5] raise self.value [ 579.079407] env[61923]: ERROR nova.compute.manager [instance: 6a2d7ede-e0cc-4a36-b1b2-0b36b269ecb5] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 579.079407] env[61923]: ERROR nova.compute.manager [instance: 6a2d7ede-e0cc-4a36-b1b2-0b36b269ecb5] updated_port = self._update_port( [ 579.079407] env[61923]: ERROR nova.compute.manager [instance: 6a2d7ede-e0cc-4a36-b1b2-0b36b269ecb5] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 579.079407] env[61923]: ERROR nova.compute.manager [instance: 6a2d7ede-e0cc-4a36-b1b2-0b36b269ecb5] _ensure_no_port_binding_failure(port) [ 579.079407] env[61923]: ERROR nova.compute.manager [instance: 6a2d7ede-e0cc-4a36-b1b2-0b36b269ecb5] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 579.079407] env[61923]: ERROR nova.compute.manager [instance: 6a2d7ede-e0cc-4a36-b1b2-0b36b269ecb5] raise exception.PortBindingFailed(port_id=port['id']) [ 579.079758] env[61923]: ERROR nova.compute.manager [instance: 6a2d7ede-e0cc-4a36-b1b2-0b36b269ecb5] nova.exception.PortBindingFailed: Binding failed for port 3b3effc4-2768-46ad-8e11-d2719c57d383, please check neutron logs for more information. [ 579.079758] env[61923]: ERROR nova.compute.manager [instance: 6a2d7ede-e0cc-4a36-b1b2-0b36b269ecb5] [ 579.079758] env[61923]: INFO nova.compute.manager [None req-fba27a13-474d-4fe3-9187-c31143f40aa5 tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] [instance: 6a2d7ede-e0cc-4a36-b1b2-0b36b269ecb5] Terminating instance [ 579.081109] env[61923]: DEBUG oslo_concurrency.lockutils [None req-fba27a13-474d-4fe3-9187-c31143f40aa5 tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] Acquiring lock "refresh_cache-6a2d7ede-e0cc-4a36-b1b2-0b36b269ecb5" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 579.081109] env[61923]: DEBUG oslo_concurrency.lockutils [None req-fba27a13-474d-4fe3-9187-c31143f40aa5 tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] Acquired lock "refresh_cache-6a2d7ede-e0cc-4a36-b1b2-0b36b269ecb5" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 579.081109] env[61923]: DEBUG nova.network.neutron [None req-fba27a13-474d-4fe3-9187-c31143f40aa5 tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] [instance: 6a2d7ede-e0cc-4a36-b1b2-0b36b269ecb5] Building network info cache for instance {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 579.128020] env[61923]: DEBUG nova.network.neutron [None req-d9e2f878-40e6-4b41-8b07-da611f2f81f1 tempest-VolumesAssistedSnapshotsTest-2043366068 tempest-VolumesAssistedSnapshotsTest-2043366068-project-member] [instance: 06bda5ee-c7a8-47d1-9e0c-a5af67d40006] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 579.260745] env[61923]: DEBUG oslo_vmware.api [None req-ae7ff058-a25b-42a3-af4b-0715677fedd4 tempest-ServersAdmin275Test-1076075580 tempest-ServersAdmin275Test-1076075580-project-admin] Task: {'id': task-1377470, 'name': ReconfigVM_Task, 'duration_secs': 0.274108} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 579.261036] env[61923]: DEBUG nova.virt.vmwareapi.volumeops [None req-ae7ff058-a25b-42a3-af4b-0715677fedd4 tempest-ServersAdmin275Test-1076075580 tempest-ServersAdmin275Test-1076075580-project-admin] [instance: f7cc960c-e06a-4c58-9367-ec8771fe09d5] Reconfigured VM instance instance-0000000a to attach disk [datastore1] f7cc960c-e06a-4c58-9367-ec8771fe09d5/f7cc960c-e06a-4c58-9367-ec8771fe09d5.vmdk or device None with type sparse {{(pid=61923) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 579.261637] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-69f27628-8ace-486a-848c-13522e931245 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 579.271787] env[61923]: DEBUG oslo_vmware.api [None req-ae7ff058-a25b-42a3-af4b-0715677fedd4 tempest-ServersAdmin275Test-1076075580 tempest-ServersAdmin275Test-1076075580-project-admin] Waiting for the task: (returnval){ [ 579.271787] env[61923]: value = "task-1377472" [ 579.271787] env[61923]: _type = "Task" [ 579.271787] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 579.280426] env[61923]: DEBUG oslo_vmware.api [None req-ae7ff058-a25b-42a3-af4b-0715677fedd4 tempest-ServersAdmin275Test-1076075580 tempest-ServersAdmin275Test-1076075580-project-admin] Task: {'id': task-1377472, 'name': Rename_Task} progress is 6%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 579.533460] env[61923]: DEBUG nova.scheduler.client.report [None req-fba27a13-474d-4fe3-9187-c31143f40aa5 tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 579.615377] env[61923]: DEBUG nova.network.neutron [None req-fba27a13-474d-4fe3-9187-c31143f40aa5 tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] [instance: 6a2d7ede-e0cc-4a36-b1b2-0b36b269ecb5] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 579.630346] env[61923]: INFO nova.compute.manager [None req-d9e2f878-40e6-4b41-8b07-da611f2f81f1 tempest-VolumesAssistedSnapshotsTest-2043366068 tempest-VolumesAssistedSnapshotsTest-2043366068-project-member] [instance: 06bda5ee-c7a8-47d1-9e0c-a5af67d40006] Took 1.04 seconds to deallocate network for instance. [ 579.721403] env[61923]: DEBUG oslo_concurrency.lockutils [None req-6833b601-845b-439d-ad97-e2f9c0927775 tempest-ServerRescueNegativeTestJSON-1435265126 tempest-ServerRescueNegativeTestJSON-1435265126-project-member] Acquiring lock "4f5348ec-f0b6-45e2-be04-31dbf1d49a3a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 579.721698] env[61923]: DEBUG oslo_concurrency.lockutils [None req-6833b601-845b-439d-ad97-e2f9c0927775 tempest-ServerRescueNegativeTestJSON-1435265126 tempest-ServerRescueNegativeTestJSON-1435265126-project-member] Lock "4f5348ec-f0b6-45e2-be04-31dbf1d49a3a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 579.722582] env[61923]: DEBUG nova.compute.manager [req-66c05c73-68e2-4dd2-bb51-b6b6d094f353 req-39cbbf3e-02c0-4955-9fac-6b56bf233e78 service nova] [instance: 6a2d7ede-e0cc-4a36-b1b2-0b36b269ecb5] Received event network-changed-3b3effc4-2768-46ad-8e11-d2719c57d383 {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 579.722782] env[61923]: DEBUG nova.compute.manager [req-66c05c73-68e2-4dd2-bb51-b6b6d094f353 req-39cbbf3e-02c0-4955-9fac-6b56bf233e78 service nova] [instance: 6a2d7ede-e0cc-4a36-b1b2-0b36b269ecb5] Refreshing instance network info cache due to event network-changed-3b3effc4-2768-46ad-8e11-d2719c57d383. {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 579.722948] env[61923]: DEBUG oslo_concurrency.lockutils [req-66c05c73-68e2-4dd2-bb51-b6b6d094f353 req-39cbbf3e-02c0-4955-9fac-6b56bf233e78 service nova] Acquiring lock "refresh_cache-6a2d7ede-e0cc-4a36-b1b2-0b36b269ecb5" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 579.744953] env[61923]: DEBUG nova.network.neutron [None req-fba27a13-474d-4fe3-9187-c31143f40aa5 tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] [instance: 6a2d7ede-e0cc-4a36-b1b2-0b36b269ecb5] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 579.784937] env[61923]: DEBUG oslo_vmware.api [None req-ae7ff058-a25b-42a3-af4b-0715677fedd4 tempest-ServersAdmin275Test-1076075580 tempest-ServersAdmin275Test-1076075580-project-admin] Task: {'id': task-1377472, 'name': Rename_Task, 'duration_secs': 0.15823} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 579.785244] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-ae7ff058-a25b-42a3-af4b-0715677fedd4 tempest-ServersAdmin275Test-1076075580 tempest-ServersAdmin275Test-1076075580-project-admin] [instance: f7cc960c-e06a-4c58-9367-ec8771fe09d5] Powering on the VM {{(pid=61923) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 579.785491] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-dce7a8f5-f8a6-4ba8-9cbc-bb0a14bd5d16 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 579.793706] env[61923]: DEBUG oslo_vmware.api [None req-ae7ff058-a25b-42a3-af4b-0715677fedd4 tempest-ServersAdmin275Test-1076075580 tempest-ServersAdmin275Test-1076075580-project-admin] Waiting for the task: (returnval){ [ 579.793706] env[61923]: value = "task-1377473" [ 579.793706] env[61923]: _type = "Task" [ 579.793706] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 579.801470] env[61923]: DEBUG oslo_vmware.api [None req-ae7ff058-a25b-42a3-af4b-0715677fedd4 tempest-ServersAdmin275Test-1076075580 tempest-ServersAdmin275Test-1076075580-project-admin] Task: {'id': task-1377473, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 579.881013] env[61923]: DEBUG nova.compute.manager [req-b896110e-ef70-4060-b718-7cd97785a652 req-07feb3a5-1bb7-4fb2-a575-76b07e0ada24 service nova] [instance: f3f872ce-cbbe-4407-b4e0-3341bb598825] Received event network-vif-deleted-793a91d7-2744-40de-8e41-428eacfdaee8 {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 580.038460] env[61923]: DEBUG oslo_concurrency.lockutils [None req-fba27a13-474d-4fe3-9187-c31143f40aa5 tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.584s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 580.038973] env[61923]: DEBUG nova.compute.manager [None req-fba27a13-474d-4fe3-9187-c31143f40aa5 tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] [instance: 9c3e2f94-91da-4255-b4d6-1c01e100ff48] Start building networks asynchronously for instance. {{(pid=61923) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 580.041864] env[61923]: DEBUG oslo_concurrency.lockutils [None req-fba27a13-474d-4fe3-9187-c31143f40aa5 tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 24.269s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 580.044159] env[61923]: INFO nova.compute.claims [None req-fba27a13-474d-4fe3-9187-c31143f40aa5 tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] [instance: 93818ffd-eb85-48f0-aebd-9774467b0a8f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 580.248127] env[61923]: DEBUG oslo_concurrency.lockutils [None req-fba27a13-474d-4fe3-9187-c31143f40aa5 tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] Releasing lock "refresh_cache-6a2d7ede-e0cc-4a36-b1b2-0b36b269ecb5" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 580.248712] env[61923]: DEBUG nova.compute.manager [None req-fba27a13-474d-4fe3-9187-c31143f40aa5 tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] [instance: 6a2d7ede-e0cc-4a36-b1b2-0b36b269ecb5] Start destroying the instance on the hypervisor. {{(pid=61923) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 580.248970] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-fba27a13-474d-4fe3-9187-c31143f40aa5 tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] [instance: 6a2d7ede-e0cc-4a36-b1b2-0b36b269ecb5] Destroying instance {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 580.249756] env[61923]: DEBUG oslo_concurrency.lockutils [req-66c05c73-68e2-4dd2-bb51-b6b6d094f353 req-39cbbf3e-02c0-4955-9fac-6b56bf233e78 service nova] Acquired lock "refresh_cache-6a2d7ede-e0cc-4a36-b1b2-0b36b269ecb5" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 580.252395] env[61923]: DEBUG nova.network.neutron [req-66c05c73-68e2-4dd2-bb51-b6b6d094f353 req-39cbbf3e-02c0-4955-9fac-6b56bf233e78 service nova] [instance: 6a2d7ede-e0cc-4a36-b1b2-0b36b269ecb5] Refreshing network info cache for port 3b3effc4-2768-46ad-8e11-d2719c57d383 {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 580.252395] env[61923]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-b7948344-82b8-4c9c-9a92-6c3a62dc8ed5 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 580.262016] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb388c66-66ec-4fba-a584-0c46db87b738 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 580.290626] env[61923]: WARNING nova.virt.vmwareapi.vmops [None req-fba27a13-474d-4fe3-9187-c31143f40aa5 tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] [instance: 6a2d7ede-e0cc-4a36-b1b2-0b36b269ecb5] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 6a2d7ede-e0cc-4a36-b1b2-0b36b269ecb5 could not be found. [ 580.291549] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-fba27a13-474d-4fe3-9187-c31143f40aa5 tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] [instance: 6a2d7ede-e0cc-4a36-b1b2-0b36b269ecb5] Instance destroyed {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 580.291549] env[61923]: INFO nova.compute.manager [None req-fba27a13-474d-4fe3-9187-c31143f40aa5 tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] [instance: 6a2d7ede-e0cc-4a36-b1b2-0b36b269ecb5] Took 0.04 seconds to destroy the instance on the hypervisor. [ 580.291549] env[61923]: DEBUG oslo.service.loopingcall [None req-fba27a13-474d-4fe3-9187-c31143f40aa5 tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61923) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 580.291706] env[61923]: DEBUG nova.compute.manager [-] [instance: 6a2d7ede-e0cc-4a36-b1b2-0b36b269ecb5] Deallocating network for instance {{(pid=61923) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 580.291706] env[61923]: DEBUG nova.network.neutron [-] [instance: 6a2d7ede-e0cc-4a36-b1b2-0b36b269ecb5] deallocate_for_instance() {{(pid=61923) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 580.302923] env[61923]: DEBUG oslo_vmware.api [None req-ae7ff058-a25b-42a3-af4b-0715677fedd4 tempest-ServersAdmin275Test-1076075580 tempest-ServersAdmin275Test-1076075580-project-admin] Task: {'id': task-1377473, 'name': PowerOnVM_Task, 'duration_secs': 0.484034} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 580.303514] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-ae7ff058-a25b-42a3-af4b-0715677fedd4 tempest-ServersAdmin275Test-1076075580 tempest-ServersAdmin275Test-1076075580-project-admin] [instance: f7cc960c-e06a-4c58-9367-ec8771fe09d5] Powered on the VM {{(pid=61923) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 580.303723] env[61923]: DEBUG nova.compute.manager [None req-ae7ff058-a25b-42a3-af4b-0715677fedd4 tempest-ServersAdmin275Test-1076075580 tempest-ServersAdmin275Test-1076075580-project-admin] [instance: f7cc960c-e06a-4c58-9367-ec8771fe09d5] Checking state {{(pid=61923) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 580.304768] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a07a7dd-66b8-4775-ac09-964ca76a772f {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 580.319264] env[61923]: DEBUG nova.network.neutron [-] [instance: 6a2d7ede-e0cc-4a36-b1b2-0b36b269ecb5] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 580.550572] env[61923]: DEBUG nova.compute.utils [None req-fba27a13-474d-4fe3-9187-c31143f40aa5 tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] Using /dev/sd instead of None {{(pid=61923) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 580.554175] env[61923]: DEBUG nova.compute.manager [None req-fba27a13-474d-4fe3-9187-c31143f40aa5 tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] [instance: 9c3e2f94-91da-4255-b4d6-1c01e100ff48] Allocating IP information in the background. {{(pid=61923) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 580.554175] env[61923]: DEBUG nova.network.neutron [None req-fba27a13-474d-4fe3-9187-c31143f40aa5 tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] [instance: 9c3e2f94-91da-4255-b4d6-1c01e100ff48] allocate_for_instance() {{(pid=61923) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 580.675190] env[61923]: DEBUG nova.policy [None req-fba27a13-474d-4fe3-9187-c31143f40aa5 tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '74859dc3474543f0829d2d7bea84bebc', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '609c22f73ddf4b9b9e764d428891f2a8', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61923) authorize /opt/stack/nova/nova/policy.py:201}} [ 580.680513] env[61923]: INFO nova.scheduler.client.report [None req-d9e2f878-40e6-4b41-8b07-da611f2f81f1 tempest-VolumesAssistedSnapshotsTest-2043366068 tempest-VolumesAssistedSnapshotsTest-2043366068-project-member] Deleted allocations for instance 06bda5ee-c7a8-47d1-9e0c-a5af67d40006 [ 580.781602] env[61923]: DEBUG nova.network.neutron [req-66c05c73-68e2-4dd2-bb51-b6b6d094f353 req-39cbbf3e-02c0-4955-9fac-6b56bf233e78 service nova] [instance: 6a2d7ede-e0cc-4a36-b1b2-0b36b269ecb5] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 580.826021] env[61923]: DEBUG nova.network.neutron [-] [instance: 6a2d7ede-e0cc-4a36-b1b2-0b36b269ecb5] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 580.829934] env[61923]: DEBUG oslo_concurrency.lockutils [None req-ae7ff058-a25b-42a3-af4b-0715677fedd4 tempest-ServersAdmin275Test-1076075580 tempest-ServersAdmin275Test-1076075580-project-admin] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 580.929528] env[61923]: DEBUG nova.network.neutron [req-66c05c73-68e2-4dd2-bb51-b6b6d094f353 req-39cbbf3e-02c0-4955-9fac-6b56bf233e78 service nova] [instance: 6a2d7ede-e0cc-4a36-b1b2-0b36b269ecb5] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 581.062977] env[61923]: DEBUG nova.compute.manager [None req-fba27a13-474d-4fe3-9187-c31143f40aa5 tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] [instance: 9c3e2f94-91da-4255-b4d6-1c01e100ff48] Start building block device mappings for instance. {{(pid=61923) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 581.187766] env[61923]: DEBUG oslo_concurrency.lockutils [None req-d9e2f878-40e6-4b41-8b07-da611f2f81f1 tempest-VolumesAssistedSnapshotsTest-2043366068 tempest-VolumesAssistedSnapshotsTest-2043366068-project-member] Lock "06bda5ee-c7a8-47d1-9e0c-a5af67d40006" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 48.637s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 581.331433] env[61923]: INFO nova.compute.manager [-] [instance: 6a2d7ede-e0cc-4a36-b1b2-0b36b269ecb5] Took 1.04 seconds to deallocate network for instance. [ 581.333990] env[61923]: DEBUG nova.compute.claims [None req-fba27a13-474d-4fe3-9187-c31143f40aa5 tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] [instance: 6a2d7ede-e0cc-4a36-b1b2-0b36b269ecb5] Aborting claim: {{(pid=61923) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 581.334223] env[61923]: DEBUG oslo_concurrency.lockutils [None req-fba27a13-474d-4fe3-9187-c31143f40aa5 tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 581.432097] env[61923]: DEBUG oslo_concurrency.lockutils [req-66c05c73-68e2-4dd2-bb51-b6b6d094f353 req-39cbbf3e-02c0-4955-9fac-6b56bf233e78 service nova] Releasing lock "refresh_cache-6a2d7ede-e0cc-4a36-b1b2-0b36b269ecb5" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 581.547780] env[61923]: DEBUG nova.network.neutron [None req-fba27a13-474d-4fe3-9187-c31143f40aa5 tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] [instance: 9c3e2f94-91da-4255-b4d6-1c01e100ff48] Successfully created port: cf3f1ca2-6afa-4f12-99b7-24adee47529f {{(pid=61923) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 581.580232] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6278d8db-6fe1-4de8-b0bf-1e2712f44620 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 581.587130] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06d5d5db-223d-4d87-93df-950031ce9352 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 581.626306] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab58053a-dc96-42f8-beba-c1f00aa44ccd {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 581.638401] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a00039e-8a39-4630-815c-af1078f7a0a5 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 581.652519] env[61923]: DEBUG nova.compute.provider_tree [None req-fba27a13-474d-4fe3-9187-c31143f40aa5 tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 581.694381] env[61923]: DEBUG nova.compute.manager [None req-f3aa5656-98df-4c46-9436-b968a8770dbc tempest-ListServerFiltersTestJSON-483030130 tempest-ListServerFiltersTestJSON-483030130-project-member] [instance: 896dd689-4810-4f23-af0a-d2f557a0796b] Starting instance... {{(pid=61923) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 582.077575] env[61923]: DEBUG nova.compute.manager [None req-fba27a13-474d-4fe3-9187-c31143f40aa5 tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] [instance: 9c3e2f94-91da-4255-b4d6-1c01e100ff48] Start spawning the instance on the hypervisor. {{(pid=61923) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 582.119717] env[61923]: DEBUG nova.virt.hardware [None req-fba27a13-474d-4fe3-9187-c31143f40aa5 tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-29T20:07:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-29T20:07:35Z,direct_url=,disk_format='vmdk',id=0f153f63-ae0a-45b1-b7f5-7f9b673c947f,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='4e7b5e3b3c3443c8bd5c70f8a15739f5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-29T20:07:36Z,virtual_size=,visibility=), allow threads: False {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 582.121689] env[61923]: DEBUG nova.virt.hardware [None req-fba27a13-474d-4fe3-9187-c31143f40aa5 tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] Flavor limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 582.121689] env[61923]: DEBUG nova.virt.hardware [None req-fba27a13-474d-4fe3-9187-c31143f40aa5 tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] Image limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 582.121689] env[61923]: DEBUG nova.virt.hardware [None req-fba27a13-474d-4fe3-9187-c31143f40aa5 tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] Flavor pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 582.121689] env[61923]: DEBUG nova.virt.hardware [None req-fba27a13-474d-4fe3-9187-c31143f40aa5 tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] Image pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 582.121996] env[61923]: DEBUG nova.virt.hardware [None req-fba27a13-474d-4fe3-9187-c31143f40aa5 tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 582.122081] env[61923]: DEBUG nova.virt.hardware [None req-fba27a13-474d-4fe3-9187-c31143f40aa5 tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 582.122853] env[61923]: DEBUG nova.virt.hardware [None req-fba27a13-474d-4fe3-9187-c31143f40aa5 tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 582.122853] env[61923]: DEBUG nova.virt.hardware [None req-fba27a13-474d-4fe3-9187-c31143f40aa5 tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] Got 1 possible topologies {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 582.122853] env[61923]: DEBUG nova.virt.hardware [None req-fba27a13-474d-4fe3-9187-c31143f40aa5 tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 582.123027] env[61923]: DEBUG nova.virt.hardware [None req-fba27a13-474d-4fe3-9187-c31143f40aa5 tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 582.124128] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-484a7f11-85db-4860-9b67-0ac313333392 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 582.136427] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad0213f2-edf7-4af6-9e2d-bc8f8e254693 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 582.156035] env[61923]: DEBUG nova.scheduler.client.report [None req-fba27a13-474d-4fe3-9187-c31143f40aa5 tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 582.228304] env[61923]: DEBUG oslo_concurrency.lockutils [None req-f3aa5656-98df-4c46-9436-b968a8770dbc tempest-ListServerFiltersTestJSON-483030130 tempest-ListServerFiltersTestJSON-483030130-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 582.597830] env[61923]: DEBUG oslo_concurrency.lockutils [None req-47902537-98f9-4fe7-afdb-4b2d4d440d10 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] Acquiring lock "f7cc960c-e06a-4c58-9367-ec8771fe09d5" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 582.598173] env[61923]: DEBUG oslo_concurrency.lockutils [None req-47902537-98f9-4fe7-afdb-4b2d4d440d10 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] Lock "f7cc960c-e06a-4c58-9367-ec8771fe09d5" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 582.598415] env[61923]: DEBUG oslo_concurrency.lockutils [None req-47902537-98f9-4fe7-afdb-4b2d4d440d10 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] Acquiring lock "f7cc960c-e06a-4c58-9367-ec8771fe09d5-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 582.598632] env[61923]: DEBUG oslo_concurrency.lockutils [None req-47902537-98f9-4fe7-afdb-4b2d4d440d10 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] Lock "f7cc960c-e06a-4c58-9367-ec8771fe09d5-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 582.598846] env[61923]: DEBUG oslo_concurrency.lockutils [None req-47902537-98f9-4fe7-afdb-4b2d4d440d10 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] Lock "f7cc960c-e06a-4c58-9367-ec8771fe09d5-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 582.602240] env[61923]: INFO nova.compute.manager [None req-47902537-98f9-4fe7-afdb-4b2d4d440d10 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] [instance: f7cc960c-e06a-4c58-9367-ec8771fe09d5] Terminating instance [ 582.604134] env[61923]: DEBUG oslo_concurrency.lockutils [None req-47902537-98f9-4fe7-afdb-4b2d4d440d10 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] Acquiring lock "refresh_cache-f7cc960c-e06a-4c58-9367-ec8771fe09d5" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 582.604360] env[61923]: DEBUG oslo_concurrency.lockutils [None req-47902537-98f9-4fe7-afdb-4b2d4d440d10 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] Acquired lock "refresh_cache-f7cc960c-e06a-4c58-9367-ec8771fe09d5" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 582.604513] env[61923]: DEBUG nova.network.neutron [None req-47902537-98f9-4fe7-afdb-4b2d4d440d10 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] [instance: f7cc960c-e06a-4c58-9367-ec8771fe09d5] Building network info cache for instance {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 582.664109] env[61923]: DEBUG oslo_concurrency.lockutils [None req-fba27a13-474d-4fe3-9187-c31143f40aa5 tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.622s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 582.664967] env[61923]: DEBUG nova.compute.manager [None req-fba27a13-474d-4fe3-9187-c31143f40aa5 tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] [instance: 93818ffd-eb85-48f0-aebd-9774467b0a8f] Start building networks asynchronously for instance. {{(pid=61923) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 582.667885] env[61923]: DEBUG oslo_concurrency.lockutils [None req-543d2693-878a-4ade-aaa2-339016c76eca tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 23.796s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 582.669462] env[61923]: INFO nova.compute.claims [None req-543d2693-878a-4ade-aaa2-339016c76eca tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] [instance: 045c0b05-333e-4f95-94b0-8f51d87dca0a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 583.126118] env[61923]: DEBUG nova.network.neutron [None req-47902537-98f9-4fe7-afdb-4b2d4d440d10 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] [instance: f7cc960c-e06a-4c58-9367-ec8771fe09d5] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 583.180313] env[61923]: DEBUG nova.compute.utils [None req-fba27a13-474d-4fe3-9187-c31143f40aa5 tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] Using /dev/sd instead of None {{(pid=61923) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 583.184046] env[61923]: DEBUG nova.compute.manager [None req-fba27a13-474d-4fe3-9187-c31143f40aa5 tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] [instance: 93818ffd-eb85-48f0-aebd-9774467b0a8f] Allocating IP information in the background. {{(pid=61923) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 583.184046] env[61923]: DEBUG nova.network.neutron [None req-fba27a13-474d-4fe3-9187-c31143f40aa5 tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] [instance: 93818ffd-eb85-48f0-aebd-9774467b0a8f] allocate_for_instance() {{(pid=61923) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 583.184721] env[61923]: DEBUG nova.network.neutron [None req-47902537-98f9-4fe7-afdb-4b2d4d440d10 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] [instance: f7cc960c-e06a-4c58-9367-ec8771fe09d5] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 583.275194] env[61923]: DEBUG nova.policy [None req-fba27a13-474d-4fe3-9187-c31143f40aa5 tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '74859dc3474543f0829d2d7bea84bebc', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '609c22f73ddf4b9b9e764d428891f2a8', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61923) authorize /opt/stack/nova/nova/policy.py:201}} [ 583.698027] env[61923]: DEBUG nova.compute.manager [None req-fba27a13-474d-4fe3-9187-c31143f40aa5 tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] [instance: 93818ffd-eb85-48f0-aebd-9774467b0a8f] Start building block device mappings for instance. {{(pid=61923) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 583.703041] env[61923]: DEBUG oslo_concurrency.lockutils [None req-47902537-98f9-4fe7-afdb-4b2d4d440d10 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] Releasing lock "refresh_cache-f7cc960c-e06a-4c58-9367-ec8771fe09d5" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 583.703041] env[61923]: DEBUG nova.compute.manager [None req-47902537-98f9-4fe7-afdb-4b2d4d440d10 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] [instance: f7cc960c-e06a-4c58-9367-ec8771fe09d5] Start destroying the instance on the hypervisor. {{(pid=61923) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 583.703041] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-47902537-98f9-4fe7-afdb-4b2d4d440d10 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] [instance: f7cc960c-e06a-4c58-9367-ec8771fe09d5] Destroying instance {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 583.703041] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4582653-6be2-4d7c-a249-bb85cbc6862b {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 583.712097] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-47902537-98f9-4fe7-afdb-4b2d4d440d10 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] [instance: f7cc960c-e06a-4c58-9367-ec8771fe09d5] Powering off the VM {{(pid=61923) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 583.712588] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f267c22e-3c45-425e-a09e-cb429dab3427 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 583.723196] env[61923]: DEBUG oslo_vmware.api [None req-47902537-98f9-4fe7-afdb-4b2d4d440d10 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] Waiting for the task: (returnval){ [ 583.723196] env[61923]: value = "task-1377476" [ 583.723196] env[61923]: _type = "Task" [ 583.723196] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 583.736560] env[61923]: DEBUG oslo_vmware.api [None req-47902537-98f9-4fe7-afdb-4b2d4d440d10 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] Task: {'id': task-1377476, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 583.760553] env[61923]: DEBUG nova.compute.manager [req-7357608c-78c7-46e9-86ec-c5fb5c472020 req-86e261b4-80d1-4405-aaf7-cb51918460cb service nova] [instance: 6a2d7ede-e0cc-4a36-b1b2-0b36b269ecb5] Received event network-vif-deleted-3b3effc4-2768-46ad-8e11-d2719c57d383 {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 583.955538] env[61923]: DEBUG oslo_concurrency.lockutils [None req-df4eb7d8-5b64-4db4-b042-683b16e0b991 tempest-ServersTestBootFromVolume-2104901628 tempest-ServersTestBootFromVolume-2104901628-project-member] Acquiring lock "fb548893-2df5-4629-bf3a-3b0f807f6357" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 583.955823] env[61923]: DEBUG oslo_concurrency.lockutils [None req-df4eb7d8-5b64-4db4-b042-683b16e0b991 tempest-ServersTestBootFromVolume-2104901628 tempest-ServersTestBootFromVolume-2104901628-project-member] Lock "fb548893-2df5-4629-bf3a-3b0f807f6357" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 584.077225] env[61923]: DEBUG nova.network.neutron [None req-fba27a13-474d-4fe3-9187-c31143f40aa5 tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] [instance: 93818ffd-eb85-48f0-aebd-9774467b0a8f] Successfully created port: e0c01f6c-546d-4532-b730-becd9e5dfcda {{(pid=61923) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 584.130531] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38697bb3-a1cc-40eb-a050-4488a2b0ef36 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 584.139518] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4948a5a6-d36f-4e9d-b326-bb6aae7dfd14 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 584.174400] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f76af52-28c7-4303-8946-8971736ca2a8 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 584.183440] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8309539d-3ac4-4eab-ae43-58a3f28f1d6e {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 584.198202] env[61923]: DEBUG nova.compute.provider_tree [None req-543d2693-878a-4ade-aaa2-339016c76eca tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 584.234710] env[61923]: DEBUG oslo_vmware.api [None req-47902537-98f9-4fe7-afdb-4b2d4d440d10 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] Task: {'id': task-1377476, 'name': PowerOffVM_Task, 'duration_secs': 0.129813} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 584.234941] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-47902537-98f9-4fe7-afdb-4b2d4d440d10 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] [instance: f7cc960c-e06a-4c58-9367-ec8771fe09d5] Powered off the VM {{(pid=61923) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 584.235351] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-47902537-98f9-4fe7-afdb-4b2d4d440d10 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] [instance: f7cc960c-e06a-4c58-9367-ec8771fe09d5] Unregistering the VM {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 584.235442] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-1aa13572-6fe9-41b0-a35f-a677798b6dac {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 584.262849] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-47902537-98f9-4fe7-afdb-4b2d4d440d10 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] [instance: f7cc960c-e06a-4c58-9367-ec8771fe09d5] Unregistered the VM {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 584.263034] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-47902537-98f9-4fe7-afdb-4b2d4d440d10 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] [instance: f7cc960c-e06a-4c58-9367-ec8771fe09d5] Deleting contents of the VM from datastore datastore1 {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 584.263215] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-47902537-98f9-4fe7-afdb-4b2d4d440d10 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] Deleting the datastore file [datastore1] f7cc960c-e06a-4c58-9367-ec8771fe09d5 {{(pid=61923) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 584.263465] env[61923]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-2b803420-e368-4f29-a79d-49e62074674d {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 584.272083] env[61923]: DEBUG oslo_vmware.api [None req-47902537-98f9-4fe7-afdb-4b2d4d440d10 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] Waiting for the task: (returnval){ [ 584.272083] env[61923]: value = "task-1377478" [ 584.272083] env[61923]: _type = "Task" [ 584.272083] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 584.282173] env[61923]: DEBUG oslo_vmware.api [None req-47902537-98f9-4fe7-afdb-4b2d4d440d10 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] Task: {'id': task-1377478, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 584.496866] env[61923]: ERROR nova.compute.manager [None req-fba27a13-474d-4fe3-9187-c31143f40aa5 tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port cf3f1ca2-6afa-4f12-99b7-24adee47529f, please check neutron logs for more information. [ 584.496866] env[61923]: ERROR nova.compute.manager Traceback (most recent call last): [ 584.496866] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 584.496866] env[61923]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 584.496866] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 584.496866] env[61923]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 584.496866] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 584.496866] env[61923]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 584.496866] env[61923]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 584.496866] env[61923]: ERROR nova.compute.manager self.force_reraise() [ 584.496866] env[61923]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 584.496866] env[61923]: ERROR nova.compute.manager raise self.value [ 584.496866] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 584.496866] env[61923]: ERROR nova.compute.manager updated_port = self._update_port( [ 584.496866] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 584.496866] env[61923]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 584.497393] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 584.497393] env[61923]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 584.497393] env[61923]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port cf3f1ca2-6afa-4f12-99b7-24adee47529f, please check neutron logs for more information. [ 584.497393] env[61923]: ERROR nova.compute.manager [ 584.497393] env[61923]: Traceback (most recent call last): [ 584.497393] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 584.497393] env[61923]: listener.cb(fileno) [ 584.497393] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 584.497393] env[61923]: result = function(*args, **kwargs) [ 584.497393] env[61923]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 584.497393] env[61923]: return func(*args, **kwargs) [ 584.497393] env[61923]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 584.497393] env[61923]: raise e [ 584.497393] env[61923]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 584.497393] env[61923]: nwinfo = self.network_api.allocate_for_instance( [ 584.497393] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 584.497393] env[61923]: created_port_ids = self._update_ports_for_instance( [ 584.497393] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 584.497393] env[61923]: with excutils.save_and_reraise_exception(): [ 584.497393] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 584.497393] env[61923]: self.force_reraise() [ 584.497393] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 584.497393] env[61923]: raise self.value [ 584.497393] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 584.497393] env[61923]: updated_port = self._update_port( [ 584.497393] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 584.497393] env[61923]: _ensure_no_port_binding_failure(port) [ 584.497393] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 584.497393] env[61923]: raise exception.PortBindingFailed(port_id=port['id']) [ 584.498304] env[61923]: nova.exception.PortBindingFailed: Binding failed for port cf3f1ca2-6afa-4f12-99b7-24adee47529f, please check neutron logs for more information. [ 584.498304] env[61923]: Removing descriptor: 14 [ 584.498304] env[61923]: ERROR nova.compute.manager [None req-fba27a13-474d-4fe3-9187-c31143f40aa5 tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] [instance: 9c3e2f94-91da-4255-b4d6-1c01e100ff48] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port cf3f1ca2-6afa-4f12-99b7-24adee47529f, please check neutron logs for more information. [ 584.498304] env[61923]: ERROR nova.compute.manager [instance: 9c3e2f94-91da-4255-b4d6-1c01e100ff48] Traceback (most recent call last): [ 584.498304] env[61923]: ERROR nova.compute.manager [instance: 9c3e2f94-91da-4255-b4d6-1c01e100ff48] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 584.498304] env[61923]: ERROR nova.compute.manager [instance: 9c3e2f94-91da-4255-b4d6-1c01e100ff48] yield resources [ 584.498304] env[61923]: ERROR nova.compute.manager [instance: 9c3e2f94-91da-4255-b4d6-1c01e100ff48] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 584.498304] env[61923]: ERROR nova.compute.manager [instance: 9c3e2f94-91da-4255-b4d6-1c01e100ff48] self.driver.spawn(context, instance, image_meta, [ 584.498304] env[61923]: ERROR nova.compute.manager [instance: 9c3e2f94-91da-4255-b4d6-1c01e100ff48] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 584.498304] env[61923]: ERROR nova.compute.manager [instance: 9c3e2f94-91da-4255-b4d6-1c01e100ff48] self._vmops.spawn(context, instance, image_meta, injected_files, [ 584.498304] env[61923]: ERROR nova.compute.manager [instance: 9c3e2f94-91da-4255-b4d6-1c01e100ff48] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 584.498304] env[61923]: ERROR nova.compute.manager [instance: 9c3e2f94-91da-4255-b4d6-1c01e100ff48] vm_ref = self.build_virtual_machine(instance, [ 584.498669] env[61923]: ERROR nova.compute.manager [instance: 9c3e2f94-91da-4255-b4d6-1c01e100ff48] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 584.498669] env[61923]: ERROR nova.compute.manager [instance: 9c3e2f94-91da-4255-b4d6-1c01e100ff48] vif_infos = vmwarevif.get_vif_info(self._session, [ 584.498669] env[61923]: ERROR nova.compute.manager [instance: 9c3e2f94-91da-4255-b4d6-1c01e100ff48] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 584.498669] env[61923]: ERROR nova.compute.manager [instance: 9c3e2f94-91da-4255-b4d6-1c01e100ff48] for vif in network_info: [ 584.498669] env[61923]: ERROR nova.compute.manager [instance: 9c3e2f94-91da-4255-b4d6-1c01e100ff48] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 584.498669] env[61923]: ERROR nova.compute.manager [instance: 9c3e2f94-91da-4255-b4d6-1c01e100ff48] return self._sync_wrapper(fn, *args, **kwargs) [ 584.498669] env[61923]: ERROR nova.compute.manager [instance: 9c3e2f94-91da-4255-b4d6-1c01e100ff48] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 584.498669] env[61923]: ERROR nova.compute.manager [instance: 9c3e2f94-91da-4255-b4d6-1c01e100ff48] self.wait() [ 584.498669] env[61923]: ERROR nova.compute.manager [instance: 9c3e2f94-91da-4255-b4d6-1c01e100ff48] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 584.498669] env[61923]: ERROR nova.compute.manager [instance: 9c3e2f94-91da-4255-b4d6-1c01e100ff48] self[:] = self._gt.wait() [ 584.498669] env[61923]: ERROR nova.compute.manager [instance: 9c3e2f94-91da-4255-b4d6-1c01e100ff48] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 584.498669] env[61923]: ERROR nova.compute.manager [instance: 9c3e2f94-91da-4255-b4d6-1c01e100ff48] return self._exit_event.wait() [ 584.498669] env[61923]: ERROR nova.compute.manager [instance: 9c3e2f94-91da-4255-b4d6-1c01e100ff48] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 584.499055] env[61923]: ERROR nova.compute.manager [instance: 9c3e2f94-91da-4255-b4d6-1c01e100ff48] result = hub.switch() [ 584.499055] env[61923]: ERROR nova.compute.manager [instance: 9c3e2f94-91da-4255-b4d6-1c01e100ff48] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 584.499055] env[61923]: ERROR nova.compute.manager [instance: 9c3e2f94-91da-4255-b4d6-1c01e100ff48] return self.greenlet.switch() [ 584.499055] env[61923]: ERROR nova.compute.manager [instance: 9c3e2f94-91da-4255-b4d6-1c01e100ff48] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 584.499055] env[61923]: ERROR nova.compute.manager [instance: 9c3e2f94-91da-4255-b4d6-1c01e100ff48] result = function(*args, **kwargs) [ 584.499055] env[61923]: ERROR nova.compute.manager [instance: 9c3e2f94-91da-4255-b4d6-1c01e100ff48] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 584.499055] env[61923]: ERROR nova.compute.manager [instance: 9c3e2f94-91da-4255-b4d6-1c01e100ff48] return func(*args, **kwargs) [ 584.499055] env[61923]: ERROR nova.compute.manager [instance: 9c3e2f94-91da-4255-b4d6-1c01e100ff48] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 584.499055] env[61923]: ERROR nova.compute.manager [instance: 9c3e2f94-91da-4255-b4d6-1c01e100ff48] raise e [ 584.499055] env[61923]: ERROR nova.compute.manager [instance: 9c3e2f94-91da-4255-b4d6-1c01e100ff48] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 584.499055] env[61923]: ERROR nova.compute.manager [instance: 9c3e2f94-91da-4255-b4d6-1c01e100ff48] nwinfo = self.network_api.allocate_for_instance( [ 584.499055] env[61923]: ERROR nova.compute.manager [instance: 9c3e2f94-91da-4255-b4d6-1c01e100ff48] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 584.499055] env[61923]: ERROR nova.compute.manager [instance: 9c3e2f94-91da-4255-b4d6-1c01e100ff48] created_port_ids = self._update_ports_for_instance( [ 584.499494] env[61923]: ERROR nova.compute.manager [instance: 9c3e2f94-91da-4255-b4d6-1c01e100ff48] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 584.499494] env[61923]: ERROR nova.compute.manager [instance: 9c3e2f94-91da-4255-b4d6-1c01e100ff48] with excutils.save_and_reraise_exception(): [ 584.499494] env[61923]: ERROR nova.compute.manager [instance: 9c3e2f94-91da-4255-b4d6-1c01e100ff48] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 584.499494] env[61923]: ERROR nova.compute.manager [instance: 9c3e2f94-91da-4255-b4d6-1c01e100ff48] self.force_reraise() [ 584.499494] env[61923]: ERROR nova.compute.manager [instance: 9c3e2f94-91da-4255-b4d6-1c01e100ff48] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 584.499494] env[61923]: ERROR nova.compute.manager [instance: 9c3e2f94-91da-4255-b4d6-1c01e100ff48] raise self.value [ 584.499494] env[61923]: ERROR nova.compute.manager [instance: 9c3e2f94-91da-4255-b4d6-1c01e100ff48] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 584.499494] env[61923]: ERROR nova.compute.manager [instance: 9c3e2f94-91da-4255-b4d6-1c01e100ff48] updated_port = self._update_port( [ 584.499494] env[61923]: ERROR nova.compute.manager [instance: 9c3e2f94-91da-4255-b4d6-1c01e100ff48] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 584.499494] env[61923]: ERROR nova.compute.manager [instance: 9c3e2f94-91da-4255-b4d6-1c01e100ff48] _ensure_no_port_binding_failure(port) [ 584.499494] env[61923]: ERROR nova.compute.manager [instance: 9c3e2f94-91da-4255-b4d6-1c01e100ff48] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 584.499494] env[61923]: ERROR nova.compute.manager [instance: 9c3e2f94-91da-4255-b4d6-1c01e100ff48] raise exception.PortBindingFailed(port_id=port['id']) [ 584.499998] env[61923]: ERROR nova.compute.manager [instance: 9c3e2f94-91da-4255-b4d6-1c01e100ff48] nova.exception.PortBindingFailed: Binding failed for port cf3f1ca2-6afa-4f12-99b7-24adee47529f, please check neutron logs for more information. [ 584.499998] env[61923]: ERROR nova.compute.manager [instance: 9c3e2f94-91da-4255-b4d6-1c01e100ff48] [ 584.499998] env[61923]: INFO nova.compute.manager [None req-fba27a13-474d-4fe3-9187-c31143f40aa5 tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] [instance: 9c3e2f94-91da-4255-b4d6-1c01e100ff48] Terminating instance [ 584.500908] env[61923]: DEBUG oslo_concurrency.lockutils [None req-fba27a13-474d-4fe3-9187-c31143f40aa5 tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] Acquiring lock "refresh_cache-9c3e2f94-91da-4255-b4d6-1c01e100ff48" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 584.504743] env[61923]: DEBUG oslo_concurrency.lockutils [None req-fba27a13-474d-4fe3-9187-c31143f40aa5 tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] Acquired lock "refresh_cache-9c3e2f94-91da-4255-b4d6-1c01e100ff48" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 584.504743] env[61923]: DEBUG nova.network.neutron [None req-fba27a13-474d-4fe3-9187-c31143f40aa5 tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] [instance: 9c3e2f94-91da-4255-b4d6-1c01e100ff48] Building network info cache for instance {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 584.702113] env[61923]: DEBUG nova.scheduler.client.report [None req-543d2693-878a-4ade-aaa2-339016c76eca tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 584.708970] env[61923]: DEBUG nova.compute.manager [None req-fba27a13-474d-4fe3-9187-c31143f40aa5 tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] [instance: 93818ffd-eb85-48f0-aebd-9774467b0a8f] Start spawning the instance on the hypervisor. {{(pid=61923) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 584.741770] env[61923]: DEBUG nova.virt.hardware [None req-fba27a13-474d-4fe3-9187-c31143f40aa5 tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-29T20:07:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-29T20:07:35Z,direct_url=,disk_format='vmdk',id=0f153f63-ae0a-45b1-b7f5-7f9b673c947f,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='4e7b5e3b3c3443c8bd5c70f8a15739f5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-29T20:07:36Z,virtual_size=,visibility=), allow threads: False {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 584.742016] env[61923]: DEBUG nova.virt.hardware [None req-fba27a13-474d-4fe3-9187-c31143f40aa5 tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] Flavor limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 584.742168] env[61923]: DEBUG nova.virt.hardware [None req-fba27a13-474d-4fe3-9187-c31143f40aa5 tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] Image limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 584.742341] env[61923]: DEBUG nova.virt.hardware [None req-fba27a13-474d-4fe3-9187-c31143f40aa5 tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] Flavor pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 584.742475] env[61923]: DEBUG nova.virt.hardware [None req-fba27a13-474d-4fe3-9187-c31143f40aa5 tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] Image pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 584.742612] env[61923]: DEBUG nova.virt.hardware [None req-fba27a13-474d-4fe3-9187-c31143f40aa5 tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 584.742807] env[61923]: DEBUG nova.virt.hardware [None req-fba27a13-474d-4fe3-9187-c31143f40aa5 tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 584.742957] env[61923]: DEBUG nova.virt.hardware [None req-fba27a13-474d-4fe3-9187-c31143f40aa5 tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 584.743177] env[61923]: DEBUG nova.virt.hardware [None req-fba27a13-474d-4fe3-9187-c31143f40aa5 tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] Got 1 possible topologies {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 584.743409] env[61923]: DEBUG nova.virt.hardware [None req-fba27a13-474d-4fe3-9187-c31143f40aa5 tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 584.743587] env[61923]: DEBUG nova.virt.hardware [None req-fba27a13-474d-4fe3-9187-c31143f40aa5 tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 584.744555] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-976e009e-1fb0-45ee-b270-c6cd83c9c841 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 584.755203] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6f756b0-60e9-422a-a936-022565961238 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 584.783188] env[61923]: DEBUG oslo_vmware.api [None req-47902537-98f9-4fe7-afdb-4b2d4d440d10 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] Task: {'id': task-1377478, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.087362} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 584.783505] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-47902537-98f9-4fe7-afdb-4b2d4d440d10 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] Deleted the datastore file {{(pid=61923) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 584.783752] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-47902537-98f9-4fe7-afdb-4b2d4d440d10 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] [instance: f7cc960c-e06a-4c58-9367-ec8771fe09d5] Deleted contents of the VM from datastore datastore1 {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 584.784064] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-47902537-98f9-4fe7-afdb-4b2d4d440d10 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] [instance: f7cc960c-e06a-4c58-9367-ec8771fe09d5] Instance destroyed {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 584.784247] env[61923]: INFO nova.compute.manager [None req-47902537-98f9-4fe7-afdb-4b2d4d440d10 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] [instance: f7cc960c-e06a-4c58-9367-ec8771fe09d5] Took 1.08 seconds to destroy the instance on the hypervisor. [ 584.784525] env[61923]: DEBUG oslo.service.loopingcall [None req-47902537-98f9-4fe7-afdb-4b2d4d440d10 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61923) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 584.784779] env[61923]: DEBUG nova.compute.manager [-] [instance: f7cc960c-e06a-4c58-9367-ec8771fe09d5] Deallocating network for instance {{(pid=61923) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 584.784896] env[61923]: DEBUG nova.network.neutron [-] [instance: f7cc960c-e06a-4c58-9367-ec8771fe09d5] deallocate_for_instance() {{(pid=61923) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 584.808660] env[61923]: DEBUG nova.network.neutron [-] [instance: f7cc960c-e06a-4c58-9367-ec8771fe09d5] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 585.034282] env[61923]: DEBUG nova.network.neutron [None req-fba27a13-474d-4fe3-9187-c31143f40aa5 tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] [instance: 9c3e2f94-91da-4255-b4d6-1c01e100ff48] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 585.213167] env[61923]: DEBUG oslo_concurrency.lockutils [None req-543d2693-878a-4ade-aaa2-339016c76eca tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.545s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 585.214010] env[61923]: DEBUG nova.compute.manager [None req-543d2693-878a-4ade-aaa2-339016c76eca tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] [instance: 045c0b05-333e-4f95-94b0-8f51d87dca0a] Start building networks asynchronously for instance. {{(pid=61923) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 585.219087] env[61923]: DEBUG oslo_concurrency.lockutils [None req-58d8b214-0b32-4993-bdd8-d652de7f9397 tempest-ServerActionsTestOtherA-1581037736 tempest-ServerActionsTestOtherA-1581037736-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 24.607s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 585.220678] env[61923]: INFO nova.compute.claims [None req-58d8b214-0b32-4993-bdd8-d652de7f9397 tempest-ServerActionsTestOtherA-1581037736 tempest-ServerActionsTestOtherA-1581037736-project-member] [instance: af5feda7-204a-47f1-b7fa-afe44832b9a2] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 585.269681] env[61923]: DEBUG nova.network.neutron [None req-fba27a13-474d-4fe3-9187-c31143f40aa5 tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] [instance: 9c3e2f94-91da-4255-b4d6-1c01e100ff48] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 585.313042] env[61923]: DEBUG nova.network.neutron [-] [instance: f7cc960c-e06a-4c58-9367-ec8771fe09d5] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 585.720809] env[61923]: DEBUG nova.compute.utils [None req-543d2693-878a-4ade-aaa2-339016c76eca tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Using /dev/sd instead of None {{(pid=61923) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 585.722213] env[61923]: DEBUG nova.compute.manager [None req-543d2693-878a-4ade-aaa2-339016c76eca tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] [instance: 045c0b05-333e-4f95-94b0-8f51d87dca0a] Allocating IP information in the background. {{(pid=61923) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 585.722403] env[61923]: DEBUG nova.network.neutron [None req-543d2693-878a-4ade-aaa2-339016c76eca tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] [instance: 045c0b05-333e-4f95-94b0-8f51d87dca0a] allocate_for_instance() {{(pid=61923) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 585.771866] env[61923]: DEBUG oslo_concurrency.lockutils [None req-fba27a13-474d-4fe3-9187-c31143f40aa5 tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] Releasing lock "refresh_cache-9c3e2f94-91da-4255-b4d6-1c01e100ff48" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 585.772322] env[61923]: DEBUG nova.compute.manager [None req-fba27a13-474d-4fe3-9187-c31143f40aa5 tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] [instance: 9c3e2f94-91da-4255-b4d6-1c01e100ff48] Start destroying the instance on the hypervisor. {{(pid=61923) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 585.772513] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-fba27a13-474d-4fe3-9187-c31143f40aa5 tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] [instance: 9c3e2f94-91da-4255-b4d6-1c01e100ff48] Destroying instance {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 585.774774] env[61923]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-8f3e1e12-0079-4aa8-8c53-039fe4bef370 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 585.785670] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-198eccb9-6103-44aa-9fff-e4ce6685212d {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 585.807260] env[61923]: WARNING nova.virt.vmwareapi.vmops [None req-fba27a13-474d-4fe3-9187-c31143f40aa5 tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] [instance: 9c3e2f94-91da-4255-b4d6-1c01e100ff48] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 9c3e2f94-91da-4255-b4d6-1c01e100ff48 could not be found. [ 585.807494] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-fba27a13-474d-4fe3-9187-c31143f40aa5 tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] [instance: 9c3e2f94-91da-4255-b4d6-1c01e100ff48] Instance destroyed {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 585.807744] env[61923]: INFO nova.compute.manager [None req-fba27a13-474d-4fe3-9187-c31143f40aa5 tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] [instance: 9c3e2f94-91da-4255-b4d6-1c01e100ff48] Took 0.04 seconds to destroy the instance on the hypervisor. [ 585.808034] env[61923]: DEBUG oslo.service.loopingcall [None req-fba27a13-474d-4fe3-9187-c31143f40aa5 tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61923) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 585.808272] env[61923]: DEBUG nova.compute.manager [-] [instance: 9c3e2f94-91da-4255-b4d6-1c01e100ff48] Deallocating network for instance {{(pid=61923) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 585.808367] env[61923]: DEBUG nova.network.neutron [-] [instance: 9c3e2f94-91da-4255-b4d6-1c01e100ff48] deallocate_for_instance() {{(pid=61923) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 585.814940] env[61923]: INFO nova.compute.manager [-] [instance: f7cc960c-e06a-4c58-9367-ec8771fe09d5] Took 1.03 seconds to deallocate network for instance. [ 585.828755] env[61923]: DEBUG nova.network.neutron [-] [instance: 9c3e2f94-91da-4255-b4d6-1c01e100ff48] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 585.995671] env[61923]: DEBUG nova.policy [None req-543d2693-878a-4ade-aaa2-339016c76eca tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'cd5039bf8f4344e490df828e643d2779', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '912300a09b80452e85e3cd13c4a644cc', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61923) authorize /opt/stack/nova/nova/policy.py:201}} [ 586.228104] env[61923]: DEBUG nova.compute.manager [None req-543d2693-878a-4ade-aaa2-339016c76eca tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] [instance: 045c0b05-333e-4f95-94b0-8f51d87dca0a] Start building block device mappings for instance. {{(pid=61923) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 586.290786] env[61923]: ERROR nova.compute.manager [None req-fba27a13-474d-4fe3-9187-c31143f40aa5 tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port e0c01f6c-546d-4532-b730-becd9e5dfcda, please check neutron logs for more information. [ 586.290786] env[61923]: ERROR nova.compute.manager Traceback (most recent call last): [ 586.290786] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 586.290786] env[61923]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 586.290786] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 586.290786] env[61923]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 586.290786] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 586.290786] env[61923]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 586.290786] env[61923]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 586.290786] env[61923]: ERROR nova.compute.manager self.force_reraise() [ 586.290786] env[61923]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 586.290786] env[61923]: ERROR nova.compute.manager raise self.value [ 586.290786] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 586.290786] env[61923]: ERROR nova.compute.manager updated_port = self._update_port( [ 586.290786] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 586.290786] env[61923]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 586.291575] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 586.291575] env[61923]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 586.291575] env[61923]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port e0c01f6c-546d-4532-b730-becd9e5dfcda, please check neutron logs for more information. [ 586.291575] env[61923]: ERROR nova.compute.manager [ 586.291575] env[61923]: Traceback (most recent call last): [ 586.291575] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 586.291575] env[61923]: listener.cb(fileno) [ 586.291575] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 586.291575] env[61923]: result = function(*args, **kwargs) [ 586.291575] env[61923]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 586.291575] env[61923]: return func(*args, **kwargs) [ 586.291575] env[61923]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 586.291575] env[61923]: raise e [ 586.291575] env[61923]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 586.291575] env[61923]: nwinfo = self.network_api.allocate_for_instance( [ 586.291575] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 586.291575] env[61923]: created_port_ids = self._update_ports_for_instance( [ 586.291575] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 586.291575] env[61923]: with excutils.save_and_reraise_exception(): [ 586.291575] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 586.291575] env[61923]: self.force_reraise() [ 586.291575] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 586.291575] env[61923]: raise self.value [ 586.291575] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 586.291575] env[61923]: updated_port = self._update_port( [ 586.291575] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 586.291575] env[61923]: _ensure_no_port_binding_failure(port) [ 586.291575] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 586.291575] env[61923]: raise exception.PortBindingFailed(port_id=port['id']) [ 586.296758] env[61923]: nova.exception.PortBindingFailed: Binding failed for port e0c01f6c-546d-4532-b730-becd9e5dfcda, please check neutron logs for more information. [ 586.296758] env[61923]: Removing descriptor: 17 [ 586.296758] env[61923]: ERROR nova.compute.manager [None req-fba27a13-474d-4fe3-9187-c31143f40aa5 tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] [instance: 93818ffd-eb85-48f0-aebd-9774467b0a8f] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port e0c01f6c-546d-4532-b730-becd9e5dfcda, please check neutron logs for more information. [ 586.296758] env[61923]: ERROR nova.compute.manager [instance: 93818ffd-eb85-48f0-aebd-9774467b0a8f] Traceback (most recent call last): [ 586.296758] env[61923]: ERROR nova.compute.manager [instance: 93818ffd-eb85-48f0-aebd-9774467b0a8f] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 586.296758] env[61923]: ERROR nova.compute.manager [instance: 93818ffd-eb85-48f0-aebd-9774467b0a8f] yield resources [ 586.296758] env[61923]: ERROR nova.compute.manager [instance: 93818ffd-eb85-48f0-aebd-9774467b0a8f] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 586.296758] env[61923]: ERROR nova.compute.manager [instance: 93818ffd-eb85-48f0-aebd-9774467b0a8f] self.driver.spawn(context, instance, image_meta, [ 586.296758] env[61923]: ERROR nova.compute.manager [instance: 93818ffd-eb85-48f0-aebd-9774467b0a8f] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 586.296758] env[61923]: ERROR nova.compute.manager [instance: 93818ffd-eb85-48f0-aebd-9774467b0a8f] self._vmops.spawn(context, instance, image_meta, injected_files, [ 586.296758] env[61923]: ERROR nova.compute.manager [instance: 93818ffd-eb85-48f0-aebd-9774467b0a8f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 586.296758] env[61923]: ERROR nova.compute.manager [instance: 93818ffd-eb85-48f0-aebd-9774467b0a8f] vm_ref = self.build_virtual_machine(instance, [ 586.297730] env[61923]: ERROR nova.compute.manager [instance: 93818ffd-eb85-48f0-aebd-9774467b0a8f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 586.297730] env[61923]: ERROR nova.compute.manager [instance: 93818ffd-eb85-48f0-aebd-9774467b0a8f] vif_infos = vmwarevif.get_vif_info(self._session, [ 586.297730] env[61923]: ERROR nova.compute.manager [instance: 93818ffd-eb85-48f0-aebd-9774467b0a8f] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 586.297730] env[61923]: ERROR nova.compute.manager [instance: 93818ffd-eb85-48f0-aebd-9774467b0a8f] for vif in network_info: [ 586.297730] env[61923]: ERROR nova.compute.manager [instance: 93818ffd-eb85-48f0-aebd-9774467b0a8f] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 586.297730] env[61923]: ERROR nova.compute.manager [instance: 93818ffd-eb85-48f0-aebd-9774467b0a8f] return self._sync_wrapper(fn, *args, **kwargs) [ 586.297730] env[61923]: ERROR nova.compute.manager [instance: 93818ffd-eb85-48f0-aebd-9774467b0a8f] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 586.297730] env[61923]: ERROR nova.compute.manager [instance: 93818ffd-eb85-48f0-aebd-9774467b0a8f] self.wait() [ 586.297730] env[61923]: ERROR nova.compute.manager [instance: 93818ffd-eb85-48f0-aebd-9774467b0a8f] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 586.297730] env[61923]: ERROR nova.compute.manager [instance: 93818ffd-eb85-48f0-aebd-9774467b0a8f] self[:] = self._gt.wait() [ 586.297730] env[61923]: ERROR nova.compute.manager [instance: 93818ffd-eb85-48f0-aebd-9774467b0a8f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 586.297730] env[61923]: ERROR nova.compute.manager [instance: 93818ffd-eb85-48f0-aebd-9774467b0a8f] return self._exit_event.wait() [ 586.297730] env[61923]: ERROR nova.compute.manager [instance: 93818ffd-eb85-48f0-aebd-9774467b0a8f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 586.298652] env[61923]: ERROR nova.compute.manager [instance: 93818ffd-eb85-48f0-aebd-9774467b0a8f] result = hub.switch() [ 586.298652] env[61923]: ERROR nova.compute.manager [instance: 93818ffd-eb85-48f0-aebd-9774467b0a8f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 586.298652] env[61923]: ERROR nova.compute.manager [instance: 93818ffd-eb85-48f0-aebd-9774467b0a8f] return self.greenlet.switch() [ 586.298652] env[61923]: ERROR nova.compute.manager [instance: 93818ffd-eb85-48f0-aebd-9774467b0a8f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 586.298652] env[61923]: ERROR nova.compute.manager [instance: 93818ffd-eb85-48f0-aebd-9774467b0a8f] result = function(*args, **kwargs) [ 586.298652] env[61923]: ERROR nova.compute.manager [instance: 93818ffd-eb85-48f0-aebd-9774467b0a8f] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 586.298652] env[61923]: ERROR nova.compute.manager [instance: 93818ffd-eb85-48f0-aebd-9774467b0a8f] return func(*args, **kwargs) [ 586.298652] env[61923]: ERROR nova.compute.manager [instance: 93818ffd-eb85-48f0-aebd-9774467b0a8f] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 586.298652] env[61923]: ERROR nova.compute.manager [instance: 93818ffd-eb85-48f0-aebd-9774467b0a8f] raise e [ 586.298652] env[61923]: ERROR nova.compute.manager [instance: 93818ffd-eb85-48f0-aebd-9774467b0a8f] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 586.298652] env[61923]: ERROR nova.compute.manager [instance: 93818ffd-eb85-48f0-aebd-9774467b0a8f] nwinfo = self.network_api.allocate_for_instance( [ 586.298652] env[61923]: ERROR nova.compute.manager [instance: 93818ffd-eb85-48f0-aebd-9774467b0a8f] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 586.298652] env[61923]: ERROR nova.compute.manager [instance: 93818ffd-eb85-48f0-aebd-9774467b0a8f] created_port_ids = self._update_ports_for_instance( [ 586.299202] env[61923]: ERROR nova.compute.manager [instance: 93818ffd-eb85-48f0-aebd-9774467b0a8f] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 586.299202] env[61923]: ERROR nova.compute.manager [instance: 93818ffd-eb85-48f0-aebd-9774467b0a8f] with excutils.save_and_reraise_exception(): [ 586.299202] env[61923]: ERROR nova.compute.manager [instance: 93818ffd-eb85-48f0-aebd-9774467b0a8f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 586.299202] env[61923]: ERROR nova.compute.manager [instance: 93818ffd-eb85-48f0-aebd-9774467b0a8f] self.force_reraise() [ 586.299202] env[61923]: ERROR nova.compute.manager [instance: 93818ffd-eb85-48f0-aebd-9774467b0a8f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 586.299202] env[61923]: ERROR nova.compute.manager [instance: 93818ffd-eb85-48f0-aebd-9774467b0a8f] raise self.value [ 586.299202] env[61923]: ERROR nova.compute.manager [instance: 93818ffd-eb85-48f0-aebd-9774467b0a8f] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 586.299202] env[61923]: ERROR nova.compute.manager [instance: 93818ffd-eb85-48f0-aebd-9774467b0a8f] updated_port = self._update_port( [ 586.299202] env[61923]: ERROR nova.compute.manager [instance: 93818ffd-eb85-48f0-aebd-9774467b0a8f] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 586.299202] env[61923]: ERROR nova.compute.manager [instance: 93818ffd-eb85-48f0-aebd-9774467b0a8f] _ensure_no_port_binding_failure(port) [ 586.299202] env[61923]: ERROR nova.compute.manager [instance: 93818ffd-eb85-48f0-aebd-9774467b0a8f] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 586.299202] env[61923]: ERROR nova.compute.manager [instance: 93818ffd-eb85-48f0-aebd-9774467b0a8f] raise exception.PortBindingFailed(port_id=port['id']) [ 586.299617] env[61923]: ERROR nova.compute.manager [instance: 93818ffd-eb85-48f0-aebd-9774467b0a8f] nova.exception.PortBindingFailed: Binding failed for port e0c01f6c-546d-4532-b730-becd9e5dfcda, please check neutron logs for more information. [ 586.299617] env[61923]: ERROR nova.compute.manager [instance: 93818ffd-eb85-48f0-aebd-9774467b0a8f] [ 586.299617] env[61923]: INFO nova.compute.manager [None req-fba27a13-474d-4fe3-9187-c31143f40aa5 tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] [instance: 93818ffd-eb85-48f0-aebd-9774467b0a8f] Terminating instance [ 586.299617] env[61923]: DEBUG oslo_concurrency.lockutils [None req-fba27a13-474d-4fe3-9187-c31143f40aa5 tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] Acquiring lock "refresh_cache-93818ffd-eb85-48f0-aebd-9774467b0a8f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 586.299617] env[61923]: DEBUG oslo_concurrency.lockutils [None req-fba27a13-474d-4fe3-9187-c31143f40aa5 tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] Acquired lock "refresh_cache-93818ffd-eb85-48f0-aebd-9774467b0a8f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 586.299617] env[61923]: DEBUG nova.network.neutron [None req-fba27a13-474d-4fe3-9187-c31143f40aa5 tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] [instance: 93818ffd-eb85-48f0-aebd-9774467b0a8f] Building network info cache for instance {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 586.322797] env[61923]: DEBUG oslo_concurrency.lockutils [None req-47902537-98f9-4fe7-afdb-4b2d4d440d10 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 586.331492] env[61923]: DEBUG nova.network.neutron [-] [instance: 9c3e2f94-91da-4255-b4d6-1c01e100ff48] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 586.671683] env[61923]: DEBUG nova.network.neutron [None req-543d2693-878a-4ade-aaa2-339016c76eca tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] [instance: 045c0b05-333e-4f95-94b0-8f51d87dca0a] Successfully created port: 3d9f5735-20ed-42f4-9565-073437eef966 {{(pid=61923) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 586.693762] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-722a0624-ec2d-4265-9470-b7629c9fcf67 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 586.704066] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97b531fc-26f5-4d20-9e56-4bfe3f6660a6 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 586.743041] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae53989a-b7b5-4ac3-9489-6f04c7232203 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 586.752583] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e02277d2-e88e-4fc6-ada7-6b9c7a0f75ba {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 586.759262] env[61923]: DEBUG nova.compute.manager [req-7b60f5b3-565a-42a0-a979-c15e64c090c9 req-55cc91aa-b846-4e4d-b24f-905a0dad7a73 service nova] [instance: 9c3e2f94-91da-4255-b4d6-1c01e100ff48] Received event network-changed-cf3f1ca2-6afa-4f12-99b7-24adee47529f {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 586.759482] env[61923]: DEBUG nova.compute.manager [req-7b60f5b3-565a-42a0-a979-c15e64c090c9 req-55cc91aa-b846-4e4d-b24f-905a0dad7a73 service nova] [instance: 9c3e2f94-91da-4255-b4d6-1c01e100ff48] Refreshing instance network info cache due to event network-changed-cf3f1ca2-6afa-4f12-99b7-24adee47529f. {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 586.759817] env[61923]: DEBUG oslo_concurrency.lockutils [req-7b60f5b3-565a-42a0-a979-c15e64c090c9 req-55cc91aa-b846-4e4d-b24f-905a0dad7a73 service nova] Acquiring lock "refresh_cache-9c3e2f94-91da-4255-b4d6-1c01e100ff48" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 586.759963] env[61923]: DEBUG oslo_concurrency.lockutils [req-7b60f5b3-565a-42a0-a979-c15e64c090c9 req-55cc91aa-b846-4e4d-b24f-905a0dad7a73 service nova] Acquired lock "refresh_cache-9c3e2f94-91da-4255-b4d6-1c01e100ff48" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 586.760167] env[61923]: DEBUG nova.network.neutron [req-7b60f5b3-565a-42a0-a979-c15e64c090c9 req-55cc91aa-b846-4e4d-b24f-905a0dad7a73 service nova] [instance: 9c3e2f94-91da-4255-b4d6-1c01e100ff48] Refreshing network info cache for port cf3f1ca2-6afa-4f12-99b7-24adee47529f {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 586.773820] env[61923]: DEBUG nova.compute.provider_tree [None req-58d8b214-0b32-4993-bdd8-d652de7f9397 tempest-ServerActionsTestOtherA-1581037736 tempest-ServerActionsTestOtherA-1581037736-project-member] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 586.829666] env[61923]: DEBUG nova.network.neutron [None req-fba27a13-474d-4fe3-9187-c31143f40aa5 tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] [instance: 93818ffd-eb85-48f0-aebd-9774467b0a8f] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 586.836166] env[61923]: INFO nova.compute.manager [-] [instance: 9c3e2f94-91da-4255-b4d6-1c01e100ff48] Took 1.03 seconds to deallocate network for instance. [ 586.838787] env[61923]: DEBUG nova.compute.claims [None req-fba27a13-474d-4fe3-9187-c31143f40aa5 tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] [instance: 9c3e2f94-91da-4255-b4d6-1c01e100ff48] Aborting claim: {{(pid=61923) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 586.838987] env[61923]: DEBUG oslo_concurrency.lockutils [None req-fba27a13-474d-4fe3-9187-c31143f40aa5 tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 587.062346] env[61923]: DEBUG nova.network.neutron [None req-fba27a13-474d-4fe3-9187-c31143f40aa5 tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] [instance: 93818ffd-eb85-48f0-aebd-9774467b0a8f] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 587.247526] env[61923]: DEBUG nova.compute.manager [None req-543d2693-878a-4ade-aaa2-339016c76eca tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] [instance: 045c0b05-333e-4f95-94b0-8f51d87dca0a] Start spawning the instance on the hypervisor. {{(pid=61923) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 587.277453] env[61923]: DEBUG nova.scheduler.client.report [None req-58d8b214-0b32-4993-bdd8-d652de7f9397 tempest-ServerActionsTestOtherA-1581037736 tempest-ServerActionsTestOtherA-1581037736-project-member] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 587.284025] env[61923]: DEBUG nova.virt.hardware [None req-543d2693-878a-4ade-aaa2-339016c76eca tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-29T20:07:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-29T20:07:35Z,direct_url=,disk_format='vmdk',id=0f153f63-ae0a-45b1-b7f5-7f9b673c947f,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='4e7b5e3b3c3443c8bd5c70f8a15739f5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-29T20:07:36Z,virtual_size=,visibility=), allow threads: False {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 587.284025] env[61923]: DEBUG nova.virt.hardware [None req-543d2693-878a-4ade-aaa2-339016c76eca tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Flavor limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 587.284025] env[61923]: DEBUG nova.virt.hardware [None req-543d2693-878a-4ade-aaa2-339016c76eca tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Image limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 587.284285] env[61923]: DEBUG nova.virt.hardware [None req-543d2693-878a-4ade-aaa2-339016c76eca tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Flavor pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 587.284285] env[61923]: DEBUG nova.virt.hardware [None req-543d2693-878a-4ade-aaa2-339016c76eca tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Image pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 587.284285] env[61923]: DEBUG nova.virt.hardware [None req-543d2693-878a-4ade-aaa2-339016c76eca tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 587.284285] env[61923]: DEBUG nova.virt.hardware [None req-543d2693-878a-4ade-aaa2-339016c76eca tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 587.284484] env[61923]: DEBUG nova.virt.hardware [None req-543d2693-878a-4ade-aaa2-339016c76eca tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 587.284638] env[61923]: DEBUG nova.virt.hardware [None req-543d2693-878a-4ade-aaa2-339016c76eca tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Got 1 possible topologies {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 587.284788] env[61923]: DEBUG nova.virt.hardware [None req-543d2693-878a-4ade-aaa2-339016c76eca tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 587.284948] env[61923]: DEBUG nova.virt.hardware [None req-543d2693-878a-4ade-aaa2-339016c76eca tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 587.286113] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87777f4d-ab2a-4557-8ca8-7fb44ee4c9a4 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 587.294953] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22f651ca-093b-4773-809c-e2c6665cd7f3 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 587.301054] env[61923]: DEBUG nova.network.neutron [req-7b60f5b3-565a-42a0-a979-c15e64c090c9 req-55cc91aa-b846-4e4d-b24f-905a0dad7a73 service nova] [instance: 9c3e2f94-91da-4255-b4d6-1c01e100ff48] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 587.477820] env[61923]: DEBUG nova.network.neutron [req-7b60f5b3-565a-42a0-a979-c15e64c090c9 req-55cc91aa-b846-4e4d-b24f-905a0dad7a73 service nova] [instance: 9c3e2f94-91da-4255-b4d6-1c01e100ff48] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 587.565451] env[61923]: DEBUG oslo_concurrency.lockutils [None req-fba27a13-474d-4fe3-9187-c31143f40aa5 tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] Releasing lock "refresh_cache-93818ffd-eb85-48f0-aebd-9774467b0a8f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 587.565880] env[61923]: DEBUG nova.compute.manager [None req-fba27a13-474d-4fe3-9187-c31143f40aa5 tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] [instance: 93818ffd-eb85-48f0-aebd-9774467b0a8f] Start destroying the instance on the hypervisor. {{(pid=61923) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 587.566354] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-fba27a13-474d-4fe3-9187-c31143f40aa5 tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] [instance: 93818ffd-eb85-48f0-aebd-9774467b0a8f] Destroying instance {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 587.566596] env[61923]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-a78bf29f-7e88-4444-a260-790c83a98811 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 587.576596] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce6e0d26-9c6d-47b4-b99f-b71b8b8a406c {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 587.603947] env[61923]: WARNING nova.virt.vmwareapi.vmops [None req-fba27a13-474d-4fe3-9187-c31143f40aa5 tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] [instance: 93818ffd-eb85-48f0-aebd-9774467b0a8f] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 93818ffd-eb85-48f0-aebd-9774467b0a8f could not be found. [ 587.605348] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-fba27a13-474d-4fe3-9187-c31143f40aa5 tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] [instance: 93818ffd-eb85-48f0-aebd-9774467b0a8f] Instance destroyed {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 587.605348] env[61923]: INFO nova.compute.manager [None req-fba27a13-474d-4fe3-9187-c31143f40aa5 tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] [instance: 93818ffd-eb85-48f0-aebd-9774467b0a8f] Took 0.04 seconds to destroy the instance on the hypervisor. [ 587.605348] env[61923]: DEBUG oslo.service.loopingcall [None req-fba27a13-474d-4fe3-9187-c31143f40aa5 tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61923) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 587.605348] env[61923]: DEBUG nova.compute.manager [-] [instance: 93818ffd-eb85-48f0-aebd-9774467b0a8f] Deallocating network for instance {{(pid=61923) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 587.605348] env[61923]: DEBUG nova.network.neutron [-] [instance: 93818ffd-eb85-48f0-aebd-9774467b0a8f] deallocate_for_instance() {{(pid=61923) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 587.637146] env[61923]: DEBUG nova.network.neutron [-] [instance: 93818ffd-eb85-48f0-aebd-9774467b0a8f] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 587.782567] env[61923]: DEBUG oslo_concurrency.lockutils [None req-58d8b214-0b32-4993-bdd8-d652de7f9397 tempest-ServerActionsTestOtherA-1581037736 tempest-ServerActionsTestOtherA-1581037736-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.563s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 587.783118] env[61923]: DEBUG nova.compute.manager [None req-58d8b214-0b32-4993-bdd8-d652de7f9397 tempest-ServerActionsTestOtherA-1581037736 tempest-ServerActionsTestOtherA-1581037736-project-member] [instance: af5feda7-204a-47f1-b7fa-afe44832b9a2] Start building networks asynchronously for instance. {{(pid=61923) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 587.786759] env[61923]: DEBUG oslo_concurrency.lockutils [None req-15b9e1af-fe57-407e-a058-c08239279239 tempest-ServersWithSpecificFlavorTestJSON-1806421899 tempest-ServersWithSpecificFlavorTestJSON-1806421899-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 23.040s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 587.980977] env[61923]: DEBUG oslo_concurrency.lockutils [req-7b60f5b3-565a-42a0-a979-c15e64c090c9 req-55cc91aa-b846-4e4d-b24f-905a0dad7a73 service nova] Releasing lock "refresh_cache-9c3e2f94-91da-4255-b4d6-1c01e100ff48" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 587.981299] env[61923]: DEBUG nova.compute.manager [req-7b60f5b3-565a-42a0-a979-c15e64c090c9 req-55cc91aa-b846-4e4d-b24f-905a0dad7a73 service nova] [instance: 9c3e2f94-91da-4255-b4d6-1c01e100ff48] Received event network-vif-deleted-cf3f1ca2-6afa-4f12-99b7-24adee47529f {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 587.981481] env[61923]: DEBUG nova.compute.manager [req-7b60f5b3-565a-42a0-a979-c15e64c090c9 req-55cc91aa-b846-4e4d-b24f-905a0dad7a73 service nova] [instance: 93818ffd-eb85-48f0-aebd-9774467b0a8f] Received event network-changed-e0c01f6c-546d-4532-b730-becd9e5dfcda {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 587.981637] env[61923]: DEBUG nova.compute.manager [req-7b60f5b3-565a-42a0-a979-c15e64c090c9 req-55cc91aa-b846-4e4d-b24f-905a0dad7a73 service nova] [instance: 93818ffd-eb85-48f0-aebd-9774467b0a8f] Refreshing instance network info cache due to event network-changed-e0c01f6c-546d-4532-b730-becd9e5dfcda. {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 587.981842] env[61923]: DEBUG oslo_concurrency.lockutils [req-7b60f5b3-565a-42a0-a979-c15e64c090c9 req-55cc91aa-b846-4e4d-b24f-905a0dad7a73 service nova] Acquiring lock "refresh_cache-93818ffd-eb85-48f0-aebd-9774467b0a8f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 587.982053] env[61923]: DEBUG oslo_concurrency.lockutils [req-7b60f5b3-565a-42a0-a979-c15e64c090c9 req-55cc91aa-b846-4e4d-b24f-905a0dad7a73 service nova] Acquired lock "refresh_cache-93818ffd-eb85-48f0-aebd-9774467b0a8f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 587.982152] env[61923]: DEBUG nova.network.neutron [req-7b60f5b3-565a-42a0-a979-c15e64c090c9 req-55cc91aa-b846-4e4d-b24f-905a0dad7a73 service nova] [instance: 93818ffd-eb85-48f0-aebd-9774467b0a8f] Refreshing network info cache for port e0c01f6c-546d-4532-b730-becd9e5dfcda {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 588.140125] env[61923]: DEBUG nova.network.neutron [-] [instance: 93818ffd-eb85-48f0-aebd-9774467b0a8f] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 588.293136] env[61923]: DEBUG nova.compute.utils [None req-58d8b214-0b32-4993-bdd8-d652de7f9397 tempest-ServerActionsTestOtherA-1581037736 tempest-ServerActionsTestOtherA-1581037736-project-member] Using /dev/sd instead of None {{(pid=61923) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 588.304329] env[61923]: DEBUG nova.compute.manager [None req-58d8b214-0b32-4993-bdd8-d652de7f9397 tempest-ServerActionsTestOtherA-1581037736 tempest-ServerActionsTestOtherA-1581037736-project-member] [instance: af5feda7-204a-47f1-b7fa-afe44832b9a2] Allocating IP information in the background. {{(pid=61923) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 588.304329] env[61923]: DEBUG nova.network.neutron [None req-58d8b214-0b32-4993-bdd8-d652de7f9397 tempest-ServerActionsTestOtherA-1581037736 tempest-ServerActionsTestOtherA-1581037736-project-member] [instance: af5feda7-204a-47f1-b7fa-afe44832b9a2] allocate_for_instance() {{(pid=61923) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 588.434689] env[61923]: DEBUG nova.policy [None req-58d8b214-0b32-4993-bdd8-d652de7f9397 tempest-ServerActionsTestOtherA-1581037736 tempest-ServerActionsTestOtherA-1581037736-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e1ffecaec7a24f56ba8fe3c7dc0e0d0f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b4c11af681cf425a87c605a6b4449451', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61923) authorize /opt/stack/nova/nova/policy.py:201}} [ 588.508525] env[61923]: DEBUG nova.network.neutron [req-7b60f5b3-565a-42a0-a979-c15e64c090c9 req-55cc91aa-b846-4e4d-b24f-905a0dad7a73 service nova] [instance: 93818ffd-eb85-48f0-aebd-9774467b0a8f] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 588.643979] env[61923]: INFO nova.compute.manager [-] [instance: 93818ffd-eb85-48f0-aebd-9774467b0a8f] Took 1.04 seconds to deallocate network for instance. [ 588.648889] env[61923]: DEBUG nova.compute.claims [None req-fba27a13-474d-4fe3-9187-c31143f40aa5 tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] [instance: 93818ffd-eb85-48f0-aebd-9774467b0a8f] Aborting claim: {{(pid=61923) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 588.649132] env[61923]: DEBUG oslo_concurrency.lockutils [None req-fba27a13-474d-4fe3-9187-c31143f40aa5 tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 588.650499] env[61923]: DEBUG nova.network.neutron [req-7b60f5b3-565a-42a0-a979-c15e64c090c9 req-55cc91aa-b846-4e4d-b24f-905a0dad7a73 service nova] [instance: 93818ffd-eb85-48f0-aebd-9774467b0a8f] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 588.759671] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42a2b1b3-96fc-45c5-a659-fca8f02c6eff {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 588.768261] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a861bf91-ea41-4db5-9234-460207d8424d {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 588.809376] env[61923]: DEBUG nova.compute.manager [None req-58d8b214-0b32-4993-bdd8-d652de7f9397 tempest-ServerActionsTestOtherA-1581037736 tempest-ServerActionsTestOtherA-1581037736-project-member] [instance: af5feda7-204a-47f1-b7fa-afe44832b9a2] Start building block device mappings for instance. {{(pid=61923) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 588.814344] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84c1e11a-d119-4dc0-9a4a-f2050045ca7f {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 588.825309] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce6475a7-4948-44a1-954d-c6687c521146 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 588.844425] env[61923]: DEBUG nova.compute.provider_tree [None req-15b9e1af-fe57-407e-a058-c08239279239 tempest-ServersWithSpecificFlavorTestJSON-1806421899 tempest-ServersWithSpecificFlavorTestJSON-1806421899-project-member] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 589.153861] env[61923]: DEBUG oslo_concurrency.lockutils [req-7b60f5b3-565a-42a0-a979-c15e64c090c9 req-55cc91aa-b846-4e4d-b24f-905a0dad7a73 service nova] Releasing lock "refresh_cache-93818ffd-eb85-48f0-aebd-9774467b0a8f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 589.348543] env[61923]: DEBUG nova.scheduler.client.report [None req-15b9e1af-fe57-407e-a058-c08239279239 tempest-ServersWithSpecificFlavorTestJSON-1806421899 tempest-ServersWithSpecificFlavorTestJSON-1806421899-project-member] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 589.567513] env[61923]: DEBUG nova.compute.manager [req-12664cf9-7472-4b13-b39b-6b7045a9d3e3 req-eeb5b0d1-eeaa-45d6-8ff0-2084e88b252f service nova] [instance: 93818ffd-eb85-48f0-aebd-9774467b0a8f] Received event network-vif-deleted-e0c01f6c-546d-4532-b730-becd9e5dfcda {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 589.613920] env[61923]: ERROR nova.compute.manager [None req-543d2693-878a-4ade-aaa2-339016c76eca tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 3d9f5735-20ed-42f4-9565-073437eef966, please check neutron logs for more information. [ 589.613920] env[61923]: ERROR nova.compute.manager Traceback (most recent call last): [ 589.613920] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 589.613920] env[61923]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 589.613920] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 589.613920] env[61923]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 589.613920] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 589.613920] env[61923]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 589.613920] env[61923]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 589.613920] env[61923]: ERROR nova.compute.manager self.force_reraise() [ 589.613920] env[61923]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 589.613920] env[61923]: ERROR nova.compute.manager raise self.value [ 589.613920] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 589.613920] env[61923]: ERROR nova.compute.manager updated_port = self._update_port( [ 589.613920] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 589.613920] env[61923]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 589.614510] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 589.614510] env[61923]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 589.614510] env[61923]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 3d9f5735-20ed-42f4-9565-073437eef966, please check neutron logs for more information. [ 589.614510] env[61923]: ERROR nova.compute.manager [ 589.614510] env[61923]: Traceback (most recent call last): [ 589.614510] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 589.614510] env[61923]: listener.cb(fileno) [ 589.614510] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 589.614510] env[61923]: result = function(*args, **kwargs) [ 589.614510] env[61923]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 589.614510] env[61923]: return func(*args, **kwargs) [ 589.614510] env[61923]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 589.614510] env[61923]: raise e [ 589.614510] env[61923]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 589.614510] env[61923]: nwinfo = self.network_api.allocate_for_instance( [ 589.614510] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 589.614510] env[61923]: created_port_ids = self._update_ports_for_instance( [ 589.614510] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 589.614510] env[61923]: with excutils.save_and_reraise_exception(): [ 589.614510] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 589.614510] env[61923]: self.force_reraise() [ 589.614510] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 589.614510] env[61923]: raise self.value [ 589.614510] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 589.614510] env[61923]: updated_port = self._update_port( [ 589.614510] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 589.614510] env[61923]: _ensure_no_port_binding_failure(port) [ 589.614510] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 589.614510] env[61923]: raise exception.PortBindingFailed(port_id=port['id']) [ 589.615489] env[61923]: nova.exception.PortBindingFailed: Binding failed for port 3d9f5735-20ed-42f4-9565-073437eef966, please check neutron logs for more information. [ 589.615489] env[61923]: Removing descriptor: 14 [ 589.615489] env[61923]: ERROR nova.compute.manager [None req-543d2693-878a-4ade-aaa2-339016c76eca tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] [instance: 045c0b05-333e-4f95-94b0-8f51d87dca0a] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 3d9f5735-20ed-42f4-9565-073437eef966, please check neutron logs for more information. [ 589.615489] env[61923]: ERROR nova.compute.manager [instance: 045c0b05-333e-4f95-94b0-8f51d87dca0a] Traceback (most recent call last): [ 589.615489] env[61923]: ERROR nova.compute.manager [instance: 045c0b05-333e-4f95-94b0-8f51d87dca0a] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 589.615489] env[61923]: ERROR nova.compute.manager [instance: 045c0b05-333e-4f95-94b0-8f51d87dca0a] yield resources [ 589.615489] env[61923]: ERROR nova.compute.manager [instance: 045c0b05-333e-4f95-94b0-8f51d87dca0a] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 589.615489] env[61923]: ERROR nova.compute.manager [instance: 045c0b05-333e-4f95-94b0-8f51d87dca0a] self.driver.spawn(context, instance, image_meta, [ 589.615489] env[61923]: ERROR nova.compute.manager [instance: 045c0b05-333e-4f95-94b0-8f51d87dca0a] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 589.615489] env[61923]: ERROR nova.compute.manager [instance: 045c0b05-333e-4f95-94b0-8f51d87dca0a] self._vmops.spawn(context, instance, image_meta, injected_files, [ 589.615489] env[61923]: ERROR nova.compute.manager [instance: 045c0b05-333e-4f95-94b0-8f51d87dca0a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 589.615489] env[61923]: ERROR nova.compute.manager [instance: 045c0b05-333e-4f95-94b0-8f51d87dca0a] vm_ref = self.build_virtual_machine(instance, [ 589.616256] env[61923]: ERROR nova.compute.manager [instance: 045c0b05-333e-4f95-94b0-8f51d87dca0a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 589.616256] env[61923]: ERROR nova.compute.manager [instance: 045c0b05-333e-4f95-94b0-8f51d87dca0a] vif_infos = vmwarevif.get_vif_info(self._session, [ 589.616256] env[61923]: ERROR nova.compute.manager [instance: 045c0b05-333e-4f95-94b0-8f51d87dca0a] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 589.616256] env[61923]: ERROR nova.compute.manager [instance: 045c0b05-333e-4f95-94b0-8f51d87dca0a] for vif in network_info: [ 589.616256] env[61923]: ERROR nova.compute.manager [instance: 045c0b05-333e-4f95-94b0-8f51d87dca0a] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 589.616256] env[61923]: ERROR nova.compute.manager [instance: 045c0b05-333e-4f95-94b0-8f51d87dca0a] return self._sync_wrapper(fn, *args, **kwargs) [ 589.616256] env[61923]: ERROR nova.compute.manager [instance: 045c0b05-333e-4f95-94b0-8f51d87dca0a] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 589.616256] env[61923]: ERROR nova.compute.manager [instance: 045c0b05-333e-4f95-94b0-8f51d87dca0a] self.wait() [ 589.616256] env[61923]: ERROR nova.compute.manager [instance: 045c0b05-333e-4f95-94b0-8f51d87dca0a] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 589.616256] env[61923]: ERROR nova.compute.manager [instance: 045c0b05-333e-4f95-94b0-8f51d87dca0a] self[:] = self._gt.wait() [ 589.616256] env[61923]: ERROR nova.compute.manager [instance: 045c0b05-333e-4f95-94b0-8f51d87dca0a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 589.616256] env[61923]: ERROR nova.compute.manager [instance: 045c0b05-333e-4f95-94b0-8f51d87dca0a] return self._exit_event.wait() [ 589.616256] env[61923]: ERROR nova.compute.manager [instance: 045c0b05-333e-4f95-94b0-8f51d87dca0a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 589.616708] env[61923]: ERROR nova.compute.manager [instance: 045c0b05-333e-4f95-94b0-8f51d87dca0a] result = hub.switch() [ 589.616708] env[61923]: ERROR nova.compute.manager [instance: 045c0b05-333e-4f95-94b0-8f51d87dca0a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 589.616708] env[61923]: ERROR nova.compute.manager [instance: 045c0b05-333e-4f95-94b0-8f51d87dca0a] return self.greenlet.switch() [ 589.616708] env[61923]: ERROR nova.compute.manager [instance: 045c0b05-333e-4f95-94b0-8f51d87dca0a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 589.616708] env[61923]: ERROR nova.compute.manager [instance: 045c0b05-333e-4f95-94b0-8f51d87dca0a] result = function(*args, **kwargs) [ 589.616708] env[61923]: ERROR nova.compute.manager [instance: 045c0b05-333e-4f95-94b0-8f51d87dca0a] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 589.616708] env[61923]: ERROR nova.compute.manager [instance: 045c0b05-333e-4f95-94b0-8f51d87dca0a] return func(*args, **kwargs) [ 589.616708] env[61923]: ERROR nova.compute.manager [instance: 045c0b05-333e-4f95-94b0-8f51d87dca0a] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 589.616708] env[61923]: ERROR nova.compute.manager [instance: 045c0b05-333e-4f95-94b0-8f51d87dca0a] raise e [ 589.616708] env[61923]: ERROR nova.compute.manager [instance: 045c0b05-333e-4f95-94b0-8f51d87dca0a] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 589.616708] env[61923]: ERROR nova.compute.manager [instance: 045c0b05-333e-4f95-94b0-8f51d87dca0a] nwinfo = self.network_api.allocate_for_instance( [ 589.616708] env[61923]: ERROR nova.compute.manager [instance: 045c0b05-333e-4f95-94b0-8f51d87dca0a] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 589.616708] env[61923]: ERROR nova.compute.manager [instance: 045c0b05-333e-4f95-94b0-8f51d87dca0a] created_port_ids = self._update_ports_for_instance( [ 589.617557] env[61923]: ERROR nova.compute.manager [instance: 045c0b05-333e-4f95-94b0-8f51d87dca0a] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 589.617557] env[61923]: ERROR nova.compute.manager [instance: 045c0b05-333e-4f95-94b0-8f51d87dca0a] with excutils.save_and_reraise_exception(): [ 589.617557] env[61923]: ERROR nova.compute.manager [instance: 045c0b05-333e-4f95-94b0-8f51d87dca0a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 589.617557] env[61923]: ERROR nova.compute.manager [instance: 045c0b05-333e-4f95-94b0-8f51d87dca0a] self.force_reraise() [ 589.617557] env[61923]: ERROR nova.compute.manager [instance: 045c0b05-333e-4f95-94b0-8f51d87dca0a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 589.617557] env[61923]: ERROR nova.compute.manager [instance: 045c0b05-333e-4f95-94b0-8f51d87dca0a] raise self.value [ 589.617557] env[61923]: ERROR nova.compute.manager [instance: 045c0b05-333e-4f95-94b0-8f51d87dca0a] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 589.617557] env[61923]: ERROR nova.compute.manager [instance: 045c0b05-333e-4f95-94b0-8f51d87dca0a] updated_port = self._update_port( [ 589.617557] env[61923]: ERROR nova.compute.manager [instance: 045c0b05-333e-4f95-94b0-8f51d87dca0a] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 589.617557] env[61923]: ERROR nova.compute.manager [instance: 045c0b05-333e-4f95-94b0-8f51d87dca0a] _ensure_no_port_binding_failure(port) [ 589.617557] env[61923]: ERROR nova.compute.manager [instance: 045c0b05-333e-4f95-94b0-8f51d87dca0a] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 589.617557] env[61923]: ERROR nova.compute.manager [instance: 045c0b05-333e-4f95-94b0-8f51d87dca0a] raise exception.PortBindingFailed(port_id=port['id']) [ 589.618162] env[61923]: ERROR nova.compute.manager [instance: 045c0b05-333e-4f95-94b0-8f51d87dca0a] nova.exception.PortBindingFailed: Binding failed for port 3d9f5735-20ed-42f4-9565-073437eef966, please check neutron logs for more information. [ 589.618162] env[61923]: ERROR nova.compute.manager [instance: 045c0b05-333e-4f95-94b0-8f51d87dca0a] [ 589.618162] env[61923]: INFO nova.compute.manager [None req-543d2693-878a-4ade-aaa2-339016c76eca tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] [instance: 045c0b05-333e-4f95-94b0-8f51d87dca0a] Terminating instance [ 589.618162] env[61923]: DEBUG oslo_concurrency.lockutils [None req-543d2693-878a-4ade-aaa2-339016c76eca tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Acquiring lock "refresh_cache-045c0b05-333e-4f95-94b0-8f51d87dca0a" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 589.618162] env[61923]: DEBUG oslo_concurrency.lockutils [None req-543d2693-878a-4ade-aaa2-339016c76eca tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Acquired lock "refresh_cache-045c0b05-333e-4f95-94b0-8f51d87dca0a" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 589.618412] env[61923]: DEBUG nova.network.neutron [None req-543d2693-878a-4ade-aaa2-339016c76eca tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] [instance: 045c0b05-333e-4f95-94b0-8f51d87dca0a] Building network info cache for instance {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 589.626358] env[61923]: DEBUG nova.network.neutron [None req-58d8b214-0b32-4993-bdd8-d652de7f9397 tempest-ServerActionsTestOtherA-1581037736 tempest-ServerActionsTestOtherA-1581037736-project-member] [instance: af5feda7-204a-47f1-b7fa-afe44832b9a2] Successfully created port: 1b61be5f-12b3-4ed1-a87e-41558aedeada {{(pid=61923) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 589.820594] env[61923]: DEBUG nova.compute.manager [None req-58d8b214-0b32-4993-bdd8-d652de7f9397 tempest-ServerActionsTestOtherA-1581037736 tempest-ServerActionsTestOtherA-1581037736-project-member] [instance: af5feda7-204a-47f1-b7fa-afe44832b9a2] Start spawning the instance on the hypervisor. {{(pid=61923) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 589.851722] env[61923]: DEBUG nova.virt.hardware [None req-58d8b214-0b32-4993-bdd8-d652de7f9397 tempest-ServerActionsTestOtherA-1581037736 tempest-ServerActionsTestOtherA-1581037736-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-29T20:07:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-29T20:07:35Z,direct_url=,disk_format='vmdk',id=0f153f63-ae0a-45b1-b7f5-7f9b673c947f,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='4e7b5e3b3c3443c8bd5c70f8a15739f5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-29T20:07:36Z,virtual_size=,visibility=), allow threads: False {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 589.851977] env[61923]: DEBUG nova.virt.hardware [None req-58d8b214-0b32-4993-bdd8-d652de7f9397 tempest-ServerActionsTestOtherA-1581037736 tempest-ServerActionsTestOtherA-1581037736-project-member] Flavor limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 589.852148] env[61923]: DEBUG nova.virt.hardware [None req-58d8b214-0b32-4993-bdd8-d652de7f9397 tempest-ServerActionsTestOtherA-1581037736 tempest-ServerActionsTestOtherA-1581037736-project-member] Image limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 589.852330] env[61923]: DEBUG nova.virt.hardware [None req-58d8b214-0b32-4993-bdd8-d652de7f9397 tempest-ServerActionsTestOtherA-1581037736 tempest-ServerActionsTestOtherA-1581037736-project-member] Flavor pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 589.852471] env[61923]: DEBUG nova.virt.hardware [None req-58d8b214-0b32-4993-bdd8-d652de7f9397 tempest-ServerActionsTestOtherA-1581037736 tempest-ServerActionsTestOtherA-1581037736-project-member] Image pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 589.852670] env[61923]: DEBUG nova.virt.hardware [None req-58d8b214-0b32-4993-bdd8-d652de7f9397 tempest-ServerActionsTestOtherA-1581037736 tempest-ServerActionsTestOtherA-1581037736-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 589.852815] env[61923]: DEBUG nova.virt.hardware [None req-58d8b214-0b32-4993-bdd8-d652de7f9397 tempest-ServerActionsTestOtherA-1581037736 tempest-ServerActionsTestOtherA-1581037736-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 589.852979] env[61923]: DEBUG nova.virt.hardware [None req-58d8b214-0b32-4993-bdd8-d652de7f9397 tempest-ServerActionsTestOtherA-1581037736 tempest-ServerActionsTestOtherA-1581037736-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 589.853602] env[61923]: DEBUG nova.virt.hardware [None req-58d8b214-0b32-4993-bdd8-d652de7f9397 tempest-ServerActionsTestOtherA-1581037736 tempest-ServerActionsTestOtherA-1581037736-project-member] Got 1 possible topologies {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 589.853983] env[61923]: DEBUG nova.virt.hardware [None req-58d8b214-0b32-4993-bdd8-d652de7f9397 tempest-ServerActionsTestOtherA-1581037736 tempest-ServerActionsTestOtherA-1581037736-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 589.853983] env[61923]: DEBUG nova.virt.hardware [None req-58d8b214-0b32-4993-bdd8-d652de7f9397 tempest-ServerActionsTestOtherA-1581037736 tempest-ServerActionsTestOtherA-1581037736-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 589.854847] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9f35911-1e1d-4dff-8e92-b2bcf1c58bf7 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 589.858488] env[61923]: DEBUG oslo_concurrency.lockutils [None req-15b9e1af-fe57-407e-a058-c08239279239 tempest-ServersWithSpecificFlavorTestJSON-1806421899 tempest-ServersWithSpecificFlavorTestJSON-1806421899-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.072s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 589.859174] env[61923]: ERROR nova.compute.manager [None req-15b9e1af-fe57-407e-a058-c08239279239 tempest-ServersWithSpecificFlavorTestJSON-1806421899 tempest-ServersWithSpecificFlavorTestJSON-1806421899-project-member] [instance: 4cec7bce-0136-43ad-898f-086058a505f2] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port fe27e2eb-c2cb-49f1-84e8-03cd81ad3229, please check neutron logs for more information. [ 589.859174] env[61923]: ERROR nova.compute.manager [instance: 4cec7bce-0136-43ad-898f-086058a505f2] Traceback (most recent call last): [ 589.859174] env[61923]: ERROR nova.compute.manager [instance: 4cec7bce-0136-43ad-898f-086058a505f2] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 589.859174] env[61923]: ERROR nova.compute.manager [instance: 4cec7bce-0136-43ad-898f-086058a505f2] self.driver.spawn(context, instance, image_meta, [ 589.859174] env[61923]: ERROR nova.compute.manager [instance: 4cec7bce-0136-43ad-898f-086058a505f2] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 589.859174] env[61923]: ERROR nova.compute.manager [instance: 4cec7bce-0136-43ad-898f-086058a505f2] self._vmops.spawn(context, instance, image_meta, injected_files, [ 589.859174] env[61923]: ERROR nova.compute.manager [instance: 4cec7bce-0136-43ad-898f-086058a505f2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 589.859174] env[61923]: ERROR nova.compute.manager [instance: 4cec7bce-0136-43ad-898f-086058a505f2] vm_ref = self.build_virtual_machine(instance, [ 589.859174] env[61923]: ERROR nova.compute.manager [instance: 4cec7bce-0136-43ad-898f-086058a505f2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 589.859174] env[61923]: ERROR nova.compute.manager [instance: 4cec7bce-0136-43ad-898f-086058a505f2] vif_infos = vmwarevif.get_vif_info(self._session, [ 589.859174] env[61923]: ERROR nova.compute.manager [instance: 4cec7bce-0136-43ad-898f-086058a505f2] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 589.859595] env[61923]: ERROR nova.compute.manager [instance: 4cec7bce-0136-43ad-898f-086058a505f2] for vif in network_info: [ 589.859595] env[61923]: ERROR nova.compute.manager [instance: 4cec7bce-0136-43ad-898f-086058a505f2] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 589.859595] env[61923]: ERROR nova.compute.manager [instance: 4cec7bce-0136-43ad-898f-086058a505f2] return self._sync_wrapper(fn, *args, **kwargs) [ 589.859595] env[61923]: ERROR nova.compute.manager [instance: 4cec7bce-0136-43ad-898f-086058a505f2] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 589.859595] env[61923]: ERROR nova.compute.manager [instance: 4cec7bce-0136-43ad-898f-086058a505f2] self.wait() [ 589.859595] env[61923]: ERROR nova.compute.manager [instance: 4cec7bce-0136-43ad-898f-086058a505f2] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 589.859595] env[61923]: ERROR nova.compute.manager [instance: 4cec7bce-0136-43ad-898f-086058a505f2] self[:] = self._gt.wait() [ 589.859595] env[61923]: ERROR nova.compute.manager [instance: 4cec7bce-0136-43ad-898f-086058a505f2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 589.859595] env[61923]: ERROR nova.compute.manager [instance: 4cec7bce-0136-43ad-898f-086058a505f2] return self._exit_event.wait() [ 589.859595] env[61923]: ERROR nova.compute.manager [instance: 4cec7bce-0136-43ad-898f-086058a505f2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 589.859595] env[61923]: ERROR nova.compute.manager [instance: 4cec7bce-0136-43ad-898f-086058a505f2] result = hub.switch() [ 589.859595] env[61923]: ERROR nova.compute.manager [instance: 4cec7bce-0136-43ad-898f-086058a505f2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 589.859595] env[61923]: ERROR nova.compute.manager [instance: 4cec7bce-0136-43ad-898f-086058a505f2] return self.greenlet.switch() [ 589.860029] env[61923]: ERROR nova.compute.manager [instance: 4cec7bce-0136-43ad-898f-086058a505f2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 589.860029] env[61923]: ERROR nova.compute.manager [instance: 4cec7bce-0136-43ad-898f-086058a505f2] result = function(*args, **kwargs) [ 589.860029] env[61923]: ERROR nova.compute.manager [instance: 4cec7bce-0136-43ad-898f-086058a505f2] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 589.860029] env[61923]: ERROR nova.compute.manager [instance: 4cec7bce-0136-43ad-898f-086058a505f2] return func(*args, **kwargs) [ 589.860029] env[61923]: ERROR nova.compute.manager [instance: 4cec7bce-0136-43ad-898f-086058a505f2] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 589.860029] env[61923]: ERROR nova.compute.manager [instance: 4cec7bce-0136-43ad-898f-086058a505f2] raise e [ 589.860029] env[61923]: ERROR nova.compute.manager [instance: 4cec7bce-0136-43ad-898f-086058a505f2] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 589.860029] env[61923]: ERROR nova.compute.manager [instance: 4cec7bce-0136-43ad-898f-086058a505f2] nwinfo = self.network_api.allocate_for_instance( [ 589.860029] env[61923]: ERROR nova.compute.manager [instance: 4cec7bce-0136-43ad-898f-086058a505f2] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 589.860029] env[61923]: ERROR nova.compute.manager [instance: 4cec7bce-0136-43ad-898f-086058a505f2] created_port_ids = self._update_ports_for_instance( [ 589.860029] env[61923]: ERROR nova.compute.manager [instance: 4cec7bce-0136-43ad-898f-086058a505f2] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 589.860029] env[61923]: ERROR nova.compute.manager [instance: 4cec7bce-0136-43ad-898f-086058a505f2] with excutils.save_and_reraise_exception(): [ 589.860029] env[61923]: ERROR nova.compute.manager [instance: 4cec7bce-0136-43ad-898f-086058a505f2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 589.860435] env[61923]: ERROR nova.compute.manager [instance: 4cec7bce-0136-43ad-898f-086058a505f2] self.force_reraise() [ 589.860435] env[61923]: ERROR nova.compute.manager [instance: 4cec7bce-0136-43ad-898f-086058a505f2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 589.860435] env[61923]: ERROR nova.compute.manager [instance: 4cec7bce-0136-43ad-898f-086058a505f2] raise self.value [ 589.860435] env[61923]: ERROR nova.compute.manager [instance: 4cec7bce-0136-43ad-898f-086058a505f2] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 589.860435] env[61923]: ERROR nova.compute.manager [instance: 4cec7bce-0136-43ad-898f-086058a505f2] updated_port = self._update_port( [ 589.860435] env[61923]: ERROR nova.compute.manager [instance: 4cec7bce-0136-43ad-898f-086058a505f2] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 589.860435] env[61923]: ERROR nova.compute.manager [instance: 4cec7bce-0136-43ad-898f-086058a505f2] _ensure_no_port_binding_failure(port) [ 589.860435] env[61923]: ERROR nova.compute.manager [instance: 4cec7bce-0136-43ad-898f-086058a505f2] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 589.860435] env[61923]: ERROR nova.compute.manager [instance: 4cec7bce-0136-43ad-898f-086058a505f2] raise exception.PortBindingFailed(port_id=port['id']) [ 589.860435] env[61923]: ERROR nova.compute.manager [instance: 4cec7bce-0136-43ad-898f-086058a505f2] nova.exception.PortBindingFailed: Binding failed for port fe27e2eb-c2cb-49f1-84e8-03cd81ad3229, please check neutron logs for more information. [ 589.860435] env[61923]: ERROR nova.compute.manager [instance: 4cec7bce-0136-43ad-898f-086058a505f2] [ 589.860775] env[61923]: DEBUG nova.compute.utils [None req-15b9e1af-fe57-407e-a058-c08239279239 tempest-ServersWithSpecificFlavorTestJSON-1806421899 tempest-ServersWithSpecificFlavorTestJSON-1806421899-project-member] [instance: 4cec7bce-0136-43ad-898f-086058a505f2] Binding failed for port fe27e2eb-c2cb-49f1-84e8-03cd81ad3229, please check neutron logs for more information. {{(pid=61923) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 589.861200] env[61923]: DEBUG oslo_concurrency.lockutils [None req-1c44cc0c-5505-4b32-99eb-cbff6c4edcc0 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 20.162s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 589.861394] env[61923]: DEBUG nova.objects.instance [None req-1c44cc0c-5505-4b32-99eb-cbff6c4edcc0 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] [instance: f7cc960c-e06a-4c58-9367-ec8771fe09d5] Trying to apply a migration context that does not seem to be set for this instance {{(pid=61923) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 589.864996] env[61923]: DEBUG nova.compute.manager [None req-15b9e1af-fe57-407e-a058-c08239279239 tempest-ServersWithSpecificFlavorTestJSON-1806421899 tempest-ServersWithSpecificFlavorTestJSON-1806421899-project-member] [instance: 4cec7bce-0136-43ad-898f-086058a505f2] Build of instance 4cec7bce-0136-43ad-898f-086058a505f2 was re-scheduled: Binding failed for port fe27e2eb-c2cb-49f1-84e8-03cd81ad3229, please check neutron logs for more information. {{(pid=61923) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 589.865459] env[61923]: DEBUG nova.compute.manager [None req-15b9e1af-fe57-407e-a058-c08239279239 tempest-ServersWithSpecificFlavorTestJSON-1806421899 tempest-ServersWithSpecificFlavorTestJSON-1806421899-project-member] [instance: 4cec7bce-0136-43ad-898f-086058a505f2] Unplugging VIFs for instance {{(pid=61923) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 589.865689] env[61923]: DEBUG oslo_concurrency.lockutils [None req-15b9e1af-fe57-407e-a058-c08239279239 tempest-ServersWithSpecificFlavorTestJSON-1806421899 tempest-ServersWithSpecificFlavorTestJSON-1806421899-project-member] Acquiring lock "refresh_cache-4cec7bce-0136-43ad-898f-086058a505f2" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 589.867446] env[61923]: DEBUG oslo_concurrency.lockutils [None req-15b9e1af-fe57-407e-a058-c08239279239 tempest-ServersWithSpecificFlavorTestJSON-1806421899 tempest-ServersWithSpecificFlavorTestJSON-1806421899-project-member] Acquired lock "refresh_cache-4cec7bce-0136-43ad-898f-086058a505f2" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 589.867446] env[61923]: DEBUG nova.network.neutron [None req-15b9e1af-fe57-407e-a058-c08239279239 tempest-ServersWithSpecificFlavorTestJSON-1806421899 tempest-ServersWithSpecificFlavorTestJSON-1806421899-project-member] [instance: 4cec7bce-0136-43ad-898f-086058a505f2] Building network info cache for instance {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 589.875735] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0772562-455a-4faa-b110-728ebb029b4f {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 590.159603] env[61923]: DEBUG nova.network.neutron [None req-543d2693-878a-4ade-aaa2-339016c76eca tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] [instance: 045c0b05-333e-4f95-94b0-8f51d87dca0a] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 590.400322] env[61923]: DEBUG nova.network.neutron [None req-15b9e1af-fe57-407e-a058-c08239279239 tempest-ServersWithSpecificFlavorTestJSON-1806421899 tempest-ServersWithSpecificFlavorTestJSON-1806421899-project-member] [instance: 4cec7bce-0136-43ad-898f-086058a505f2] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 590.437833] env[61923]: DEBUG nova.network.neutron [None req-543d2693-878a-4ade-aaa2-339016c76eca tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] [instance: 045c0b05-333e-4f95-94b0-8f51d87dca0a] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 590.523026] env[61923]: DEBUG nova.network.neutron [None req-15b9e1af-fe57-407e-a058-c08239279239 tempest-ServersWithSpecificFlavorTestJSON-1806421899 tempest-ServersWithSpecificFlavorTestJSON-1806421899-project-member] [instance: 4cec7bce-0136-43ad-898f-086058a505f2] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 590.880790] env[61923]: DEBUG oslo_concurrency.lockutils [None req-1c44cc0c-5505-4b32-99eb-cbff6c4edcc0 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.019s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 590.885947] env[61923]: DEBUG oslo_concurrency.lockutils [None req-595ebb22-536b-4375-9cbe-6fc472a5a17d tempest-ListServerFiltersTestJSON-483030130 tempest-ListServerFiltersTestJSON-483030130-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.549s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 590.885947] env[61923]: INFO nova.compute.claims [None req-595ebb22-536b-4375-9cbe-6fc472a5a17d tempest-ListServerFiltersTestJSON-483030130 tempest-ListServerFiltersTestJSON-483030130-project-member] [instance: 71d9eccc-3b90-4b0a-a59a-d6a8e08a9417] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 590.941189] env[61923]: DEBUG oslo_concurrency.lockutils [None req-543d2693-878a-4ade-aaa2-339016c76eca tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Releasing lock "refresh_cache-045c0b05-333e-4f95-94b0-8f51d87dca0a" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 590.942736] env[61923]: DEBUG nova.compute.manager [None req-543d2693-878a-4ade-aaa2-339016c76eca tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] [instance: 045c0b05-333e-4f95-94b0-8f51d87dca0a] Start destroying the instance on the hypervisor. {{(pid=61923) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 590.942974] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-543d2693-878a-4ade-aaa2-339016c76eca tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] [instance: 045c0b05-333e-4f95-94b0-8f51d87dca0a] Destroying instance {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 590.943356] env[61923]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-1dcbdead-a663-47bf-a1c1-9d0fe1dff5af {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 590.955819] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c45eb5b3-0f26-411b-a7fc-f9f916cca8c0 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 590.985021] env[61923]: WARNING nova.virt.vmwareapi.vmops [None req-543d2693-878a-4ade-aaa2-339016c76eca tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] [instance: 045c0b05-333e-4f95-94b0-8f51d87dca0a] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 045c0b05-333e-4f95-94b0-8f51d87dca0a could not be found. [ 590.985106] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-543d2693-878a-4ade-aaa2-339016c76eca tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] [instance: 045c0b05-333e-4f95-94b0-8f51d87dca0a] Instance destroyed {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 590.985358] env[61923]: INFO nova.compute.manager [None req-543d2693-878a-4ade-aaa2-339016c76eca tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] [instance: 045c0b05-333e-4f95-94b0-8f51d87dca0a] Took 0.04 seconds to destroy the instance on the hypervisor. [ 590.985736] env[61923]: DEBUG oslo.service.loopingcall [None req-543d2693-878a-4ade-aaa2-339016c76eca tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61923) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 590.986308] env[61923]: DEBUG nova.compute.manager [-] [instance: 045c0b05-333e-4f95-94b0-8f51d87dca0a] Deallocating network for instance {{(pid=61923) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 590.986308] env[61923]: DEBUG nova.network.neutron [-] [instance: 045c0b05-333e-4f95-94b0-8f51d87dca0a] deallocate_for_instance() {{(pid=61923) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 591.020895] env[61923]: DEBUG nova.network.neutron [-] [instance: 045c0b05-333e-4f95-94b0-8f51d87dca0a] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 591.028021] env[61923]: DEBUG oslo_concurrency.lockutils [None req-15b9e1af-fe57-407e-a058-c08239279239 tempest-ServersWithSpecificFlavorTestJSON-1806421899 tempest-ServersWithSpecificFlavorTestJSON-1806421899-project-member] Releasing lock "refresh_cache-4cec7bce-0136-43ad-898f-086058a505f2" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 591.028021] env[61923]: DEBUG nova.compute.manager [None req-15b9e1af-fe57-407e-a058-c08239279239 tempest-ServersWithSpecificFlavorTestJSON-1806421899 tempest-ServersWithSpecificFlavorTestJSON-1806421899-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61923) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 591.028021] env[61923]: DEBUG nova.compute.manager [None req-15b9e1af-fe57-407e-a058-c08239279239 tempest-ServersWithSpecificFlavorTestJSON-1806421899 tempest-ServersWithSpecificFlavorTestJSON-1806421899-project-member] [instance: 4cec7bce-0136-43ad-898f-086058a505f2] Deallocating network for instance {{(pid=61923) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 591.028021] env[61923]: DEBUG nova.network.neutron [None req-15b9e1af-fe57-407e-a058-c08239279239 tempest-ServersWithSpecificFlavorTestJSON-1806421899 tempest-ServersWithSpecificFlavorTestJSON-1806421899-project-member] [instance: 4cec7bce-0136-43ad-898f-086058a505f2] deallocate_for_instance() {{(pid=61923) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 591.052347] env[61923]: DEBUG nova.network.neutron [None req-15b9e1af-fe57-407e-a058-c08239279239 tempest-ServersWithSpecificFlavorTestJSON-1806421899 tempest-ServersWithSpecificFlavorTestJSON-1806421899-project-member] [instance: 4cec7bce-0136-43ad-898f-086058a505f2] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 591.527926] env[61923]: DEBUG nova.network.neutron [-] [instance: 045c0b05-333e-4f95-94b0-8f51d87dca0a] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 591.555547] env[61923]: DEBUG nova.network.neutron [None req-15b9e1af-fe57-407e-a058-c08239279239 tempest-ServersWithSpecificFlavorTestJSON-1806421899 tempest-ServersWithSpecificFlavorTestJSON-1806421899-project-member] [instance: 4cec7bce-0136-43ad-898f-086058a505f2] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 592.033950] env[61923]: INFO nova.compute.manager [-] [instance: 045c0b05-333e-4f95-94b0-8f51d87dca0a] Took 1.05 seconds to deallocate network for instance. [ 592.038437] env[61923]: DEBUG nova.compute.claims [None req-543d2693-878a-4ade-aaa2-339016c76eca tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] [instance: 045c0b05-333e-4f95-94b0-8f51d87dca0a] Aborting claim: {{(pid=61923) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 592.038655] env[61923]: DEBUG oslo_concurrency.lockutils [None req-543d2693-878a-4ade-aaa2-339016c76eca tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 592.061245] env[61923]: INFO nova.compute.manager [None req-15b9e1af-fe57-407e-a058-c08239279239 tempest-ServersWithSpecificFlavorTestJSON-1806421899 tempest-ServersWithSpecificFlavorTestJSON-1806421899-project-member] [instance: 4cec7bce-0136-43ad-898f-086058a505f2] Took 1.04 seconds to deallocate network for instance. [ 592.340574] env[61923]: DEBUG oslo_concurrency.lockutils [None req-44a76586-0f76-448f-af41-a5d846bfda81 tempest-ServerActionsV293TestJSON-849444864 tempest-ServerActionsV293TestJSON-849444864-project-member] Acquiring lock "62144317-0354-4751-8f2b-4eae588c9da7" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 592.340574] env[61923]: DEBUG oslo_concurrency.lockutils [None req-44a76586-0f76-448f-af41-a5d846bfda81 tempest-ServerActionsV293TestJSON-849444864 tempest-ServerActionsV293TestJSON-849444864-project-member] Lock "62144317-0354-4751-8f2b-4eae588c9da7" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 592.390502] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5479f5ac-641f-4ee6-b653-6535a3dc13ab {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 592.399594] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-871ebf22-79d0-4f71-9a3b-59c5c709ec74 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 592.433394] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7720c44b-f30e-414c-87a0-a442efb5d324 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 592.446624] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52054af7-bf35-44d9-baf6-689ce398fe3b {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 592.465379] env[61923]: DEBUG nova.compute.provider_tree [None req-595ebb22-536b-4375-9cbe-6fc472a5a17d tempest-ListServerFiltersTestJSON-483030130 tempest-ListServerFiltersTestJSON-483030130-project-member] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 592.482175] env[61923]: DEBUG nova.compute.manager [req-6fbf57b9-e432-4670-95ad-3d7df4a4c8b4 req-3687c9f6-a4f8-42b4-95ca-b59ec280d1d3 service nova] [instance: 045c0b05-333e-4f95-94b0-8f51d87dca0a] Received event network-changed-3d9f5735-20ed-42f4-9565-073437eef966 {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 592.482342] env[61923]: DEBUG nova.compute.manager [req-6fbf57b9-e432-4670-95ad-3d7df4a4c8b4 req-3687c9f6-a4f8-42b4-95ca-b59ec280d1d3 service nova] [instance: 045c0b05-333e-4f95-94b0-8f51d87dca0a] Refreshing instance network info cache due to event network-changed-3d9f5735-20ed-42f4-9565-073437eef966. {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 592.482548] env[61923]: DEBUG oslo_concurrency.lockutils [req-6fbf57b9-e432-4670-95ad-3d7df4a4c8b4 req-3687c9f6-a4f8-42b4-95ca-b59ec280d1d3 service nova] Acquiring lock "refresh_cache-045c0b05-333e-4f95-94b0-8f51d87dca0a" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 592.482684] env[61923]: DEBUG oslo_concurrency.lockutils [req-6fbf57b9-e432-4670-95ad-3d7df4a4c8b4 req-3687c9f6-a4f8-42b4-95ca-b59ec280d1d3 service nova] Acquired lock "refresh_cache-045c0b05-333e-4f95-94b0-8f51d87dca0a" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 592.484260] env[61923]: DEBUG nova.network.neutron [req-6fbf57b9-e432-4670-95ad-3d7df4a4c8b4 req-3687c9f6-a4f8-42b4-95ca-b59ec280d1d3 service nova] [instance: 045c0b05-333e-4f95-94b0-8f51d87dca0a] Refreshing network info cache for port 3d9f5735-20ed-42f4-9565-073437eef966 {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 592.510376] env[61923]: ERROR nova.compute.manager [None req-58d8b214-0b32-4993-bdd8-d652de7f9397 tempest-ServerActionsTestOtherA-1581037736 tempest-ServerActionsTestOtherA-1581037736-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 1b61be5f-12b3-4ed1-a87e-41558aedeada, please check neutron logs for more information. [ 592.510376] env[61923]: ERROR nova.compute.manager Traceback (most recent call last): [ 592.510376] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 592.510376] env[61923]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 592.510376] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 592.510376] env[61923]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 592.510376] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 592.510376] env[61923]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 592.510376] env[61923]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 592.510376] env[61923]: ERROR nova.compute.manager self.force_reraise() [ 592.510376] env[61923]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 592.510376] env[61923]: ERROR nova.compute.manager raise self.value [ 592.510376] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 592.510376] env[61923]: ERROR nova.compute.manager updated_port = self._update_port( [ 592.510376] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 592.510376] env[61923]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 592.510978] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 592.510978] env[61923]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 592.510978] env[61923]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 1b61be5f-12b3-4ed1-a87e-41558aedeada, please check neutron logs for more information. [ 592.510978] env[61923]: ERROR nova.compute.manager [ 592.510978] env[61923]: Traceback (most recent call last): [ 592.510978] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 592.510978] env[61923]: listener.cb(fileno) [ 592.510978] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 592.510978] env[61923]: result = function(*args, **kwargs) [ 592.510978] env[61923]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 592.510978] env[61923]: return func(*args, **kwargs) [ 592.510978] env[61923]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 592.510978] env[61923]: raise e [ 592.510978] env[61923]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 592.510978] env[61923]: nwinfo = self.network_api.allocate_for_instance( [ 592.510978] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 592.510978] env[61923]: created_port_ids = self._update_ports_for_instance( [ 592.510978] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 592.510978] env[61923]: with excutils.save_and_reraise_exception(): [ 592.510978] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 592.510978] env[61923]: self.force_reraise() [ 592.510978] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 592.510978] env[61923]: raise self.value [ 592.510978] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 592.510978] env[61923]: updated_port = self._update_port( [ 592.510978] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 592.510978] env[61923]: _ensure_no_port_binding_failure(port) [ 592.510978] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 592.510978] env[61923]: raise exception.PortBindingFailed(port_id=port['id']) [ 592.512035] env[61923]: nova.exception.PortBindingFailed: Binding failed for port 1b61be5f-12b3-4ed1-a87e-41558aedeada, please check neutron logs for more information. [ 592.512035] env[61923]: Removing descriptor: 17 [ 592.512035] env[61923]: ERROR nova.compute.manager [None req-58d8b214-0b32-4993-bdd8-d652de7f9397 tempest-ServerActionsTestOtherA-1581037736 tempest-ServerActionsTestOtherA-1581037736-project-member] [instance: af5feda7-204a-47f1-b7fa-afe44832b9a2] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 1b61be5f-12b3-4ed1-a87e-41558aedeada, please check neutron logs for more information. [ 592.512035] env[61923]: ERROR nova.compute.manager [instance: af5feda7-204a-47f1-b7fa-afe44832b9a2] Traceback (most recent call last): [ 592.512035] env[61923]: ERROR nova.compute.manager [instance: af5feda7-204a-47f1-b7fa-afe44832b9a2] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 592.512035] env[61923]: ERROR nova.compute.manager [instance: af5feda7-204a-47f1-b7fa-afe44832b9a2] yield resources [ 592.512035] env[61923]: ERROR nova.compute.manager [instance: af5feda7-204a-47f1-b7fa-afe44832b9a2] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 592.512035] env[61923]: ERROR nova.compute.manager [instance: af5feda7-204a-47f1-b7fa-afe44832b9a2] self.driver.spawn(context, instance, image_meta, [ 592.512035] env[61923]: ERROR nova.compute.manager [instance: af5feda7-204a-47f1-b7fa-afe44832b9a2] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 592.512035] env[61923]: ERROR nova.compute.manager [instance: af5feda7-204a-47f1-b7fa-afe44832b9a2] self._vmops.spawn(context, instance, image_meta, injected_files, [ 592.512035] env[61923]: ERROR nova.compute.manager [instance: af5feda7-204a-47f1-b7fa-afe44832b9a2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 592.512035] env[61923]: ERROR nova.compute.manager [instance: af5feda7-204a-47f1-b7fa-afe44832b9a2] vm_ref = self.build_virtual_machine(instance, [ 592.512450] env[61923]: ERROR nova.compute.manager [instance: af5feda7-204a-47f1-b7fa-afe44832b9a2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 592.512450] env[61923]: ERROR nova.compute.manager [instance: af5feda7-204a-47f1-b7fa-afe44832b9a2] vif_infos = vmwarevif.get_vif_info(self._session, [ 592.512450] env[61923]: ERROR nova.compute.manager [instance: af5feda7-204a-47f1-b7fa-afe44832b9a2] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 592.512450] env[61923]: ERROR nova.compute.manager [instance: af5feda7-204a-47f1-b7fa-afe44832b9a2] for vif in network_info: [ 592.512450] env[61923]: ERROR nova.compute.manager [instance: af5feda7-204a-47f1-b7fa-afe44832b9a2] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 592.512450] env[61923]: ERROR nova.compute.manager [instance: af5feda7-204a-47f1-b7fa-afe44832b9a2] return self._sync_wrapper(fn, *args, **kwargs) [ 592.512450] env[61923]: ERROR nova.compute.manager [instance: af5feda7-204a-47f1-b7fa-afe44832b9a2] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 592.512450] env[61923]: ERROR nova.compute.manager [instance: af5feda7-204a-47f1-b7fa-afe44832b9a2] self.wait() [ 592.512450] env[61923]: ERROR nova.compute.manager [instance: af5feda7-204a-47f1-b7fa-afe44832b9a2] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 592.512450] env[61923]: ERROR nova.compute.manager [instance: af5feda7-204a-47f1-b7fa-afe44832b9a2] self[:] = self._gt.wait() [ 592.512450] env[61923]: ERROR nova.compute.manager [instance: af5feda7-204a-47f1-b7fa-afe44832b9a2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 592.512450] env[61923]: ERROR nova.compute.manager [instance: af5feda7-204a-47f1-b7fa-afe44832b9a2] return self._exit_event.wait() [ 592.512450] env[61923]: ERROR nova.compute.manager [instance: af5feda7-204a-47f1-b7fa-afe44832b9a2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 592.512899] env[61923]: ERROR nova.compute.manager [instance: af5feda7-204a-47f1-b7fa-afe44832b9a2] result = hub.switch() [ 592.512899] env[61923]: ERROR nova.compute.manager [instance: af5feda7-204a-47f1-b7fa-afe44832b9a2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 592.512899] env[61923]: ERROR nova.compute.manager [instance: af5feda7-204a-47f1-b7fa-afe44832b9a2] return self.greenlet.switch() [ 592.512899] env[61923]: ERROR nova.compute.manager [instance: af5feda7-204a-47f1-b7fa-afe44832b9a2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 592.512899] env[61923]: ERROR nova.compute.manager [instance: af5feda7-204a-47f1-b7fa-afe44832b9a2] result = function(*args, **kwargs) [ 592.512899] env[61923]: ERROR nova.compute.manager [instance: af5feda7-204a-47f1-b7fa-afe44832b9a2] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 592.512899] env[61923]: ERROR nova.compute.manager [instance: af5feda7-204a-47f1-b7fa-afe44832b9a2] return func(*args, **kwargs) [ 592.512899] env[61923]: ERROR nova.compute.manager [instance: af5feda7-204a-47f1-b7fa-afe44832b9a2] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 592.512899] env[61923]: ERROR nova.compute.manager [instance: af5feda7-204a-47f1-b7fa-afe44832b9a2] raise e [ 592.512899] env[61923]: ERROR nova.compute.manager [instance: af5feda7-204a-47f1-b7fa-afe44832b9a2] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 592.512899] env[61923]: ERROR nova.compute.manager [instance: af5feda7-204a-47f1-b7fa-afe44832b9a2] nwinfo = self.network_api.allocate_for_instance( [ 592.512899] env[61923]: ERROR nova.compute.manager [instance: af5feda7-204a-47f1-b7fa-afe44832b9a2] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 592.512899] env[61923]: ERROR nova.compute.manager [instance: af5feda7-204a-47f1-b7fa-afe44832b9a2] created_port_ids = self._update_ports_for_instance( [ 592.513408] env[61923]: ERROR nova.compute.manager [instance: af5feda7-204a-47f1-b7fa-afe44832b9a2] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 592.513408] env[61923]: ERROR nova.compute.manager [instance: af5feda7-204a-47f1-b7fa-afe44832b9a2] with excutils.save_and_reraise_exception(): [ 592.513408] env[61923]: ERROR nova.compute.manager [instance: af5feda7-204a-47f1-b7fa-afe44832b9a2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 592.513408] env[61923]: ERROR nova.compute.manager [instance: af5feda7-204a-47f1-b7fa-afe44832b9a2] self.force_reraise() [ 592.513408] env[61923]: ERROR nova.compute.manager [instance: af5feda7-204a-47f1-b7fa-afe44832b9a2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 592.513408] env[61923]: ERROR nova.compute.manager [instance: af5feda7-204a-47f1-b7fa-afe44832b9a2] raise self.value [ 592.513408] env[61923]: ERROR nova.compute.manager [instance: af5feda7-204a-47f1-b7fa-afe44832b9a2] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 592.513408] env[61923]: ERROR nova.compute.manager [instance: af5feda7-204a-47f1-b7fa-afe44832b9a2] updated_port = self._update_port( [ 592.513408] env[61923]: ERROR nova.compute.manager [instance: af5feda7-204a-47f1-b7fa-afe44832b9a2] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 592.513408] env[61923]: ERROR nova.compute.manager [instance: af5feda7-204a-47f1-b7fa-afe44832b9a2] _ensure_no_port_binding_failure(port) [ 592.513408] env[61923]: ERROR nova.compute.manager [instance: af5feda7-204a-47f1-b7fa-afe44832b9a2] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 592.513408] env[61923]: ERROR nova.compute.manager [instance: af5feda7-204a-47f1-b7fa-afe44832b9a2] raise exception.PortBindingFailed(port_id=port['id']) [ 592.513843] env[61923]: ERROR nova.compute.manager [instance: af5feda7-204a-47f1-b7fa-afe44832b9a2] nova.exception.PortBindingFailed: Binding failed for port 1b61be5f-12b3-4ed1-a87e-41558aedeada, please check neutron logs for more information. [ 592.513843] env[61923]: ERROR nova.compute.manager [instance: af5feda7-204a-47f1-b7fa-afe44832b9a2] [ 592.513843] env[61923]: INFO nova.compute.manager [None req-58d8b214-0b32-4993-bdd8-d652de7f9397 tempest-ServerActionsTestOtherA-1581037736 tempest-ServerActionsTestOtherA-1581037736-project-member] [instance: af5feda7-204a-47f1-b7fa-afe44832b9a2] Terminating instance [ 592.520719] env[61923]: DEBUG oslo_concurrency.lockutils [None req-58d8b214-0b32-4993-bdd8-d652de7f9397 tempest-ServerActionsTestOtherA-1581037736 tempest-ServerActionsTestOtherA-1581037736-project-member] Acquiring lock "refresh_cache-af5feda7-204a-47f1-b7fa-afe44832b9a2" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 592.520719] env[61923]: DEBUG oslo_concurrency.lockutils [None req-58d8b214-0b32-4993-bdd8-d652de7f9397 tempest-ServerActionsTestOtherA-1581037736 tempest-ServerActionsTestOtherA-1581037736-project-member] Acquired lock "refresh_cache-af5feda7-204a-47f1-b7fa-afe44832b9a2" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 592.520719] env[61923]: DEBUG nova.network.neutron [None req-58d8b214-0b32-4993-bdd8-d652de7f9397 tempest-ServerActionsTestOtherA-1581037736 tempest-ServerActionsTestOtherA-1581037736-project-member] [instance: af5feda7-204a-47f1-b7fa-afe44832b9a2] Building network info cache for instance {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 592.969055] env[61923]: DEBUG nova.scheduler.client.report [None req-595ebb22-536b-4375-9cbe-6fc472a5a17d tempest-ListServerFiltersTestJSON-483030130 tempest-ListServerFiltersTestJSON-483030130-project-member] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 593.016065] env[61923]: DEBUG nova.network.neutron [req-6fbf57b9-e432-4670-95ad-3d7df4a4c8b4 req-3687c9f6-a4f8-42b4-95ca-b59ec280d1d3 service nova] [instance: 045c0b05-333e-4f95-94b0-8f51d87dca0a] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 593.042017] env[61923]: DEBUG nova.network.neutron [None req-58d8b214-0b32-4993-bdd8-d652de7f9397 tempest-ServerActionsTestOtherA-1581037736 tempest-ServerActionsTestOtherA-1581037736-project-member] [instance: af5feda7-204a-47f1-b7fa-afe44832b9a2] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 593.126530] env[61923]: INFO nova.scheduler.client.report [None req-15b9e1af-fe57-407e-a058-c08239279239 tempest-ServersWithSpecificFlavorTestJSON-1806421899 tempest-ServersWithSpecificFlavorTestJSON-1806421899-project-member] Deleted allocations for instance 4cec7bce-0136-43ad-898f-086058a505f2 [ 593.160842] env[61923]: DEBUG nova.network.neutron [None req-58d8b214-0b32-4993-bdd8-d652de7f9397 tempest-ServerActionsTestOtherA-1581037736 tempest-ServerActionsTestOtherA-1581037736-project-member] [instance: af5feda7-204a-47f1-b7fa-afe44832b9a2] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 593.254852] env[61923]: DEBUG nova.network.neutron [req-6fbf57b9-e432-4670-95ad-3d7df4a4c8b4 req-3687c9f6-a4f8-42b4-95ca-b59ec280d1d3 service nova] [instance: 045c0b05-333e-4f95-94b0-8f51d87dca0a] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 593.475638] env[61923]: DEBUG oslo_concurrency.lockutils [None req-595ebb22-536b-4375-9cbe-6fc472a5a17d tempest-ListServerFiltersTestJSON-483030130 tempest-ListServerFiltersTestJSON-483030130-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.593s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 593.476213] env[61923]: DEBUG nova.compute.manager [None req-595ebb22-536b-4375-9cbe-6fc472a5a17d tempest-ListServerFiltersTestJSON-483030130 tempest-ListServerFiltersTestJSON-483030130-project-member] [instance: 71d9eccc-3b90-4b0a-a59a-d6a8e08a9417] Start building networks asynchronously for instance. {{(pid=61923) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 593.481011] env[61923]: DEBUG oslo_concurrency.lockutils [None req-03cb7270-244b-47b9-9dd5-e358ebdfa946 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.593s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 593.482514] env[61923]: INFO nova.compute.claims [None req-03cb7270-244b-47b9-9dd5-e358ebdfa946 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: cb998da7-8db0-4725-bed2-657d7df748b5] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 593.643736] env[61923]: DEBUG oslo_concurrency.lockutils [None req-15b9e1af-fe57-407e-a058-c08239279239 tempest-ServersWithSpecificFlavorTestJSON-1806421899 tempest-ServersWithSpecificFlavorTestJSON-1806421899-project-member] Lock "4cec7bce-0136-43ad-898f-086058a505f2" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 57.577s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 593.665276] env[61923]: DEBUG oslo_concurrency.lockutils [None req-58d8b214-0b32-4993-bdd8-d652de7f9397 tempest-ServerActionsTestOtherA-1581037736 tempest-ServerActionsTestOtherA-1581037736-project-member] Releasing lock "refresh_cache-af5feda7-204a-47f1-b7fa-afe44832b9a2" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 593.666024] env[61923]: DEBUG nova.compute.manager [None req-58d8b214-0b32-4993-bdd8-d652de7f9397 tempest-ServerActionsTestOtherA-1581037736 tempest-ServerActionsTestOtherA-1581037736-project-member] [instance: af5feda7-204a-47f1-b7fa-afe44832b9a2] Start destroying the instance on the hypervisor. {{(pid=61923) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 593.666126] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-58d8b214-0b32-4993-bdd8-d652de7f9397 tempest-ServerActionsTestOtherA-1581037736 tempest-ServerActionsTestOtherA-1581037736-project-member] [instance: af5feda7-204a-47f1-b7fa-afe44832b9a2] Destroying instance {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 593.667173] env[61923]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-03897e6b-db9f-409f-9744-b15f0cd8c394 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 593.688288] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af50a16a-da8d-41ed-9a2c-b9dd1ffcfa46 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 593.721978] env[61923]: WARNING nova.virt.vmwareapi.vmops [None req-58d8b214-0b32-4993-bdd8-d652de7f9397 tempest-ServerActionsTestOtherA-1581037736 tempest-ServerActionsTestOtherA-1581037736-project-member] [instance: af5feda7-204a-47f1-b7fa-afe44832b9a2] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance af5feda7-204a-47f1-b7fa-afe44832b9a2 could not be found. [ 593.722245] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-58d8b214-0b32-4993-bdd8-d652de7f9397 tempest-ServerActionsTestOtherA-1581037736 tempest-ServerActionsTestOtherA-1581037736-project-member] [instance: af5feda7-204a-47f1-b7fa-afe44832b9a2] Instance destroyed {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 593.722430] env[61923]: INFO nova.compute.manager [None req-58d8b214-0b32-4993-bdd8-d652de7f9397 tempest-ServerActionsTestOtherA-1581037736 tempest-ServerActionsTestOtherA-1581037736-project-member] [instance: af5feda7-204a-47f1-b7fa-afe44832b9a2] Took 0.06 seconds to destroy the instance on the hypervisor. [ 593.722690] env[61923]: DEBUG oslo.service.loopingcall [None req-58d8b214-0b32-4993-bdd8-d652de7f9397 tempest-ServerActionsTestOtherA-1581037736 tempest-ServerActionsTestOtherA-1581037736-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61923) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 593.722949] env[61923]: DEBUG nova.compute.manager [-] [instance: af5feda7-204a-47f1-b7fa-afe44832b9a2] Deallocating network for instance {{(pid=61923) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 593.723079] env[61923]: DEBUG nova.network.neutron [-] [instance: af5feda7-204a-47f1-b7fa-afe44832b9a2] deallocate_for_instance() {{(pid=61923) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 593.759255] env[61923]: DEBUG oslo_concurrency.lockutils [req-6fbf57b9-e432-4670-95ad-3d7df4a4c8b4 req-3687c9f6-a4f8-42b4-95ca-b59ec280d1d3 service nova] Releasing lock "refresh_cache-045c0b05-333e-4f95-94b0-8f51d87dca0a" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 593.759688] env[61923]: DEBUG nova.compute.manager [req-6fbf57b9-e432-4670-95ad-3d7df4a4c8b4 req-3687c9f6-a4f8-42b4-95ca-b59ec280d1d3 service nova] [instance: 045c0b05-333e-4f95-94b0-8f51d87dca0a] Received event network-vif-deleted-3d9f5735-20ed-42f4-9565-073437eef966 {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 593.760918] env[61923]: DEBUG nova.network.neutron [-] [instance: af5feda7-204a-47f1-b7fa-afe44832b9a2] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 593.986720] env[61923]: DEBUG nova.compute.utils [None req-595ebb22-536b-4375-9cbe-6fc472a5a17d tempest-ListServerFiltersTestJSON-483030130 tempest-ListServerFiltersTestJSON-483030130-project-member] Using /dev/sd instead of None {{(pid=61923) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 593.996427] env[61923]: DEBUG nova.compute.manager [None req-595ebb22-536b-4375-9cbe-6fc472a5a17d tempest-ListServerFiltersTestJSON-483030130 tempest-ListServerFiltersTestJSON-483030130-project-member] [instance: 71d9eccc-3b90-4b0a-a59a-d6a8e08a9417] Allocating IP information in the background. {{(pid=61923) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 593.997014] env[61923]: DEBUG nova.network.neutron [None req-595ebb22-536b-4375-9cbe-6fc472a5a17d tempest-ListServerFiltersTestJSON-483030130 tempest-ListServerFiltersTestJSON-483030130-project-member] [instance: 71d9eccc-3b90-4b0a-a59a-d6a8e08a9417] allocate_for_instance() {{(pid=61923) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 594.103225] env[61923]: DEBUG nova.policy [None req-595ebb22-536b-4375-9cbe-6fc472a5a17d tempest-ListServerFiltersTestJSON-483030130 tempest-ListServerFiltersTestJSON-483030130-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '16836760c008401586e5b67b19a53cc0', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '6c5ff14dc18843cc875b21b949eca824', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61923) authorize /opt/stack/nova/nova/policy.py:201}} [ 594.147235] env[61923]: DEBUG nova.compute.manager [None req-b2fe22fb-da8f-41ba-8845-95fc6177390e tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] [instance: 119bb4a7-2d71-44d7-ae81-9238512937c8] Starting instance... {{(pid=61923) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 594.264036] env[61923]: DEBUG nova.network.neutron [-] [instance: af5feda7-204a-47f1-b7fa-afe44832b9a2] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 594.341394] env[61923]: DEBUG oslo_concurrency.lockutils [None req-c09a27e9-8e33-411f-b17b-7cd8d66a175c tempest-ServerMetadataNegativeTestJSON-1976245157 tempest-ServerMetadataNegativeTestJSON-1976245157-project-member] Acquiring lock "25b62db6-6e64-4e38-bfd3-82a2e9a4883a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 594.341837] env[61923]: DEBUG oslo_concurrency.lockutils [None req-c09a27e9-8e33-411f-b17b-7cd8d66a175c tempest-ServerMetadataNegativeTestJSON-1976245157 tempest-ServerMetadataNegativeTestJSON-1976245157-project-member] Lock "25b62db6-6e64-4e38-bfd3-82a2e9a4883a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 594.496570] env[61923]: DEBUG nova.compute.manager [None req-595ebb22-536b-4375-9cbe-6fc472a5a17d tempest-ListServerFiltersTestJSON-483030130 tempest-ListServerFiltersTestJSON-483030130-project-member] [instance: 71d9eccc-3b90-4b0a-a59a-d6a8e08a9417] Start building block device mappings for instance. {{(pid=61923) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 594.680627] env[61923]: DEBUG oslo_concurrency.lockutils [None req-b2fe22fb-da8f-41ba-8845-95fc6177390e tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 594.732146] env[61923]: DEBUG nova.network.neutron [None req-595ebb22-536b-4375-9cbe-6fc472a5a17d tempest-ListServerFiltersTestJSON-483030130 tempest-ListServerFiltersTestJSON-483030130-project-member] [instance: 71d9eccc-3b90-4b0a-a59a-d6a8e08a9417] Successfully created port: b42828d3-f036-4a41-949a-3505532203d9 {{(pid=61923) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 594.766523] env[61923]: INFO nova.compute.manager [-] [instance: af5feda7-204a-47f1-b7fa-afe44832b9a2] Took 1.04 seconds to deallocate network for instance. [ 594.769670] env[61923]: DEBUG nova.compute.claims [None req-58d8b214-0b32-4993-bdd8-d652de7f9397 tempest-ServerActionsTestOtherA-1581037736 tempest-ServerActionsTestOtherA-1581037736-project-member] [instance: af5feda7-204a-47f1-b7fa-afe44832b9a2] Aborting claim: {{(pid=61923) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 594.769670] env[61923]: DEBUG oslo_concurrency.lockutils [None req-58d8b214-0b32-4993-bdd8-d652de7f9397 tempest-ServerActionsTestOtherA-1581037736 tempest-ServerActionsTestOtherA-1581037736-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 594.977030] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5251394c-1aeb-471f-8811-c7b1c256e9ac {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 594.986854] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c5f2f1e-2f00-4584-a3af-f5a95eae0aa4 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 595.026024] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa1cad2d-e427-44cc-8d28-3cde0bf634bd {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 595.035172] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f9e46e4-d717-40e1-af33-85fc9aea8e2c {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 595.054427] env[61923]: DEBUG nova.compute.provider_tree [None req-03cb7270-244b-47b9-9dd5-e358ebdfa946 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 595.464250] env[61923]: DEBUG nova.compute.manager [req-d3188566-a2f5-457a-99f0-e18d5a13b54d req-55a80664-2e3d-4e3f-8923-25da16100618 service nova] [instance: af5feda7-204a-47f1-b7fa-afe44832b9a2] Received event network-changed-1b61be5f-12b3-4ed1-a87e-41558aedeada {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 595.464250] env[61923]: DEBUG nova.compute.manager [req-d3188566-a2f5-457a-99f0-e18d5a13b54d req-55a80664-2e3d-4e3f-8923-25da16100618 service nova] [instance: af5feda7-204a-47f1-b7fa-afe44832b9a2] Refreshing instance network info cache due to event network-changed-1b61be5f-12b3-4ed1-a87e-41558aedeada. {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 595.464423] env[61923]: DEBUG oslo_concurrency.lockutils [req-d3188566-a2f5-457a-99f0-e18d5a13b54d req-55a80664-2e3d-4e3f-8923-25da16100618 service nova] Acquiring lock "refresh_cache-af5feda7-204a-47f1-b7fa-afe44832b9a2" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 595.466879] env[61923]: DEBUG oslo_concurrency.lockutils [req-d3188566-a2f5-457a-99f0-e18d5a13b54d req-55a80664-2e3d-4e3f-8923-25da16100618 service nova] Acquired lock "refresh_cache-af5feda7-204a-47f1-b7fa-afe44832b9a2" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 595.466879] env[61923]: DEBUG nova.network.neutron [req-d3188566-a2f5-457a-99f0-e18d5a13b54d req-55a80664-2e3d-4e3f-8923-25da16100618 service nova] [instance: af5feda7-204a-47f1-b7fa-afe44832b9a2] Refreshing network info cache for port 1b61be5f-12b3-4ed1-a87e-41558aedeada {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 595.533137] env[61923]: DEBUG nova.compute.manager [None req-595ebb22-536b-4375-9cbe-6fc472a5a17d tempest-ListServerFiltersTestJSON-483030130 tempest-ListServerFiltersTestJSON-483030130-project-member] [instance: 71d9eccc-3b90-4b0a-a59a-d6a8e08a9417] Start spawning the instance on the hypervisor. {{(pid=61923) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 595.557658] env[61923]: DEBUG nova.virt.hardware [None req-595ebb22-536b-4375-9cbe-6fc472a5a17d tempest-ListServerFiltersTestJSON-483030130 tempest-ListServerFiltersTestJSON-483030130-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-29T20:07:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-29T20:07:35Z,direct_url=,disk_format='vmdk',id=0f153f63-ae0a-45b1-b7f5-7f9b673c947f,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='4e7b5e3b3c3443c8bd5c70f8a15739f5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-29T20:07:36Z,virtual_size=,visibility=), allow threads: False {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 595.561019] env[61923]: DEBUG nova.virt.hardware [None req-595ebb22-536b-4375-9cbe-6fc472a5a17d tempest-ListServerFiltersTestJSON-483030130 tempest-ListServerFiltersTestJSON-483030130-project-member] Flavor limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 595.561019] env[61923]: DEBUG nova.virt.hardware [None req-595ebb22-536b-4375-9cbe-6fc472a5a17d tempest-ListServerFiltersTestJSON-483030130 tempest-ListServerFiltersTestJSON-483030130-project-member] Image limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 595.561019] env[61923]: DEBUG nova.virt.hardware [None req-595ebb22-536b-4375-9cbe-6fc472a5a17d tempest-ListServerFiltersTestJSON-483030130 tempest-ListServerFiltersTestJSON-483030130-project-member] Flavor pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 595.561019] env[61923]: DEBUG nova.virt.hardware [None req-595ebb22-536b-4375-9cbe-6fc472a5a17d tempest-ListServerFiltersTestJSON-483030130 tempest-ListServerFiltersTestJSON-483030130-project-member] Image pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 595.561019] env[61923]: DEBUG nova.virt.hardware [None req-595ebb22-536b-4375-9cbe-6fc472a5a17d tempest-ListServerFiltersTestJSON-483030130 tempest-ListServerFiltersTestJSON-483030130-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 595.561398] env[61923]: DEBUG nova.virt.hardware [None req-595ebb22-536b-4375-9cbe-6fc472a5a17d tempest-ListServerFiltersTestJSON-483030130 tempest-ListServerFiltersTestJSON-483030130-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 595.561398] env[61923]: DEBUG nova.virt.hardware [None req-595ebb22-536b-4375-9cbe-6fc472a5a17d tempest-ListServerFiltersTestJSON-483030130 tempest-ListServerFiltersTestJSON-483030130-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 595.561398] env[61923]: DEBUG nova.virt.hardware [None req-595ebb22-536b-4375-9cbe-6fc472a5a17d tempest-ListServerFiltersTestJSON-483030130 tempest-ListServerFiltersTestJSON-483030130-project-member] Got 1 possible topologies {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 595.561398] env[61923]: DEBUG nova.virt.hardware [None req-595ebb22-536b-4375-9cbe-6fc472a5a17d tempest-ListServerFiltersTestJSON-483030130 tempest-ListServerFiltersTestJSON-483030130-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 595.561398] env[61923]: DEBUG nova.virt.hardware [None req-595ebb22-536b-4375-9cbe-6fc472a5a17d tempest-ListServerFiltersTestJSON-483030130 tempest-ListServerFiltersTestJSON-483030130-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 595.561598] env[61923]: DEBUG nova.scheduler.client.report [None req-03cb7270-244b-47b9-9dd5-e358ebdfa946 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 595.565795] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-757a21ff-c219-4918-8f60-a50e03d40c06 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 595.575924] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c70bff8-cbc5-4741-990d-4ca52d9377f9 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.072194] env[61923]: DEBUG oslo_concurrency.lockutils [None req-03cb7270-244b-47b9-9dd5-e358ebdfa946 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.590s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 596.072194] env[61923]: DEBUG nova.compute.manager [None req-03cb7270-244b-47b9-9dd5-e358ebdfa946 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: cb998da7-8db0-4725-bed2-657d7df748b5] Start building networks asynchronously for instance. {{(pid=61923) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 596.078021] env[61923]: DEBUG oslo_concurrency.lockutils [None req-36da9865-5d98-4c2a-a964-964bee74523c tempest-ListServerFiltersTestJSON-483030130 tempest-ListServerFiltersTestJSON-483030130-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 17.710s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 596.087600] env[61923]: DEBUG oslo_concurrency.lockutils [None req-24801cc4-3974-42da-978f-a22cb8f70af0 tempest-SecurityGroupsTestJSON-509490211 tempest-SecurityGroupsTestJSON-509490211-project-member] Acquiring lock "2939d1e5-07fe-48b3-bc87-5f67fa1a7ae4" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 596.087836] env[61923]: DEBUG oslo_concurrency.lockutils [None req-24801cc4-3974-42da-978f-a22cb8f70af0 tempest-SecurityGroupsTestJSON-509490211 tempest-SecurityGroupsTestJSON-509490211-project-member] Lock "2939d1e5-07fe-48b3-bc87-5f67fa1a7ae4" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 596.126894] env[61923]: DEBUG nova.network.neutron [req-d3188566-a2f5-457a-99f0-e18d5a13b54d req-55a80664-2e3d-4e3f-8923-25da16100618 service nova] [instance: af5feda7-204a-47f1-b7fa-afe44832b9a2] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 596.388629] env[61923]: DEBUG nova.network.neutron [req-d3188566-a2f5-457a-99f0-e18d5a13b54d req-55a80664-2e3d-4e3f-8923-25da16100618 service nova] [instance: af5feda7-204a-47f1-b7fa-afe44832b9a2] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 596.579321] env[61923]: ERROR nova.compute.manager [None req-595ebb22-536b-4375-9cbe-6fc472a5a17d tempest-ListServerFiltersTestJSON-483030130 tempest-ListServerFiltersTestJSON-483030130-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port b42828d3-f036-4a41-949a-3505532203d9, please check neutron logs for more information. [ 596.579321] env[61923]: ERROR nova.compute.manager Traceback (most recent call last): [ 596.579321] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 596.579321] env[61923]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 596.579321] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 596.579321] env[61923]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 596.579321] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 596.579321] env[61923]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 596.579321] env[61923]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 596.579321] env[61923]: ERROR nova.compute.manager self.force_reraise() [ 596.579321] env[61923]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 596.579321] env[61923]: ERROR nova.compute.manager raise self.value [ 596.579321] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 596.579321] env[61923]: ERROR nova.compute.manager updated_port = self._update_port( [ 596.579321] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 596.579321] env[61923]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 596.581124] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 596.581124] env[61923]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 596.581124] env[61923]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port b42828d3-f036-4a41-949a-3505532203d9, please check neutron logs for more information. [ 596.581124] env[61923]: ERROR nova.compute.manager [ 596.581124] env[61923]: Traceback (most recent call last): [ 596.581124] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 596.581124] env[61923]: listener.cb(fileno) [ 596.581124] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 596.581124] env[61923]: result = function(*args, **kwargs) [ 596.581124] env[61923]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 596.581124] env[61923]: return func(*args, **kwargs) [ 596.581124] env[61923]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 596.581124] env[61923]: raise e [ 596.581124] env[61923]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 596.581124] env[61923]: nwinfo = self.network_api.allocate_for_instance( [ 596.581124] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 596.581124] env[61923]: created_port_ids = self._update_ports_for_instance( [ 596.581124] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 596.581124] env[61923]: with excutils.save_and_reraise_exception(): [ 596.581124] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 596.581124] env[61923]: self.force_reraise() [ 596.581124] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 596.581124] env[61923]: raise self.value [ 596.581124] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 596.581124] env[61923]: updated_port = self._update_port( [ 596.581124] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 596.581124] env[61923]: _ensure_no_port_binding_failure(port) [ 596.581124] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 596.581124] env[61923]: raise exception.PortBindingFailed(port_id=port['id']) [ 596.583484] env[61923]: nova.exception.PortBindingFailed: Binding failed for port b42828d3-f036-4a41-949a-3505532203d9, please check neutron logs for more information. [ 596.583484] env[61923]: Removing descriptor: 17 [ 596.583484] env[61923]: ERROR nova.compute.manager [None req-595ebb22-536b-4375-9cbe-6fc472a5a17d tempest-ListServerFiltersTestJSON-483030130 tempest-ListServerFiltersTestJSON-483030130-project-member] [instance: 71d9eccc-3b90-4b0a-a59a-d6a8e08a9417] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port b42828d3-f036-4a41-949a-3505532203d9, please check neutron logs for more information. [ 596.583484] env[61923]: ERROR nova.compute.manager [instance: 71d9eccc-3b90-4b0a-a59a-d6a8e08a9417] Traceback (most recent call last): [ 596.583484] env[61923]: ERROR nova.compute.manager [instance: 71d9eccc-3b90-4b0a-a59a-d6a8e08a9417] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 596.583484] env[61923]: ERROR nova.compute.manager [instance: 71d9eccc-3b90-4b0a-a59a-d6a8e08a9417] yield resources [ 596.583484] env[61923]: ERROR nova.compute.manager [instance: 71d9eccc-3b90-4b0a-a59a-d6a8e08a9417] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 596.583484] env[61923]: ERROR nova.compute.manager [instance: 71d9eccc-3b90-4b0a-a59a-d6a8e08a9417] self.driver.spawn(context, instance, image_meta, [ 596.583484] env[61923]: ERROR nova.compute.manager [instance: 71d9eccc-3b90-4b0a-a59a-d6a8e08a9417] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 596.583484] env[61923]: ERROR nova.compute.manager [instance: 71d9eccc-3b90-4b0a-a59a-d6a8e08a9417] self._vmops.spawn(context, instance, image_meta, injected_files, [ 596.583484] env[61923]: ERROR nova.compute.manager [instance: 71d9eccc-3b90-4b0a-a59a-d6a8e08a9417] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 596.583484] env[61923]: ERROR nova.compute.manager [instance: 71d9eccc-3b90-4b0a-a59a-d6a8e08a9417] vm_ref = self.build_virtual_machine(instance, [ 596.584145] env[61923]: ERROR nova.compute.manager [instance: 71d9eccc-3b90-4b0a-a59a-d6a8e08a9417] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 596.584145] env[61923]: ERROR nova.compute.manager [instance: 71d9eccc-3b90-4b0a-a59a-d6a8e08a9417] vif_infos = vmwarevif.get_vif_info(self._session, [ 596.584145] env[61923]: ERROR nova.compute.manager [instance: 71d9eccc-3b90-4b0a-a59a-d6a8e08a9417] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 596.584145] env[61923]: ERROR nova.compute.manager [instance: 71d9eccc-3b90-4b0a-a59a-d6a8e08a9417] for vif in network_info: [ 596.584145] env[61923]: ERROR nova.compute.manager [instance: 71d9eccc-3b90-4b0a-a59a-d6a8e08a9417] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 596.584145] env[61923]: ERROR nova.compute.manager [instance: 71d9eccc-3b90-4b0a-a59a-d6a8e08a9417] return self._sync_wrapper(fn, *args, **kwargs) [ 596.584145] env[61923]: ERROR nova.compute.manager [instance: 71d9eccc-3b90-4b0a-a59a-d6a8e08a9417] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 596.584145] env[61923]: ERROR nova.compute.manager [instance: 71d9eccc-3b90-4b0a-a59a-d6a8e08a9417] self.wait() [ 596.584145] env[61923]: ERROR nova.compute.manager [instance: 71d9eccc-3b90-4b0a-a59a-d6a8e08a9417] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 596.584145] env[61923]: ERROR nova.compute.manager [instance: 71d9eccc-3b90-4b0a-a59a-d6a8e08a9417] self[:] = self._gt.wait() [ 596.584145] env[61923]: ERROR nova.compute.manager [instance: 71d9eccc-3b90-4b0a-a59a-d6a8e08a9417] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 596.584145] env[61923]: ERROR nova.compute.manager [instance: 71d9eccc-3b90-4b0a-a59a-d6a8e08a9417] return self._exit_event.wait() [ 596.584145] env[61923]: ERROR nova.compute.manager [instance: 71d9eccc-3b90-4b0a-a59a-d6a8e08a9417] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 596.584537] env[61923]: ERROR nova.compute.manager [instance: 71d9eccc-3b90-4b0a-a59a-d6a8e08a9417] result = hub.switch() [ 596.584537] env[61923]: ERROR nova.compute.manager [instance: 71d9eccc-3b90-4b0a-a59a-d6a8e08a9417] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 596.584537] env[61923]: ERROR nova.compute.manager [instance: 71d9eccc-3b90-4b0a-a59a-d6a8e08a9417] return self.greenlet.switch() [ 596.584537] env[61923]: ERROR nova.compute.manager [instance: 71d9eccc-3b90-4b0a-a59a-d6a8e08a9417] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 596.584537] env[61923]: ERROR nova.compute.manager [instance: 71d9eccc-3b90-4b0a-a59a-d6a8e08a9417] result = function(*args, **kwargs) [ 596.584537] env[61923]: ERROR nova.compute.manager [instance: 71d9eccc-3b90-4b0a-a59a-d6a8e08a9417] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 596.584537] env[61923]: ERROR nova.compute.manager [instance: 71d9eccc-3b90-4b0a-a59a-d6a8e08a9417] return func(*args, **kwargs) [ 596.584537] env[61923]: ERROR nova.compute.manager [instance: 71d9eccc-3b90-4b0a-a59a-d6a8e08a9417] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 596.584537] env[61923]: ERROR nova.compute.manager [instance: 71d9eccc-3b90-4b0a-a59a-d6a8e08a9417] raise e [ 596.584537] env[61923]: ERROR nova.compute.manager [instance: 71d9eccc-3b90-4b0a-a59a-d6a8e08a9417] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 596.584537] env[61923]: ERROR nova.compute.manager [instance: 71d9eccc-3b90-4b0a-a59a-d6a8e08a9417] nwinfo = self.network_api.allocate_for_instance( [ 596.584537] env[61923]: ERROR nova.compute.manager [instance: 71d9eccc-3b90-4b0a-a59a-d6a8e08a9417] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 596.584537] env[61923]: ERROR nova.compute.manager [instance: 71d9eccc-3b90-4b0a-a59a-d6a8e08a9417] created_port_ids = self._update_ports_for_instance( [ 596.584917] env[61923]: ERROR nova.compute.manager [instance: 71d9eccc-3b90-4b0a-a59a-d6a8e08a9417] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 596.584917] env[61923]: ERROR nova.compute.manager [instance: 71d9eccc-3b90-4b0a-a59a-d6a8e08a9417] with excutils.save_and_reraise_exception(): [ 596.584917] env[61923]: ERROR nova.compute.manager [instance: 71d9eccc-3b90-4b0a-a59a-d6a8e08a9417] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 596.584917] env[61923]: ERROR nova.compute.manager [instance: 71d9eccc-3b90-4b0a-a59a-d6a8e08a9417] self.force_reraise() [ 596.584917] env[61923]: ERROR nova.compute.manager [instance: 71d9eccc-3b90-4b0a-a59a-d6a8e08a9417] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 596.584917] env[61923]: ERROR nova.compute.manager [instance: 71d9eccc-3b90-4b0a-a59a-d6a8e08a9417] raise self.value [ 596.584917] env[61923]: ERROR nova.compute.manager [instance: 71d9eccc-3b90-4b0a-a59a-d6a8e08a9417] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 596.584917] env[61923]: ERROR nova.compute.manager [instance: 71d9eccc-3b90-4b0a-a59a-d6a8e08a9417] updated_port = self._update_port( [ 596.584917] env[61923]: ERROR nova.compute.manager [instance: 71d9eccc-3b90-4b0a-a59a-d6a8e08a9417] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 596.584917] env[61923]: ERROR nova.compute.manager [instance: 71d9eccc-3b90-4b0a-a59a-d6a8e08a9417] _ensure_no_port_binding_failure(port) [ 596.584917] env[61923]: ERROR nova.compute.manager [instance: 71d9eccc-3b90-4b0a-a59a-d6a8e08a9417] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 596.584917] env[61923]: ERROR nova.compute.manager [instance: 71d9eccc-3b90-4b0a-a59a-d6a8e08a9417] raise exception.PortBindingFailed(port_id=port['id']) [ 596.585483] env[61923]: ERROR nova.compute.manager [instance: 71d9eccc-3b90-4b0a-a59a-d6a8e08a9417] nova.exception.PortBindingFailed: Binding failed for port b42828d3-f036-4a41-949a-3505532203d9, please check neutron logs for more information. [ 596.585483] env[61923]: ERROR nova.compute.manager [instance: 71d9eccc-3b90-4b0a-a59a-d6a8e08a9417] [ 596.585483] env[61923]: INFO nova.compute.manager [None req-595ebb22-536b-4375-9cbe-6fc472a5a17d tempest-ListServerFiltersTestJSON-483030130 tempest-ListServerFiltersTestJSON-483030130-project-member] [instance: 71d9eccc-3b90-4b0a-a59a-d6a8e08a9417] Terminating instance [ 596.585483] env[61923]: DEBUG oslo_concurrency.lockutils [None req-595ebb22-536b-4375-9cbe-6fc472a5a17d tempest-ListServerFiltersTestJSON-483030130 tempest-ListServerFiltersTestJSON-483030130-project-member] Acquiring lock "refresh_cache-71d9eccc-3b90-4b0a-a59a-d6a8e08a9417" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 596.585483] env[61923]: DEBUG oslo_concurrency.lockutils [None req-595ebb22-536b-4375-9cbe-6fc472a5a17d tempest-ListServerFiltersTestJSON-483030130 tempest-ListServerFiltersTestJSON-483030130-project-member] Acquired lock "refresh_cache-71d9eccc-3b90-4b0a-a59a-d6a8e08a9417" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 596.585483] env[61923]: DEBUG nova.network.neutron [None req-595ebb22-536b-4375-9cbe-6fc472a5a17d tempest-ListServerFiltersTestJSON-483030130 tempest-ListServerFiltersTestJSON-483030130-project-member] [instance: 71d9eccc-3b90-4b0a-a59a-d6a8e08a9417] Building network info cache for instance {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 596.585694] env[61923]: DEBUG nova.compute.utils [None req-03cb7270-244b-47b9-9dd5-e358ebdfa946 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Using /dev/sd instead of None {{(pid=61923) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 596.585694] env[61923]: DEBUG nova.compute.manager [None req-03cb7270-244b-47b9-9dd5-e358ebdfa946 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: cb998da7-8db0-4725-bed2-657d7df748b5] Allocating IP information in the background. {{(pid=61923) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 596.585694] env[61923]: DEBUG nova.network.neutron [None req-03cb7270-244b-47b9-9dd5-e358ebdfa946 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: cb998da7-8db0-4725-bed2-657d7df748b5] allocate_for_instance() {{(pid=61923) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 596.805987] env[61923]: DEBUG nova.policy [None req-03cb7270-244b-47b9-9dd5-e358ebdfa946 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5b8a639e99914e75857b4571f0d58a3c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '69cc941eb6dd4780ac12aa29656c37f7', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61923) authorize /opt/stack/nova/nova/policy.py:201}} [ 596.891647] env[61923]: DEBUG oslo_concurrency.lockutils [req-d3188566-a2f5-457a-99f0-e18d5a13b54d req-55a80664-2e3d-4e3f-8923-25da16100618 service nova] Releasing lock "refresh_cache-af5feda7-204a-47f1-b7fa-afe44832b9a2" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 596.892213] env[61923]: DEBUG nova.compute.manager [req-d3188566-a2f5-457a-99f0-e18d5a13b54d req-55a80664-2e3d-4e3f-8923-25da16100618 service nova] [instance: af5feda7-204a-47f1-b7fa-afe44832b9a2] Received event network-vif-deleted-1b61be5f-12b3-4ed1-a87e-41558aedeada {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 597.088228] env[61923]: DEBUG nova.compute.manager [None req-03cb7270-244b-47b9-9dd5-e358ebdfa946 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: cb998da7-8db0-4725-bed2-657d7df748b5] Start building block device mappings for instance. {{(pid=61923) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 597.119144] env[61923]: DEBUG nova.network.neutron [None req-595ebb22-536b-4375-9cbe-6fc472a5a17d tempest-ListServerFiltersTestJSON-483030130 tempest-ListServerFiltersTestJSON-483030130-project-member] [instance: 71d9eccc-3b90-4b0a-a59a-d6a8e08a9417] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 597.145720] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ca97b23-292f-4047-9014-a80c6316a9ed {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 597.154949] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6715d25-c2fc-4a50-9e70-e9ec4675aac9 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 597.198353] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68955fb8-ccbf-49ec-bdd7-36423b6bfdc8 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 597.211455] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f697aa1e-fe98-466f-9c21-8f6c89fbbb17 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 597.234753] env[61923]: DEBUG nova.compute.provider_tree [None req-36da9865-5d98-4c2a-a964-964bee74523c tempest-ListServerFiltersTestJSON-483030130 tempest-ListServerFiltersTestJSON-483030130-project-member] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 597.602988] env[61923]: DEBUG nova.network.neutron [None req-03cb7270-244b-47b9-9dd5-e358ebdfa946 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: cb998da7-8db0-4725-bed2-657d7df748b5] Successfully created port: 225a6542-d2aa-4f13-a9d6-7c2e6c048241 {{(pid=61923) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 597.623629] env[61923]: DEBUG nova.network.neutron [None req-595ebb22-536b-4375-9cbe-6fc472a5a17d tempest-ListServerFiltersTestJSON-483030130 tempest-ListServerFiltersTestJSON-483030130-project-member] [instance: 71d9eccc-3b90-4b0a-a59a-d6a8e08a9417] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 597.715518] env[61923]: DEBUG nova.compute.manager [req-08acf91d-c666-49e1-9ada-0e524b52033e req-5f720f47-292b-4854-a620-39079a86a57a service nova] [instance: 71d9eccc-3b90-4b0a-a59a-d6a8e08a9417] Received event network-changed-b42828d3-f036-4a41-949a-3505532203d9 {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 597.715605] env[61923]: DEBUG nova.compute.manager [req-08acf91d-c666-49e1-9ada-0e524b52033e req-5f720f47-292b-4854-a620-39079a86a57a service nova] [instance: 71d9eccc-3b90-4b0a-a59a-d6a8e08a9417] Refreshing instance network info cache due to event network-changed-b42828d3-f036-4a41-949a-3505532203d9. {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 597.715802] env[61923]: DEBUG oslo_concurrency.lockutils [req-08acf91d-c666-49e1-9ada-0e524b52033e req-5f720f47-292b-4854-a620-39079a86a57a service nova] Acquiring lock "refresh_cache-71d9eccc-3b90-4b0a-a59a-d6a8e08a9417" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 597.741280] env[61923]: DEBUG nova.scheduler.client.report [None req-36da9865-5d98-4c2a-a964-964bee74523c tempest-ListServerFiltersTestJSON-483030130 tempest-ListServerFiltersTestJSON-483030130-project-member] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 598.103398] env[61923]: DEBUG nova.compute.manager [None req-03cb7270-244b-47b9-9dd5-e358ebdfa946 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: cb998da7-8db0-4725-bed2-657d7df748b5] Start spawning the instance on the hypervisor. {{(pid=61923) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 598.129021] env[61923]: DEBUG oslo_concurrency.lockutils [None req-595ebb22-536b-4375-9cbe-6fc472a5a17d tempest-ListServerFiltersTestJSON-483030130 tempest-ListServerFiltersTestJSON-483030130-project-member] Releasing lock "refresh_cache-71d9eccc-3b90-4b0a-a59a-d6a8e08a9417" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 598.129021] env[61923]: DEBUG nova.compute.manager [None req-595ebb22-536b-4375-9cbe-6fc472a5a17d tempest-ListServerFiltersTestJSON-483030130 tempest-ListServerFiltersTestJSON-483030130-project-member] [instance: 71d9eccc-3b90-4b0a-a59a-d6a8e08a9417] Start destroying the instance on the hypervisor. {{(pid=61923) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 598.129021] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-595ebb22-536b-4375-9cbe-6fc472a5a17d tempest-ListServerFiltersTestJSON-483030130 tempest-ListServerFiltersTestJSON-483030130-project-member] [instance: 71d9eccc-3b90-4b0a-a59a-d6a8e08a9417] Destroying instance {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 598.129021] env[61923]: DEBUG oslo_concurrency.lockutils [req-08acf91d-c666-49e1-9ada-0e524b52033e req-5f720f47-292b-4854-a620-39079a86a57a service nova] Acquired lock "refresh_cache-71d9eccc-3b90-4b0a-a59a-d6a8e08a9417" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 598.129021] env[61923]: DEBUG nova.network.neutron [req-08acf91d-c666-49e1-9ada-0e524b52033e req-5f720f47-292b-4854-a620-39079a86a57a service nova] [instance: 71d9eccc-3b90-4b0a-a59a-d6a8e08a9417] Refreshing network info cache for port b42828d3-f036-4a41-949a-3505532203d9 {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 598.129252] env[61923]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ab8cc81f-ed81-4f47-a80d-c3d1e8dbcce9 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 598.139326] env[61923]: DEBUG nova.virt.hardware [None req-03cb7270-244b-47b9-9dd5-e358ebdfa946 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-29T20:07:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-29T20:07:35Z,direct_url=,disk_format='vmdk',id=0f153f63-ae0a-45b1-b7f5-7f9b673c947f,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='4e7b5e3b3c3443c8bd5c70f8a15739f5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-29T20:07:36Z,virtual_size=,visibility=), allow threads: False {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 598.139564] env[61923]: DEBUG nova.virt.hardware [None req-03cb7270-244b-47b9-9dd5-e358ebdfa946 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Flavor limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 598.139720] env[61923]: DEBUG nova.virt.hardware [None req-03cb7270-244b-47b9-9dd5-e358ebdfa946 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Image limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 598.140087] env[61923]: DEBUG nova.virt.hardware [None req-03cb7270-244b-47b9-9dd5-e358ebdfa946 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Flavor pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 598.140087] env[61923]: DEBUG nova.virt.hardware [None req-03cb7270-244b-47b9-9dd5-e358ebdfa946 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Image pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 598.140223] env[61923]: DEBUG nova.virt.hardware [None req-03cb7270-244b-47b9-9dd5-e358ebdfa946 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 598.140374] env[61923]: DEBUG nova.virt.hardware [None req-03cb7270-244b-47b9-9dd5-e358ebdfa946 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 598.140521] env[61923]: DEBUG nova.virt.hardware [None req-03cb7270-244b-47b9-9dd5-e358ebdfa946 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 598.140745] env[61923]: DEBUG nova.virt.hardware [None req-03cb7270-244b-47b9-9dd5-e358ebdfa946 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Got 1 possible topologies {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 598.140825] env[61923]: DEBUG nova.virt.hardware [None req-03cb7270-244b-47b9-9dd5-e358ebdfa946 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 598.140984] env[61923]: DEBUG nova.virt.hardware [None req-03cb7270-244b-47b9-9dd5-e358ebdfa946 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 598.142167] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3cbab71a-94db-4892-8f95-f2d20c98e0b0 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 598.149419] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b57e49a8-d3b1-408e-ac9e-23f08298b410 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 598.170036] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb9c19fb-0eea-416e-be86-d440dda3809f {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 598.192191] env[61923]: WARNING nova.virt.vmwareapi.vmops [None req-595ebb22-536b-4375-9cbe-6fc472a5a17d tempest-ListServerFiltersTestJSON-483030130 tempest-ListServerFiltersTestJSON-483030130-project-member] [instance: 71d9eccc-3b90-4b0a-a59a-d6a8e08a9417] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 71d9eccc-3b90-4b0a-a59a-d6a8e08a9417 could not be found. [ 598.192434] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-595ebb22-536b-4375-9cbe-6fc472a5a17d tempest-ListServerFiltersTestJSON-483030130 tempest-ListServerFiltersTestJSON-483030130-project-member] [instance: 71d9eccc-3b90-4b0a-a59a-d6a8e08a9417] Instance destroyed {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 598.192684] env[61923]: INFO nova.compute.manager [None req-595ebb22-536b-4375-9cbe-6fc472a5a17d tempest-ListServerFiltersTestJSON-483030130 tempest-ListServerFiltersTestJSON-483030130-project-member] [instance: 71d9eccc-3b90-4b0a-a59a-d6a8e08a9417] Took 0.07 seconds to destroy the instance on the hypervisor. [ 598.192853] env[61923]: DEBUG oslo.service.loopingcall [None req-595ebb22-536b-4375-9cbe-6fc472a5a17d tempest-ListServerFiltersTestJSON-483030130 tempest-ListServerFiltersTestJSON-483030130-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61923) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 598.193091] env[61923]: DEBUG nova.compute.manager [-] [instance: 71d9eccc-3b90-4b0a-a59a-d6a8e08a9417] Deallocating network for instance {{(pid=61923) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 598.193189] env[61923]: DEBUG nova.network.neutron [-] [instance: 71d9eccc-3b90-4b0a-a59a-d6a8e08a9417] deallocate_for_instance() {{(pid=61923) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 598.229703] env[61923]: DEBUG nova.network.neutron [-] [instance: 71d9eccc-3b90-4b0a-a59a-d6a8e08a9417] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 598.243642] env[61923]: DEBUG oslo_concurrency.lockutils [None req-36da9865-5d98-4c2a-a964-964bee74523c tempest-ListServerFiltersTestJSON-483030130 tempest-ListServerFiltersTestJSON-483030130-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.169s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 598.244425] env[61923]: ERROR nova.compute.manager [None req-36da9865-5d98-4c2a-a964-964bee74523c tempest-ListServerFiltersTestJSON-483030130 tempest-ListServerFiltersTestJSON-483030130-project-member] [instance: f3f872ce-cbbe-4407-b4e0-3341bb598825] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 793a91d7-2744-40de-8e41-428eacfdaee8, please check neutron logs for more information. [ 598.244425] env[61923]: ERROR nova.compute.manager [instance: f3f872ce-cbbe-4407-b4e0-3341bb598825] Traceback (most recent call last): [ 598.244425] env[61923]: ERROR nova.compute.manager [instance: f3f872ce-cbbe-4407-b4e0-3341bb598825] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 598.244425] env[61923]: ERROR nova.compute.manager [instance: f3f872ce-cbbe-4407-b4e0-3341bb598825] self.driver.spawn(context, instance, image_meta, [ 598.244425] env[61923]: ERROR nova.compute.manager [instance: f3f872ce-cbbe-4407-b4e0-3341bb598825] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 598.244425] env[61923]: ERROR nova.compute.manager [instance: f3f872ce-cbbe-4407-b4e0-3341bb598825] self._vmops.spawn(context, instance, image_meta, injected_files, [ 598.244425] env[61923]: ERROR nova.compute.manager [instance: f3f872ce-cbbe-4407-b4e0-3341bb598825] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 598.244425] env[61923]: ERROR nova.compute.manager [instance: f3f872ce-cbbe-4407-b4e0-3341bb598825] vm_ref = self.build_virtual_machine(instance, [ 598.244425] env[61923]: ERROR nova.compute.manager [instance: f3f872ce-cbbe-4407-b4e0-3341bb598825] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 598.244425] env[61923]: ERROR nova.compute.manager [instance: f3f872ce-cbbe-4407-b4e0-3341bb598825] vif_infos = vmwarevif.get_vif_info(self._session, [ 598.244425] env[61923]: ERROR nova.compute.manager [instance: f3f872ce-cbbe-4407-b4e0-3341bb598825] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 598.244733] env[61923]: ERROR nova.compute.manager [instance: f3f872ce-cbbe-4407-b4e0-3341bb598825] for vif in network_info: [ 598.244733] env[61923]: ERROR nova.compute.manager [instance: f3f872ce-cbbe-4407-b4e0-3341bb598825] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 598.244733] env[61923]: ERROR nova.compute.manager [instance: f3f872ce-cbbe-4407-b4e0-3341bb598825] return self._sync_wrapper(fn, *args, **kwargs) [ 598.244733] env[61923]: ERROR nova.compute.manager [instance: f3f872ce-cbbe-4407-b4e0-3341bb598825] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 598.244733] env[61923]: ERROR nova.compute.manager [instance: f3f872ce-cbbe-4407-b4e0-3341bb598825] self.wait() [ 598.244733] env[61923]: ERROR nova.compute.manager [instance: f3f872ce-cbbe-4407-b4e0-3341bb598825] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 598.244733] env[61923]: ERROR nova.compute.manager [instance: f3f872ce-cbbe-4407-b4e0-3341bb598825] self[:] = self._gt.wait() [ 598.244733] env[61923]: ERROR nova.compute.manager [instance: f3f872ce-cbbe-4407-b4e0-3341bb598825] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 598.244733] env[61923]: ERROR nova.compute.manager [instance: f3f872ce-cbbe-4407-b4e0-3341bb598825] return self._exit_event.wait() [ 598.244733] env[61923]: ERROR nova.compute.manager [instance: f3f872ce-cbbe-4407-b4e0-3341bb598825] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 598.244733] env[61923]: ERROR nova.compute.manager [instance: f3f872ce-cbbe-4407-b4e0-3341bb598825] result = hub.switch() [ 598.244733] env[61923]: ERROR nova.compute.manager [instance: f3f872ce-cbbe-4407-b4e0-3341bb598825] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 598.244733] env[61923]: ERROR nova.compute.manager [instance: f3f872ce-cbbe-4407-b4e0-3341bb598825] return self.greenlet.switch() [ 598.245079] env[61923]: ERROR nova.compute.manager [instance: f3f872ce-cbbe-4407-b4e0-3341bb598825] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 598.245079] env[61923]: ERROR nova.compute.manager [instance: f3f872ce-cbbe-4407-b4e0-3341bb598825] result = function(*args, **kwargs) [ 598.245079] env[61923]: ERROR nova.compute.manager [instance: f3f872ce-cbbe-4407-b4e0-3341bb598825] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 598.245079] env[61923]: ERROR nova.compute.manager [instance: f3f872ce-cbbe-4407-b4e0-3341bb598825] return func(*args, **kwargs) [ 598.245079] env[61923]: ERROR nova.compute.manager [instance: f3f872ce-cbbe-4407-b4e0-3341bb598825] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 598.245079] env[61923]: ERROR nova.compute.manager [instance: f3f872ce-cbbe-4407-b4e0-3341bb598825] raise e [ 598.245079] env[61923]: ERROR nova.compute.manager [instance: f3f872ce-cbbe-4407-b4e0-3341bb598825] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 598.245079] env[61923]: ERROR nova.compute.manager [instance: f3f872ce-cbbe-4407-b4e0-3341bb598825] nwinfo = self.network_api.allocate_for_instance( [ 598.245079] env[61923]: ERROR nova.compute.manager [instance: f3f872ce-cbbe-4407-b4e0-3341bb598825] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 598.245079] env[61923]: ERROR nova.compute.manager [instance: f3f872ce-cbbe-4407-b4e0-3341bb598825] created_port_ids = self._update_ports_for_instance( [ 598.245079] env[61923]: ERROR nova.compute.manager [instance: f3f872ce-cbbe-4407-b4e0-3341bb598825] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 598.245079] env[61923]: ERROR nova.compute.manager [instance: f3f872ce-cbbe-4407-b4e0-3341bb598825] with excutils.save_and_reraise_exception(): [ 598.245079] env[61923]: ERROR nova.compute.manager [instance: f3f872ce-cbbe-4407-b4e0-3341bb598825] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 598.245380] env[61923]: ERROR nova.compute.manager [instance: f3f872ce-cbbe-4407-b4e0-3341bb598825] self.force_reraise() [ 598.245380] env[61923]: ERROR nova.compute.manager [instance: f3f872ce-cbbe-4407-b4e0-3341bb598825] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 598.245380] env[61923]: ERROR nova.compute.manager [instance: f3f872ce-cbbe-4407-b4e0-3341bb598825] raise self.value [ 598.245380] env[61923]: ERROR nova.compute.manager [instance: f3f872ce-cbbe-4407-b4e0-3341bb598825] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 598.245380] env[61923]: ERROR nova.compute.manager [instance: f3f872ce-cbbe-4407-b4e0-3341bb598825] updated_port = self._update_port( [ 598.245380] env[61923]: ERROR nova.compute.manager [instance: f3f872ce-cbbe-4407-b4e0-3341bb598825] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 598.245380] env[61923]: ERROR nova.compute.manager [instance: f3f872ce-cbbe-4407-b4e0-3341bb598825] _ensure_no_port_binding_failure(port) [ 598.245380] env[61923]: ERROR nova.compute.manager [instance: f3f872ce-cbbe-4407-b4e0-3341bb598825] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 598.245380] env[61923]: ERROR nova.compute.manager [instance: f3f872ce-cbbe-4407-b4e0-3341bb598825] raise exception.PortBindingFailed(port_id=port['id']) [ 598.245380] env[61923]: ERROR nova.compute.manager [instance: f3f872ce-cbbe-4407-b4e0-3341bb598825] nova.exception.PortBindingFailed: Binding failed for port 793a91d7-2744-40de-8e41-428eacfdaee8, please check neutron logs for more information. [ 598.245380] env[61923]: ERROR nova.compute.manager [instance: f3f872ce-cbbe-4407-b4e0-3341bb598825] [ 598.248139] env[61923]: DEBUG nova.compute.utils [None req-36da9865-5d98-4c2a-a964-964bee74523c tempest-ListServerFiltersTestJSON-483030130 tempest-ListServerFiltersTestJSON-483030130-project-member] [instance: f3f872ce-cbbe-4407-b4e0-3341bb598825] Binding failed for port 793a91d7-2744-40de-8e41-428eacfdaee8, please check neutron logs for more information. {{(pid=61923) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 598.248139] env[61923]: DEBUG oslo_concurrency.lockutils [None req-ae7ff058-a25b-42a3-af4b-0715677fedd4 tempest-ServersAdmin275Test-1076075580 tempest-ServersAdmin275Test-1076075580-project-admin] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 17.418s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 598.248139] env[61923]: DEBUG nova.objects.instance [None req-ae7ff058-a25b-42a3-af4b-0715677fedd4 tempest-ServersAdmin275Test-1076075580 tempest-ServersAdmin275Test-1076075580-project-admin] [instance: f7cc960c-e06a-4c58-9367-ec8771fe09d5] Trying to apply a migration context that does not seem to be set for this instance {{(pid=61923) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 598.251703] env[61923]: DEBUG nova.compute.manager [None req-36da9865-5d98-4c2a-a964-964bee74523c tempest-ListServerFiltersTestJSON-483030130 tempest-ListServerFiltersTestJSON-483030130-project-member] [instance: f3f872ce-cbbe-4407-b4e0-3341bb598825] Build of instance f3f872ce-cbbe-4407-b4e0-3341bb598825 was re-scheduled: Binding failed for port 793a91d7-2744-40de-8e41-428eacfdaee8, please check neutron logs for more information. {{(pid=61923) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 598.251703] env[61923]: DEBUG nova.compute.manager [None req-36da9865-5d98-4c2a-a964-964bee74523c tempest-ListServerFiltersTestJSON-483030130 tempest-ListServerFiltersTestJSON-483030130-project-member] [instance: f3f872ce-cbbe-4407-b4e0-3341bb598825] Unplugging VIFs for instance {{(pid=61923) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 598.251703] env[61923]: DEBUG oslo_concurrency.lockutils [None req-36da9865-5d98-4c2a-a964-964bee74523c tempest-ListServerFiltersTestJSON-483030130 tempest-ListServerFiltersTestJSON-483030130-project-member] Acquiring lock "refresh_cache-f3f872ce-cbbe-4407-b4e0-3341bb598825" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 598.251703] env[61923]: DEBUG oslo_concurrency.lockutils [None req-36da9865-5d98-4c2a-a964-964bee74523c tempest-ListServerFiltersTestJSON-483030130 tempest-ListServerFiltersTestJSON-483030130-project-member] Acquired lock "refresh_cache-f3f872ce-cbbe-4407-b4e0-3341bb598825" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 598.251948] env[61923]: DEBUG nova.network.neutron [None req-36da9865-5d98-4c2a-a964-964bee74523c tempest-ListServerFiltersTestJSON-483030130 tempest-ListServerFiltersTestJSON-483030130-project-member] [instance: f3f872ce-cbbe-4407-b4e0-3341bb598825] Building network info cache for instance {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 598.658236] env[61923]: DEBUG nova.network.neutron [req-08acf91d-c666-49e1-9ada-0e524b52033e req-5f720f47-292b-4854-a620-39079a86a57a service nova] [instance: 71d9eccc-3b90-4b0a-a59a-d6a8e08a9417] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 598.733966] env[61923]: DEBUG nova.network.neutron [-] [instance: 71d9eccc-3b90-4b0a-a59a-d6a8e08a9417] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 598.814448] env[61923]: DEBUG nova.network.neutron [None req-36da9865-5d98-4c2a-a964-964bee74523c tempest-ListServerFiltersTestJSON-483030130 tempest-ListServerFiltersTestJSON-483030130-project-member] [instance: f3f872ce-cbbe-4407-b4e0-3341bb598825] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 598.882093] env[61923]: DEBUG nova.network.neutron [req-08acf91d-c666-49e1-9ada-0e524b52033e req-5f720f47-292b-4854-a620-39079a86a57a service nova] [instance: 71d9eccc-3b90-4b0a-a59a-d6a8e08a9417] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 599.119034] env[61923]: DEBUG nova.network.neutron [None req-36da9865-5d98-4c2a-a964-964bee74523c tempest-ListServerFiltersTestJSON-483030130 tempest-ListServerFiltersTestJSON-483030130-project-member] [instance: f3f872ce-cbbe-4407-b4e0-3341bb598825] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 599.241768] env[61923]: INFO nova.compute.manager [-] [instance: 71d9eccc-3b90-4b0a-a59a-d6a8e08a9417] Took 1.05 seconds to deallocate network for instance. [ 599.242723] env[61923]: DEBUG nova.compute.claims [None req-595ebb22-536b-4375-9cbe-6fc472a5a17d tempest-ListServerFiltersTestJSON-483030130 tempest-ListServerFiltersTestJSON-483030130-project-member] [instance: 71d9eccc-3b90-4b0a-a59a-d6a8e08a9417] Aborting claim: {{(pid=61923) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 599.243484] env[61923]: DEBUG oslo_concurrency.lockutils [None req-595ebb22-536b-4375-9cbe-6fc472a5a17d tempest-ListServerFiltersTestJSON-483030130 tempest-ListServerFiltersTestJSON-483030130-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 599.261946] env[61923]: DEBUG oslo_concurrency.lockutils [None req-ae7ff058-a25b-42a3-af4b-0715677fedd4 tempest-ServersAdmin275Test-1076075580 tempest-ServersAdmin275Test-1076075580-project-admin] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.014s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 599.267145] env[61923]: DEBUG oslo_concurrency.lockutils [None req-fba27a13-474d-4fe3-9187-c31143f40aa5 tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 17.932s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 599.385254] env[61923]: DEBUG oslo_concurrency.lockutils [req-08acf91d-c666-49e1-9ada-0e524b52033e req-5f720f47-292b-4854-a620-39079a86a57a service nova] Releasing lock "refresh_cache-71d9eccc-3b90-4b0a-a59a-d6a8e08a9417" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 599.385529] env[61923]: DEBUG nova.compute.manager [req-08acf91d-c666-49e1-9ada-0e524b52033e req-5f720f47-292b-4854-a620-39079a86a57a service nova] [instance: 71d9eccc-3b90-4b0a-a59a-d6a8e08a9417] Received event network-vif-deleted-b42828d3-f036-4a41-949a-3505532203d9 {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 599.622834] env[61923]: DEBUG oslo_concurrency.lockutils [None req-36da9865-5d98-4c2a-a964-964bee74523c tempest-ListServerFiltersTestJSON-483030130 tempest-ListServerFiltersTestJSON-483030130-project-member] Releasing lock "refresh_cache-f3f872ce-cbbe-4407-b4e0-3341bb598825" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 599.623112] env[61923]: DEBUG nova.compute.manager [None req-36da9865-5d98-4c2a-a964-964bee74523c tempest-ListServerFiltersTestJSON-483030130 tempest-ListServerFiltersTestJSON-483030130-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61923) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 599.623291] env[61923]: DEBUG nova.compute.manager [None req-36da9865-5d98-4c2a-a964-964bee74523c tempest-ListServerFiltersTestJSON-483030130 tempest-ListServerFiltersTestJSON-483030130-project-member] [instance: f3f872ce-cbbe-4407-b4e0-3341bb598825] Deallocating network for instance {{(pid=61923) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 599.623477] env[61923]: DEBUG nova.network.neutron [None req-36da9865-5d98-4c2a-a964-964bee74523c tempest-ListServerFiltersTestJSON-483030130 tempest-ListServerFiltersTestJSON-483030130-project-member] [instance: f3f872ce-cbbe-4407-b4e0-3341bb598825] deallocate_for_instance() {{(pid=61923) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 599.670548] env[61923]: DEBUG nova.network.neutron [None req-36da9865-5d98-4c2a-a964-964bee74523c tempest-ListServerFiltersTestJSON-483030130 tempest-ListServerFiltersTestJSON-483030130-project-member] [instance: f3f872ce-cbbe-4407-b4e0-3341bb598825] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 600.174203] env[61923]: DEBUG nova.network.neutron [None req-36da9865-5d98-4c2a-a964-964bee74523c tempest-ListServerFiltersTestJSON-483030130 tempest-ListServerFiltersTestJSON-483030130-project-member] [instance: f3f872ce-cbbe-4407-b4e0-3341bb598825] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 600.314207] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa4a32da-4663-441e-93cb-ad29d8ba272b {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 600.323636] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-285442ab-acea-4349-abca-b2bfe0b20842 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 600.365229] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25b2114e-776d-4fc5-a125-e34363830de3 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 600.374486] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-359a0787-2904-4815-8c2b-48431a82be99 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 600.392758] env[61923]: DEBUG nova.compute.provider_tree [None req-fba27a13-474d-4fe3-9187-c31143f40aa5 tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 600.585079] env[61923]: ERROR nova.compute.manager [None req-03cb7270-244b-47b9-9dd5-e358ebdfa946 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 225a6542-d2aa-4f13-a9d6-7c2e6c048241, please check neutron logs for more information. [ 600.585079] env[61923]: ERROR nova.compute.manager Traceback (most recent call last): [ 600.585079] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 600.585079] env[61923]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 600.585079] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 600.585079] env[61923]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 600.585079] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 600.585079] env[61923]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 600.585079] env[61923]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 600.585079] env[61923]: ERROR nova.compute.manager self.force_reraise() [ 600.585079] env[61923]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 600.585079] env[61923]: ERROR nova.compute.manager raise self.value [ 600.585079] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 600.585079] env[61923]: ERROR nova.compute.manager updated_port = self._update_port( [ 600.585079] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 600.585079] env[61923]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 600.586085] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 600.586085] env[61923]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 600.586085] env[61923]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 225a6542-d2aa-4f13-a9d6-7c2e6c048241, please check neutron logs for more information. [ 600.586085] env[61923]: ERROR nova.compute.manager [ 600.586085] env[61923]: Traceback (most recent call last): [ 600.586085] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 600.586085] env[61923]: listener.cb(fileno) [ 600.586085] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 600.586085] env[61923]: result = function(*args, **kwargs) [ 600.586085] env[61923]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 600.586085] env[61923]: return func(*args, **kwargs) [ 600.586085] env[61923]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 600.586085] env[61923]: raise e [ 600.586085] env[61923]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 600.586085] env[61923]: nwinfo = self.network_api.allocate_for_instance( [ 600.586085] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 600.586085] env[61923]: created_port_ids = self._update_ports_for_instance( [ 600.586085] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 600.586085] env[61923]: with excutils.save_and_reraise_exception(): [ 600.586085] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 600.586085] env[61923]: self.force_reraise() [ 600.586085] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 600.586085] env[61923]: raise self.value [ 600.586085] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 600.586085] env[61923]: updated_port = self._update_port( [ 600.586085] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 600.586085] env[61923]: _ensure_no_port_binding_failure(port) [ 600.586085] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 600.586085] env[61923]: raise exception.PortBindingFailed(port_id=port['id']) [ 600.586750] env[61923]: nova.exception.PortBindingFailed: Binding failed for port 225a6542-d2aa-4f13-a9d6-7c2e6c048241, please check neutron logs for more information. [ 600.586750] env[61923]: Removing descriptor: 17 [ 600.586750] env[61923]: ERROR nova.compute.manager [None req-03cb7270-244b-47b9-9dd5-e358ebdfa946 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: cb998da7-8db0-4725-bed2-657d7df748b5] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 225a6542-d2aa-4f13-a9d6-7c2e6c048241, please check neutron logs for more information. [ 600.586750] env[61923]: ERROR nova.compute.manager [instance: cb998da7-8db0-4725-bed2-657d7df748b5] Traceback (most recent call last): [ 600.586750] env[61923]: ERROR nova.compute.manager [instance: cb998da7-8db0-4725-bed2-657d7df748b5] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 600.586750] env[61923]: ERROR nova.compute.manager [instance: cb998da7-8db0-4725-bed2-657d7df748b5] yield resources [ 600.586750] env[61923]: ERROR nova.compute.manager [instance: cb998da7-8db0-4725-bed2-657d7df748b5] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 600.586750] env[61923]: ERROR nova.compute.manager [instance: cb998da7-8db0-4725-bed2-657d7df748b5] self.driver.spawn(context, instance, image_meta, [ 600.586750] env[61923]: ERROR nova.compute.manager [instance: cb998da7-8db0-4725-bed2-657d7df748b5] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 600.586750] env[61923]: ERROR nova.compute.manager [instance: cb998da7-8db0-4725-bed2-657d7df748b5] self._vmops.spawn(context, instance, image_meta, injected_files, [ 600.586750] env[61923]: ERROR nova.compute.manager [instance: cb998da7-8db0-4725-bed2-657d7df748b5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 600.586750] env[61923]: ERROR nova.compute.manager [instance: cb998da7-8db0-4725-bed2-657d7df748b5] vm_ref = self.build_virtual_machine(instance, [ 600.588912] env[61923]: ERROR nova.compute.manager [instance: cb998da7-8db0-4725-bed2-657d7df748b5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 600.588912] env[61923]: ERROR nova.compute.manager [instance: cb998da7-8db0-4725-bed2-657d7df748b5] vif_infos = vmwarevif.get_vif_info(self._session, [ 600.588912] env[61923]: ERROR nova.compute.manager [instance: cb998da7-8db0-4725-bed2-657d7df748b5] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 600.588912] env[61923]: ERROR nova.compute.manager [instance: cb998da7-8db0-4725-bed2-657d7df748b5] for vif in network_info: [ 600.588912] env[61923]: ERROR nova.compute.manager [instance: cb998da7-8db0-4725-bed2-657d7df748b5] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 600.588912] env[61923]: ERROR nova.compute.manager [instance: cb998da7-8db0-4725-bed2-657d7df748b5] return self._sync_wrapper(fn, *args, **kwargs) [ 600.588912] env[61923]: ERROR nova.compute.manager [instance: cb998da7-8db0-4725-bed2-657d7df748b5] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 600.588912] env[61923]: ERROR nova.compute.manager [instance: cb998da7-8db0-4725-bed2-657d7df748b5] self.wait() [ 600.588912] env[61923]: ERROR nova.compute.manager [instance: cb998da7-8db0-4725-bed2-657d7df748b5] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 600.588912] env[61923]: ERROR nova.compute.manager [instance: cb998da7-8db0-4725-bed2-657d7df748b5] self[:] = self._gt.wait() [ 600.588912] env[61923]: ERROR nova.compute.manager [instance: cb998da7-8db0-4725-bed2-657d7df748b5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 600.588912] env[61923]: ERROR nova.compute.manager [instance: cb998da7-8db0-4725-bed2-657d7df748b5] return self._exit_event.wait() [ 600.588912] env[61923]: ERROR nova.compute.manager [instance: cb998da7-8db0-4725-bed2-657d7df748b5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 600.589365] env[61923]: ERROR nova.compute.manager [instance: cb998da7-8db0-4725-bed2-657d7df748b5] result = hub.switch() [ 600.589365] env[61923]: ERROR nova.compute.manager [instance: cb998da7-8db0-4725-bed2-657d7df748b5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 600.589365] env[61923]: ERROR nova.compute.manager [instance: cb998da7-8db0-4725-bed2-657d7df748b5] return self.greenlet.switch() [ 600.589365] env[61923]: ERROR nova.compute.manager [instance: cb998da7-8db0-4725-bed2-657d7df748b5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 600.589365] env[61923]: ERROR nova.compute.manager [instance: cb998da7-8db0-4725-bed2-657d7df748b5] result = function(*args, **kwargs) [ 600.589365] env[61923]: ERROR nova.compute.manager [instance: cb998da7-8db0-4725-bed2-657d7df748b5] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 600.589365] env[61923]: ERROR nova.compute.manager [instance: cb998da7-8db0-4725-bed2-657d7df748b5] return func(*args, **kwargs) [ 600.589365] env[61923]: ERROR nova.compute.manager [instance: cb998da7-8db0-4725-bed2-657d7df748b5] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 600.589365] env[61923]: ERROR nova.compute.manager [instance: cb998da7-8db0-4725-bed2-657d7df748b5] raise e [ 600.589365] env[61923]: ERROR nova.compute.manager [instance: cb998da7-8db0-4725-bed2-657d7df748b5] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 600.589365] env[61923]: ERROR nova.compute.manager [instance: cb998da7-8db0-4725-bed2-657d7df748b5] nwinfo = self.network_api.allocate_for_instance( [ 600.589365] env[61923]: ERROR nova.compute.manager [instance: cb998da7-8db0-4725-bed2-657d7df748b5] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 600.589365] env[61923]: ERROR nova.compute.manager [instance: cb998da7-8db0-4725-bed2-657d7df748b5] created_port_ids = self._update_ports_for_instance( [ 600.589626] env[61923]: ERROR nova.compute.manager [instance: cb998da7-8db0-4725-bed2-657d7df748b5] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 600.589626] env[61923]: ERROR nova.compute.manager [instance: cb998da7-8db0-4725-bed2-657d7df748b5] with excutils.save_and_reraise_exception(): [ 600.589626] env[61923]: ERROR nova.compute.manager [instance: cb998da7-8db0-4725-bed2-657d7df748b5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 600.589626] env[61923]: ERROR nova.compute.manager [instance: cb998da7-8db0-4725-bed2-657d7df748b5] self.force_reraise() [ 600.589626] env[61923]: ERROR nova.compute.manager [instance: cb998da7-8db0-4725-bed2-657d7df748b5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 600.589626] env[61923]: ERROR nova.compute.manager [instance: cb998da7-8db0-4725-bed2-657d7df748b5] raise self.value [ 600.589626] env[61923]: ERROR nova.compute.manager [instance: cb998da7-8db0-4725-bed2-657d7df748b5] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 600.589626] env[61923]: ERROR nova.compute.manager [instance: cb998da7-8db0-4725-bed2-657d7df748b5] updated_port = self._update_port( [ 600.589626] env[61923]: ERROR nova.compute.manager [instance: cb998da7-8db0-4725-bed2-657d7df748b5] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 600.589626] env[61923]: ERROR nova.compute.manager [instance: cb998da7-8db0-4725-bed2-657d7df748b5] _ensure_no_port_binding_failure(port) [ 600.589626] env[61923]: ERROR nova.compute.manager [instance: cb998da7-8db0-4725-bed2-657d7df748b5] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 600.589626] env[61923]: ERROR nova.compute.manager [instance: cb998da7-8db0-4725-bed2-657d7df748b5] raise exception.PortBindingFailed(port_id=port['id']) [ 600.589868] env[61923]: ERROR nova.compute.manager [instance: cb998da7-8db0-4725-bed2-657d7df748b5] nova.exception.PortBindingFailed: Binding failed for port 225a6542-d2aa-4f13-a9d6-7c2e6c048241, please check neutron logs for more information. [ 600.589868] env[61923]: ERROR nova.compute.manager [instance: cb998da7-8db0-4725-bed2-657d7df748b5] [ 600.589868] env[61923]: INFO nova.compute.manager [None req-03cb7270-244b-47b9-9dd5-e358ebdfa946 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: cb998da7-8db0-4725-bed2-657d7df748b5] Terminating instance [ 600.590280] env[61923]: DEBUG oslo_concurrency.lockutils [None req-03cb7270-244b-47b9-9dd5-e358ebdfa946 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Acquiring lock "refresh_cache-cb998da7-8db0-4725-bed2-657d7df748b5" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 600.593513] env[61923]: DEBUG oslo_concurrency.lockutils [None req-03cb7270-244b-47b9-9dd5-e358ebdfa946 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Acquired lock "refresh_cache-cb998da7-8db0-4725-bed2-657d7df748b5" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 600.593513] env[61923]: DEBUG nova.network.neutron [None req-03cb7270-244b-47b9-9dd5-e358ebdfa946 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: cb998da7-8db0-4725-bed2-657d7df748b5] Building network info cache for instance {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 600.682344] env[61923]: INFO nova.compute.manager [None req-36da9865-5d98-4c2a-a964-964bee74523c tempest-ListServerFiltersTestJSON-483030130 tempest-ListServerFiltersTestJSON-483030130-project-member] [instance: f3f872ce-cbbe-4407-b4e0-3341bb598825] Took 1.06 seconds to deallocate network for instance. [ 600.788673] env[61923]: DEBUG oslo_concurrency.lockutils [None req-425a3606-915c-4f9b-8507-50f1c8d0d8d9 tempest-ServerRescueTestJSON-521767381 tempest-ServerRescueTestJSON-521767381-project-member] Acquiring lock "74b32299-983f-41f7-b3f4-3a12815a83e6" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 600.788673] env[61923]: DEBUG oslo_concurrency.lockutils [None req-425a3606-915c-4f9b-8507-50f1c8d0d8d9 tempest-ServerRescueTestJSON-521767381 tempest-ServerRescueTestJSON-521767381-project-member] Lock "74b32299-983f-41f7-b3f4-3a12815a83e6" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 600.831650] env[61923]: DEBUG nova.compute.manager [req-3cf3ebcc-2851-480e-9e4b-d6417e272d07 req-c6293dd9-1c02-4a1c-bfd3-bf89f49e7a86 service nova] [instance: cb998da7-8db0-4725-bed2-657d7df748b5] Received event network-changed-225a6542-d2aa-4f13-a9d6-7c2e6c048241 {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 600.831926] env[61923]: DEBUG nova.compute.manager [req-3cf3ebcc-2851-480e-9e4b-d6417e272d07 req-c6293dd9-1c02-4a1c-bfd3-bf89f49e7a86 service nova] [instance: cb998da7-8db0-4725-bed2-657d7df748b5] Refreshing instance network info cache due to event network-changed-225a6542-d2aa-4f13-a9d6-7c2e6c048241. {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 600.832180] env[61923]: DEBUG oslo_concurrency.lockutils [req-3cf3ebcc-2851-480e-9e4b-d6417e272d07 req-c6293dd9-1c02-4a1c-bfd3-bf89f49e7a86 service nova] Acquiring lock "refresh_cache-cb998da7-8db0-4725-bed2-657d7df748b5" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 600.896486] env[61923]: DEBUG nova.scheduler.client.report [None req-fba27a13-474d-4fe3-9187-c31143f40aa5 tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 601.128451] env[61923]: DEBUG nova.network.neutron [None req-03cb7270-244b-47b9-9dd5-e358ebdfa946 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: cb998da7-8db0-4725-bed2-657d7df748b5] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 601.360655] env[61923]: DEBUG nova.network.neutron [None req-03cb7270-244b-47b9-9dd5-e358ebdfa946 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: cb998da7-8db0-4725-bed2-657d7df748b5] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 601.404456] env[61923]: DEBUG oslo_concurrency.lockutils [None req-fba27a13-474d-4fe3-9187-c31143f40aa5 tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.139s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 601.405536] env[61923]: ERROR nova.compute.manager [None req-fba27a13-474d-4fe3-9187-c31143f40aa5 tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] [instance: 6a2d7ede-e0cc-4a36-b1b2-0b36b269ecb5] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 3b3effc4-2768-46ad-8e11-d2719c57d383, please check neutron logs for more information. [ 601.405536] env[61923]: ERROR nova.compute.manager [instance: 6a2d7ede-e0cc-4a36-b1b2-0b36b269ecb5] Traceback (most recent call last): [ 601.405536] env[61923]: ERROR nova.compute.manager [instance: 6a2d7ede-e0cc-4a36-b1b2-0b36b269ecb5] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 601.405536] env[61923]: ERROR nova.compute.manager [instance: 6a2d7ede-e0cc-4a36-b1b2-0b36b269ecb5] self.driver.spawn(context, instance, image_meta, [ 601.405536] env[61923]: ERROR nova.compute.manager [instance: 6a2d7ede-e0cc-4a36-b1b2-0b36b269ecb5] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 601.405536] env[61923]: ERROR nova.compute.manager [instance: 6a2d7ede-e0cc-4a36-b1b2-0b36b269ecb5] self._vmops.spawn(context, instance, image_meta, injected_files, [ 601.405536] env[61923]: ERROR nova.compute.manager [instance: 6a2d7ede-e0cc-4a36-b1b2-0b36b269ecb5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 601.405536] env[61923]: ERROR nova.compute.manager [instance: 6a2d7ede-e0cc-4a36-b1b2-0b36b269ecb5] vm_ref = self.build_virtual_machine(instance, [ 601.405536] env[61923]: ERROR nova.compute.manager [instance: 6a2d7ede-e0cc-4a36-b1b2-0b36b269ecb5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 601.405536] env[61923]: ERROR nova.compute.manager [instance: 6a2d7ede-e0cc-4a36-b1b2-0b36b269ecb5] vif_infos = vmwarevif.get_vif_info(self._session, [ 601.405536] env[61923]: ERROR nova.compute.manager [instance: 6a2d7ede-e0cc-4a36-b1b2-0b36b269ecb5] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 601.405842] env[61923]: ERROR nova.compute.manager [instance: 6a2d7ede-e0cc-4a36-b1b2-0b36b269ecb5] for vif in network_info: [ 601.405842] env[61923]: ERROR nova.compute.manager [instance: 6a2d7ede-e0cc-4a36-b1b2-0b36b269ecb5] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 601.405842] env[61923]: ERROR nova.compute.manager [instance: 6a2d7ede-e0cc-4a36-b1b2-0b36b269ecb5] return self._sync_wrapper(fn, *args, **kwargs) [ 601.405842] env[61923]: ERROR nova.compute.manager [instance: 6a2d7ede-e0cc-4a36-b1b2-0b36b269ecb5] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 601.405842] env[61923]: ERROR nova.compute.manager [instance: 6a2d7ede-e0cc-4a36-b1b2-0b36b269ecb5] self.wait() [ 601.405842] env[61923]: ERROR nova.compute.manager [instance: 6a2d7ede-e0cc-4a36-b1b2-0b36b269ecb5] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 601.405842] env[61923]: ERROR nova.compute.manager [instance: 6a2d7ede-e0cc-4a36-b1b2-0b36b269ecb5] self[:] = self._gt.wait() [ 601.405842] env[61923]: ERROR nova.compute.manager [instance: 6a2d7ede-e0cc-4a36-b1b2-0b36b269ecb5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 601.405842] env[61923]: ERROR nova.compute.manager [instance: 6a2d7ede-e0cc-4a36-b1b2-0b36b269ecb5] return self._exit_event.wait() [ 601.405842] env[61923]: ERROR nova.compute.manager [instance: 6a2d7ede-e0cc-4a36-b1b2-0b36b269ecb5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 601.405842] env[61923]: ERROR nova.compute.manager [instance: 6a2d7ede-e0cc-4a36-b1b2-0b36b269ecb5] result = hub.switch() [ 601.405842] env[61923]: ERROR nova.compute.manager [instance: 6a2d7ede-e0cc-4a36-b1b2-0b36b269ecb5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 601.405842] env[61923]: ERROR nova.compute.manager [instance: 6a2d7ede-e0cc-4a36-b1b2-0b36b269ecb5] return self.greenlet.switch() [ 601.406190] env[61923]: ERROR nova.compute.manager [instance: 6a2d7ede-e0cc-4a36-b1b2-0b36b269ecb5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 601.406190] env[61923]: ERROR nova.compute.manager [instance: 6a2d7ede-e0cc-4a36-b1b2-0b36b269ecb5] result = function(*args, **kwargs) [ 601.406190] env[61923]: ERROR nova.compute.manager [instance: 6a2d7ede-e0cc-4a36-b1b2-0b36b269ecb5] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 601.406190] env[61923]: ERROR nova.compute.manager [instance: 6a2d7ede-e0cc-4a36-b1b2-0b36b269ecb5] return func(*args, **kwargs) [ 601.406190] env[61923]: ERROR nova.compute.manager [instance: 6a2d7ede-e0cc-4a36-b1b2-0b36b269ecb5] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 601.406190] env[61923]: ERROR nova.compute.manager [instance: 6a2d7ede-e0cc-4a36-b1b2-0b36b269ecb5] raise e [ 601.406190] env[61923]: ERROR nova.compute.manager [instance: 6a2d7ede-e0cc-4a36-b1b2-0b36b269ecb5] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 601.406190] env[61923]: ERROR nova.compute.manager [instance: 6a2d7ede-e0cc-4a36-b1b2-0b36b269ecb5] nwinfo = self.network_api.allocate_for_instance( [ 601.406190] env[61923]: ERROR nova.compute.manager [instance: 6a2d7ede-e0cc-4a36-b1b2-0b36b269ecb5] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 601.406190] env[61923]: ERROR nova.compute.manager [instance: 6a2d7ede-e0cc-4a36-b1b2-0b36b269ecb5] created_port_ids = self._update_ports_for_instance( [ 601.406190] env[61923]: ERROR nova.compute.manager [instance: 6a2d7ede-e0cc-4a36-b1b2-0b36b269ecb5] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 601.406190] env[61923]: ERROR nova.compute.manager [instance: 6a2d7ede-e0cc-4a36-b1b2-0b36b269ecb5] with excutils.save_and_reraise_exception(): [ 601.406190] env[61923]: ERROR nova.compute.manager [instance: 6a2d7ede-e0cc-4a36-b1b2-0b36b269ecb5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 601.406520] env[61923]: ERROR nova.compute.manager [instance: 6a2d7ede-e0cc-4a36-b1b2-0b36b269ecb5] self.force_reraise() [ 601.406520] env[61923]: ERROR nova.compute.manager [instance: 6a2d7ede-e0cc-4a36-b1b2-0b36b269ecb5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 601.406520] env[61923]: ERROR nova.compute.manager [instance: 6a2d7ede-e0cc-4a36-b1b2-0b36b269ecb5] raise self.value [ 601.406520] env[61923]: ERROR nova.compute.manager [instance: 6a2d7ede-e0cc-4a36-b1b2-0b36b269ecb5] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 601.406520] env[61923]: ERROR nova.compute.manager [instance: 6a2d7ede-e0cc-4a36-b1b2-0b36b269ecb5] updated_port = self._update_port( [ 601.406520] env[61923]: ERROR nova.compute.manager [instance: 6a2d7ede-e0cc-4a36-b1b2-0b36b269ecb5] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 601.406520] env[61923]: ERROR nova.compute.manager [instance: 6a2d7ede-e0cc-4a36-b1b2-0b36b269ecb5] _ensure_no_port_binding_failure(port) [ 601.406520] env[61923]: ERROR nova.compute.manager [instance: 6a2d7ede-e0cc-4a36-b1b2-0b36b269ecb5] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 601.406520] env[61923]: ERROR nova.compute.manager [instance: 6a2d7ede-e0cc-4a36-b1b2-0b36b269ecb5] raise exception.PortBindingFailed(port_id=port['id']) [ 601.406520] env[61923]: ERROR nova.compute.manager [instance: 6a2d7ede-e0cc-4a36-b1b2-0b36b269ecb5] nova.exception.PortBindingFailed: Binding failed for port 3b3effc4-2768-46ad-8e11-d2719c57d383, please check neutron logs for more information. [ 601.406520] env[61923]: ERROR nova.compute.manager [instance: 6a2d7ede-e0cc-4a36-b1b2-0b36b269ecb5] [ 601.406761] env[61923]: DEBUG nova.compute.utils [None req-fba27a13-474d-4fe3-9187-c31143f40aa5 tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] [instance: 6a2d7ede-e0cc-4a36-b1b2-0b36b269ecb5] Binding failed for port 3b3effc4-2768-46ad-8e11-d2719c57d383, please check neutron logs for more information. {{(pid=61923) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 601.407965] env[61923]: DEBUG oslo_concurrency.lockutils [None req-f3aa5656-98df-4c46-9436-b968a8770dbc tempest-ListServerFiltersTestJSON-483030130 tempest-ListServerFiltersTestJSON-483030130-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.179s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 601.410469] env[61923]: INFO nova.compute.claims [None req-f3aa5656-98df-4c46-9436-b968a8770dbc tempest-ListServerFiltersTestJSON-483030130 tempest-ListServerFiltersTestJSON-483030130-project-member] [instance: 896dd689-4810-4f23-af0a-d2f557a0796b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 601.414147] env[61923]: DEBUG nova.compute.manager [None req-fba27a13-474d-4fe3-9187-c31143f40aa5 tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] [instance: 6a2d7ede-e0cc-4a36-b1b2-0b36b269ecb5] Build of instance 6a2d7ede-e0cc-4a36-b1b2-0b36b269ecb5 was re-scheduled: Binding failed for port 3b3effc4-2768-46ad-8e11-d2719c57d383, please check neutron logs for more information. {{(pid=61923) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 601.414147] env[61923]: DEBUG nova.compute.manager [None req-fba27a13-474d-4fe3-9187-c31143f40aa5 tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] [instance: 6a2d7ede-e0cc-4a36-b1b2-0b36b269ecb5] Unplugging VIFs for instance {{(pid=61923) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 601.414147] env[61923]: DEBUG oslo_concurrency.lockutils [None req-fba27a13-474d-4fe3-9187-c31143f40aa5 tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] Acquiring lock "refresh_cache-6a2d7ede-e0cc-4a36-b1b2-0b36b269ecb5" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 601.414147] env[61923]: DEBUG oslo_concurrency.lockutils [None req-fba27a13-474d-4fe3-9187-c31143f40aa5 tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] Acquired lock "refresh_cache-6a2d7ede-e0cc-4a36-b1b2-0b36b269ecb5" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 601.414432] env[61923]: DEBUG nova.network.neutron [None req-fba27a13-474d-4fe3-9187-c31143f40aa5 tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] [instance: 6a2d7ede-e0cc-4a36-b1b2-0b36b269ecb5] Building network info cache for instance {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 601.726016] env[61923]: INFO nova.scheduler.client.report [None req-36da9865-5d98-4c2a-a964-964bee74523c tempest-ListServerFiltersTestJSON-483030130 tempest-ListServerFiltersTestJSON-483030130-project-member] Deleted allocations for instance f3f872ce-cbbe-4407-b4e0-3341bb598825 [ 601.866072] env[61923]: DEBUG oslo_concurrency.lockutils [None req-03cb7270-244b-47b9-9dd5-e358ebdfa946 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Releasing lock "refresh_cache-cb998da7-8db0-4725-bed2-657d7df748b5" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 601.866072] env[61923]: DEBUG nova.compute.manager [None req-03cb7270-244b-47b9-9dd5-e358ebdfa946 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: cb998da7-8db0-4725-bed2-657d7df748b5] Start destroying the instance on the hypervisor. {{(pid=61923) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 601.866072] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-03cb7270-244b-47b9-9dd5-e358ebdfa946 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: cb998da7-8db0-4725-bed2-657d7df748b5] Destroying instance {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 601.866072] env[61923]: DEBUG oslo_concurrency.lockutils [req-3cf3ebcc-2851-480e-9e4b-d6417e272d07 req-c6293dd9-1c02-4a1c-bfd3-bf89f49e7a86 service nova] Acquired lock "refresh_cache-cb998da7-8db0-4725-bed2-657d7df748b5" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 601.866072] env[61923]: DEBUG nova.network.neutron [req-3cf3ebcc-2851-480e-9e4b-d6417e272d07 req-c6293dd9-1c02-4a1c-bfd3-bf89f49e7a86 service nova] [instance: cb998da7-8db0-4725-bed2-657d7df748b5] Refreshing network info cache for port 225a6542-d2aa-4f13-a9d6-7c2e6c048241 {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 601.866320] env[61923]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-fac04e7f-3305-4257-b8b0-2ae00503c4db {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.884045] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53cd5f90-45a2-4205-b74a-1d45732897dd {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.908362] env[61923]: WARNING nova.virt.vmwareapi.vmops [None req-03cb7270-244b-47b9-9dd5-e358ebdfa946 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: cb998da7-8db0-4725-bed2-657d7df748b5] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance cb998da7-8db0-4725-bed2-657d7df748b5 could not be found. [ 601.908911] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-03cb7270-244b-47b9-9dd5-e358ebdfa946 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: cb998da7-8db0-4725-bed2-657d7df748b5] Instance destroyed {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 601.909452] env[61923]: INFO nova.compute.manager [None req-03cb7270-244b-47b9-9dd5-e358ebdfa946 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: cb998da7-8db0-4725-bed2-657d7df748b5] Took 0.05 seconds to destroy the instance on the hypervisor. [ 601.910089] env[61923]: DEBUG oslo.service.loopingcall [None req-03cb7270-244b-47b9-9dd5-e358ebdfa946 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61923) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 601.910460] env[61923]: DEBUG nova.compute.manager [-] [instance: cb998da7-8db0-4725-bed2-657d7df748b5] Deallocating network for instance {{(pid=61923) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 601.910816] env[61923]: DEBUG nova.network.neutron [-] [instance: cb998da7-8db0-4725-bed2-657d7df748b5] deallocate_for_instance() {{(pid=61923) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 601.974159] env[61923]: DEBUG nova.network.neutron [-] [instance: cb998da7-8db0-4725-bed2-657d7df748b5] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 601.974159] env[61923]: DEBUG nova.network.neutron [None req-fba27a13-474d-4fe3-9187-c31143f40aa5 tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] [instance: 6a2d7ede-e0cc-4a36-b1b2-0b36b269ecb5] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 602.176496] env[61923]: DEBUG nova.network.neutron [None req-fba27a13-474d-4fe3-9187-c31143f40aa5 tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] [instance: 6a2d7ede-e0cc-4a36-b1b2-0b36b269ecb5] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 602.244734] env[61923]: DEBUG oslo_concurrency.lockutils [None req-36da9865-5d98-4c2a-a964-964bee74523c tempest-ListServerFiltersTestJSON-483030130 tempest-ListServerFiltersTestJSON-483030130-project-member] Lock "f3f872ce-cbbe-4407-b4e0-3341bb598825" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 64.717s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 602.415468] env[61923]: DEBUG nova.network.neutron [req-3cf3ebcc-2851-480e-9e4b-d6417e272d07 req-c6293dd9-1c02-4a1c-bfd3-bf89f49e7a86 service nova] [instance: cb998da7-8db0-4725-bed2-657d7df748b5] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 602.477164] env[61923]: DEBUG nova.network.neutron [-] [instance: cb998da7-8db0-4725-bed2-657d7df748b5] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 602.598096] env[61923]: DEBUG nova.network.neutron [req-3cf3ebcc-2851-480e-9e4b-d6417e272d07 req-c6293dd9-1c02-4a1c-bfd3-bf89f49e7a86 service nova] [instance: cb998da7-8db0-4725-bed2-657d7df748b5] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 602.683440] env[61923]: DEBUG oslo_concurrency.lockutils [None req-fba27a13-474d-4fe3-9187-c31143f40aa5 tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] Releasing lock "refresh_cache-6a2d7ede-e0cc-4a36-b1b2-0b36b269ecb5" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 602.684767] env[61923]: DEBUG nova.compute.manager [None req-fba27a13-474d-4fe3-9187-c31143f40aa5 tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61923) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 602.684767] env[61923]: DEBUG nova.compute.manager [None req-fba27a13-474d-4fe3-9187-c31143f40aa5 tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] [instance: 6a2d7ede-e0cc-4a36-b1b2-0b36b269ecb5] Deallocating network for instance {{(pid=61923) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 602.684767] env[61923]: DEBUG nova.network.neutron [None req-fba27a13-474d-4fe3-9187-c31143f40aa5 tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] [instance: 6a2d7ede-e0cc-4a36-b1b2-0b36b269ecb5] deallocate_for_instance() {{(pid=61923) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 602.709423] env[61923]: DEBUG nova.network.neutron [None req-fba27a13-474d-4fe3-9187-c31143f40aa5 tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] [instance: 6a2d7ede-e0cc-4a36-b1b2-0b36b269ecb5] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 602.747181] env[61923]: DEBUG nova.compute.manager [None req-aa9c464b-ac6b-44fe-9384-28cd9bdfecbc tempest-ImagesOneServerTestJSON-63657725 tempest-ImagesOneServerTestJSON-63657725-project-member] [instance: 0c1b8117-199f-4101-ab81-4ae7ef0d1251] Starting instance... {{(pid=61923) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 602.869167] env[61923]: DEBUG nova.compute.manager [req-bb5dcdc3-7c52-4e61-903d-761e465683b5 req-add619e9-553b-401d-856e-182ba08dcc3e service nova] [instance: cb998da7-8db0-4725-bed2-657d7df748b5] Received event network-vif-deleted-225a6542-d2aa-4f13-a9d6-7c2e6c048241 {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 602.942837] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fdf80c6d-a89f-43b6-816d-d252cae7bfba {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 602.955153] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-617c9e63-b721-49de-971d-6b8fb9012532 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 602.987271] env[61923]: INFO nova.compute.manager [-] [instance: cb998da7-8db0-4725-bed2-657d7df748b5] Took 1.08 seconds to deallocate network for instance. [ 602.989847] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32dc806b-086b-4e6a-bc32-23d7eab1cd39 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 602.992646] env[61923]: DEBUG nova.compute.claims [None req-03cb7270-244b-47b9-9dd5-e358ebdfa946 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: cb998da7-8db0-4725-bed2-657d7df748b5] Aborting claim: {{(pid=61923) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 602.992826] env[61923]: DEBUG oslo_concurrency.lockutils [None req-03cb7270-244b-47b9-9dd5-e358ebdfa946 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 602.999438] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1ae5bb7-b190-40f8-977b-a811a60a3507 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.015016] env[61923]: DEBUG nova.compute.provider_tree [None req-f3aa5656-98df-4c46-9436-b968a8770dbc tempest-ListServerFiltersTestJSON-483030130 tempest-ListServerFiltersTestJSON-483030130-project-member] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 603.102599] env[61923]: DEBUG oslo_concurrency.lockutils [req-3cf3ebcc-2851-480e-9e4b-d6417e272d07 req-c6293dd9-1c02-4a1c-bfd3-bf89f49e7a86 service nova] Releasing lock "refresh_cache-cb998da7-8db0-4725-bed2-657d7df748b5" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 603.213956] env[61923]: DEBUG nova.network.neutron [None req-fba27a13-474d-4fe3-9187-c31143f40aa5 tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] [instance: 6a2d7ede-e0cc-4a36-b1b2-0b36b269ecb5] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 603.276227] env[61923]: DEBUG oslo_concurrency.lockutils [None req-aa9c464b-ac6b-44fe-9384-28cd9bdfecbc tempest-ImagesOneServerTestJSON-63657725 tempest-ImagesOneServerTestJSON-63657725-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 603.518102] env[61923]: DEBUG nova.scheduler.client.report [None req-f3aa5656-98df-4c46-9436-b968a8770dbc tempest-ListServerFiltersTestJSON-483030130 tempest-ListServerFiltersTestJSON-483030130-project-member] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 603.719740] env[61923]: INFO nova.compute.manager [None req-fba27a13-474d-4fe3-9187-c31143f40aa5 tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] [instance: 6a2d7ede-e0cc-4a36-b1b2-0b36b269ecb5] Took 1.04 seconds to deallocate network for instance. [ 604.026749] env[61923]: DEBUG oslo_concurrency.lockutils [None req-f3aa5656-98df-4c46-9436-b968a8770dbc tempest-ListServerFiltersTestJSON-483030130 tempest-ListServerFiltersTestJSON-483030130-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.620s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 604.027898] env[61923]: DEBUG nova.compute.manager [None req-f3aa5656-98df-4c46-9436-b968a8770dbc tempest-ListServerFiltersTestJSON-483030130 tempest-ListServerFiltersTestJSON-483030130-project-member] [instance: 896dd689-4810-4f23-af0a-d2f557a0796b] Start building networks asynchronously for instance. {{(pid=61923) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 604.032334] env[61923]: DEBUG oslo_concurrency.lockutils [None req-47902537-98f9-4fe7-afdb-4b2d4d440d10 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 17.710s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 604.032983] env[61923]: DEBUG nova.objects.instance [None req-47902537-98f9-4fe7-afdb-4b2d4d440d10 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] Lazy-loading 'resources' on Instance uuid f7cc960c-e06a-4c58-9367-ec8771fe09d5 {{(pid=61923) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 604.495117] env[61923]: DEBUG oslo_concurrency.lockutils [None req-40b011d4-39ab-4684-a124-b3ba87d489df tempest-ServerPasswordTestJSON-595643156 tempest-ServerPasswordTestJSON-595643156-project-member] Acquiring lock "7c0341f9-7401-42ee-9512-afd832c76940" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 604.495340] env[61923]: DEBUG oslo_concurrency.lockutils [None req-40b011d4-39ab-4684-a124-b3ba87d489df tempest-ServerPasswordTestJSON-595643156 tempest-ServerPasswordTestJSON-595643156-project-member] Lock "7c0341f9-7401-42ee-9512-afd832c76940" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 604.540112] env[61923]: DEBUG nova.compute.utils [None req-f3aa5656-98df-4c46-9436-b968a8770dbc tempest-ListServerFiltersTestJSON-483030130 tempest-ListServerFiltersTestJSON-483030130-project-member] Using /dev/sd instead of None {{(pid=61923) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 604.545144] env[61923]: DEBUG nova.compute.manager [None req-f3aa5656-98df-4c46-9436-b968a8770dbc tempest-ListServerFiltersTestJSON-483030130 tempest-ListServerFiltersTestJSON-483030130-project-member] [instance: 896dd689-4810-4f23-af0a-d2f557a0796b] Allocating IP information in the background. {{(pid=61923) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 604.545759] env[61923]: DEBUG nova.network.neutron [None req-f3aa5656-98df-4c46-9436-b968a8770dbc tempest-ListServerFiltersTestJSON-483030130 tempest-ListServerFiltersTestJSON-483030130-project-member] [instance: 896dd689-4810-4f23-af0a-d2f557a0796b] allocate_for_instance() {{(pid=61923) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 604.603893] env[61923]: DEBUG nova.policy [None req-f3aa5656-98df-4c46-9436-b968a8770dbc tempest-ListServerFiltersTestJSON-483030130 tempest-ListServerFiltersTestJSON-483030130-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '16836760c008401586e5b67b19a53cc0', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '6c5ff14dc18843cc875b21b949eca824', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61923) authorize /opt/stack/nova/nova/policy.py:201}} [ 604.766664] env[61923]: INFO nova.scheduler.client.report [None req-fba27a13-474d-4fe3-9187-c31143f40aa5 tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] Deleted allocations for instance 6a2d7ede-e0cc-4a36-b1b2-0b36b269ecb5 [ 605.042308] env[61923]: DEBUG nova.network.neutron [None req-f3aa5656-98df-4c46-9436-b968a8770dbc tempest-ListServerFiltersTestJSON-483030130 tempest-ListServerFiltersTestJSON-483030130-project-member] [instance: 896dd689-4810-4f23-af0a-d2f557a0796b] Successfully created port: 53407134-6b24-4ff7-b5d0-ef4cf1885933 {{(pid=61923) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 605.047432] env[61923]: DEBUG nova.compute.manager [None req-f3aa5656-98df-4c46-9436-b968a8770dbc tempest-ListServerFiltersTestJSON-483030130 tempest-ListServerFiltersTestJSON-483030130-project-member] [instance: 896dd689-4810-4f23-af0a-d2f557a0796b] Start building block device mappings for instance. {{(pid=61923) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 605.111873] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76a3468b-be65-4d36-a237-2b8098b3dede {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 605.120639] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6040b16a-affd-482a-abcd-8c9b0f73f799 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 605.157242] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-811b642f-ed5a-4309-ae6c-4a31a4b8c425 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 605.166380] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0fb3bd5c-9b8d-4241-8ed1-32c98a1f8ef0 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 605.180248] env[61923]: DEBUG nova.compute.provider_tree [None req-47902537-98f9-4fe7-afdb-4b2d4d440d10 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 605.275305] env[61923]: DEBUG oslo_concurrency.lockutils [None req-fba27a13-474d-4fe3-9187-c31143f40aa5 tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] Lock "6a2d7ede-e0cc-4a36-b1b2-0b36b269ecb5" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 66.307s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 605.682899] env[61923]: DEBUG nova.scheduler.client.report [None req-47902537-98f9-4fe7-afdb-4b2d4d440d10 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 605.779024] env[61923]: DEBUG nova.compute.manager [None req-0c55849d-766e-4fac-b1c4-e4e97ba688a6 tempest-DeleteServersAdminTestJSON-590116201 tempest-DeleteServersAdminTestJSON-590116201-project-member] [instance: 1b509288-2df3-4927-afcf-cb76918d3da3] Starting instance... {{(pid=61923) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 606.060850] env[61923]: DEBUG nova.compute.manager [None req-f3aa5656-98df-4c46-9436-b968a8770dbc tempest-ListServerFiltersTestJSON-483030130 tempest-ListServerFiltersTestJSON-483030130-project-member] [instance: 896dd689-4810-4f23-af0a-d2f557a0796b] Start spawning the instance on the hypervisor. {{(pid=61923) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 606.098884] env[61923]: DEBUG nova.virt.hardware [None req-f3aa5656-98df-4c46-9436-b968a8770dbc tempest-ListServerFiltersTestJSON-483030130 tempest-ListServerFiltersTestJSON-483030130-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-29T20:07:54Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-29T20:07:35Z,direct_url=,disk_format='vmdk',id=0f153f63-ae0a-45b1-b7f5-7f9b673c947f,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='4e7b5e3b3c3443c8bd5c70f8a15739f5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-29T20:07:36Z,virtual_size=,visibility=), allow threads: False {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 606.099152] env[61923]: DEBUG nova.virt.hardware [None req-f3aa5656-98df-4c46-9436-b968a8770dbc tempest-ListServerFiltersTestJSON-483030130 tempest-ListServerFiltersTestJSON-483030130-project-member] Flavor limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 606.099307] env[61923]: DEBUG nova.virt.hardware [None req-f3aa5656-98df-4c46-9436-b968a8770dbc tempest-ListServerFiltersTestJSON-483030130 tempest-ListServerFiltersTestJSON-483030130-project-member] Image limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 606.099513] env[61923]: DEBUG nova.virt.hardware [None req-f3aa5656-98df-4c46-9436-b968a8770dbc tempest-ListServerFiltersTestJSON-483030130 tempest-ListServerFiltersTestJSON-483030130-project-member] Flavor pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 606.099619] env[61923]: DEBUG nova.virt.hardware [None req-f3aa5656-98df-4c46-9436-b968a8770dbc tempest-ListServerFiltersTestJSON-483030130 tempest-ListServerFiltersTestJSON-483030130-project-member] Image pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 606.099765] env[61923]: DEBUG nova.virt.hardware [None req-f3aa5656-98df-4c46-9436-b968a8770dbc tempest-ListServerFiltersTestJSON-483030130 tempest-ListServerFiltersTestJSON-483030130-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 606.099969] env[61923]: DEBUG nova.virt.hardware [None req-f3aa5656-98df-4c46-9436-b968a8770dbc tempest-ListServerFiltersTestJSON-483030130 tempest-ListServerFiltersTestJSON-483030130-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 606.100160] env[61923]: DEBUG nova.virt.hardware [None req-f3aa5656-98df-4c46-9436-b968a8770dbc tempest-ListServerFiltersTestJSON-483030130 tempest-ListServerFiltersTestJSON-483030130-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 606.100368] env[61923]: DEBUG nova.virt.hardware [None req-f3aa5656-98df-4c46-9436-b968a8770dbc tempest-ListServerFiltersTestJSON-483030130 tempest-ListServerFiltersTestJSON-483030130-project-member] Got 1 possible topologies {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 606.100472] env[61923]: DEBUG nova.virt.hardware [None req-f3aa5656-98df-4c46-9436-b968a8770dbc tempest-ListServerFiltersTestJSON-483030130 tempest-ListServerFiltersTestJSON-483030130-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 606.100638] env[61923]: DEBUG nova.virt.hardware [None req-f3aa5656-98df-4c46-9436-b968a8770dbc tempest-ListServerFiltersTestJSON-483030130 tempest-ListServerFiltersTestJSON-483030130-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 606.101566] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7802fd8-5bad-4191-be1b-052f52682d70 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.113378] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-036aa4fb-97f6-4072-bb2f-e3af7a282757 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.189187] env[61923]: DEBUG oslo_concurrency.lockutils [None req-47902537-98f9-4fe7-afdb-4b2d4d440d10 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.157s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 606.191552] env[61923]: DEBUG oslo_concurrency.lockutils [None req-fba27a13-474d-4fe3-9187-c31143f40aa5 tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 19.352s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 606.221946] env[61923]: INFO nova.scheduler.client.report [None req-47902537-98f9-4fe7-afdb-4b2d4d440d10 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] Deleted allocations for instance f7cc960c-e06a-4c58-9367-ec8771fe09d5 [ 606.290704] env[61923]: DEBUG nova.compute.manager [req-141ed8ab-3127-4ad5-9869-de706c9f0f40 req-c6f1014c-ae25-4f9c-ab1b-e386e64a245c service nova] [instance: 896dd689-4810-4f23-af0a-d2f557a0796b] Received event network-changed-53407134-6b24-4ff7-b5d0-ef4cf1885933 {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 606.290916] env[61923]: DEBUG nova.compute.manager [req-141ed8ab-3127-4ad5-9869-de706c9f0f40 req-c6f1014c-ae25-4f9c-ab1b-e386e64a245c service nova] [instance: 896dd689-4810-4f23-af0a-d2f557a0796b] Refreshing instance network info cache due to event network-changed-53407134-6b24-4ff7-b5d0-ef4cf1885933. {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 606.291286] env[61923]: DEBUG oslo_concurrency.lockutils [req-141ed8ab-3127-4ad5-9869-de706c9f0f40 req-c6f1014c-ae25-4f9c-ab1b-e386e64a245c service nova] Acquiring lock "refresh_cache-896dd689-4810-4f23-af0a-d2f557a0796b" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 606.291360] env[61923]: DEBUG oslo_concurrency.lockutils [req-141ed8ab-3127-4ad5-9869-de706c9f0f40 req-c6f1014c-ae25-4f9c-ab1b-e386e64a245c service nova] Acquired lock "refresh_cache-896dd689-4810-4f23-af0a-d2f557a0796b" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 606.291587] env[61923]: DEBUG nova.network.neutron [req-141ed8ab-3127-4ad5-9869-de706c9f0f40 req-c6f1014c-ae25-4f9c-ab1b-e386e64a245c service nova] [instance: 896dd689-4810-4f23-af0a-d2f557a0796b] Refreshing network info cache for port 53407134-6b24-4ff7-b5d0-ef4cf1885933 {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 606.302968] env[61923]: DEBUG oslo_concurrency.lockutils [None req-0c55849d-766e-4fac-b1c4-e4e97ba688a6 tempest-DeleteServersAdminTestJSON-590116201 tempest-DeleteServersAdminTestJSON-590116201-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 606.429967] env[61923]: DEBUG oslo_concurrency.lockutils [None req-ae039c28-826e-4a1e-bb21-b1102865618f tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] Acquiring lock "93818ffd-eb85-48f0-aebd-9774467b0a8f" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 606.613262] env[61923]: DEBUG oslo_concurrency.lockutils [None req-cfefea89-2815-4df2-96e2-c7370c791e5f tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] Acquiring lock "9c3e2f94-91da-4255-b4d6-1c01e100ff48" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 606.644292] env[61923]: ERROR nova.compute.manager [None req-f3aa5656-98df-4c46-9436-b968a8770dbc tempest-ListServerFiltersTestJSON-483030130 tempest-ListServerFiltersTestJSON-483030130-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 53407134-6b24-4ff7-b5d0-ef4cf1885933, please check neutron logs for more information. [ 606.644292] env[61923]: ERROR nova.compute.manager Traceback (most recent call last): [ 606.644292] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 606.644292] env[61923]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 606.644292] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 606.644292] env[61923]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 606.644292] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 606.644292] env[61923]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 606.644292] env[61923]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 606.644292] env[61923]: ERROR nova.compute.manager self.force_reraise() [ 606.644292] env[61923]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 606.644292] env[61923]: ERROR nova.compute.manager raise self.value [ 606.644292] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 606.644292] env[61923]: ERROR nova.compute.manager updated_port = self._update_port( [ 606.644292] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 606.644292] env[61923]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 606.646416] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 606.646416] env[61923]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 606.646416] env[61923]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 53407134-6b24-4ff7-b5d0-ef4cf1885933, please check neutron logs for more information. [ 606.646416] env[61923]: ERROR nova.compute.manager [ 606.646416] env[61923]: Traceback (most recent call last): [ 606.646416] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 606.646416] env[61923]: listener.cb(fileno) [ 606.646416] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 606.646416] env[61923]: result = function(*args, **kwargs) [ 606.646416] env[61923]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 606.646416] env[61923]: return func(*args, **kwargs) [ 606.646416] env[61923]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 606.646416] env[61923]: raise e [ 606.646416] env[61923]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 606.646416] env[61923]: nwinfo = self.network_api.allocate_for_instance( [ 606.646416] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 606.646416] env[61923]: created_port_ids = self._update_ports_for_instance( [ 606.646416] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 606.646416] env[61923]: with excutils.save_and_reraise_exception(): [ 606.646416] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 606.646416] env[61923]: self.force_reraise() [ 606.646416] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 606.646416] env[61923]: raise self.value [ 606.646416] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 606.646416] env[61923]: updated_port = self._update_port( [ 606.646416] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 606.646416] env[61923]: _ensure_no_port_binding_failure(port) [ 606.646416] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 606.646416] env[61923]: raise exception.PortBindingFailed(port_id=port['id']) [ 606.647151] env[61923]: nova.exception.PortBindingFailed: Binding failed for port 53407134-6b24-4ff7-b5d0-ef4cf1885933, please check neutron logs for more information. [ 606.647151] env[61923]: Removing descriptor: 14 [ 606.647151] env[61923]: ERROR nova.compute.manager [None req-f3aa5656-98df-4c46-9436-b968a8770dbc tempest-ListServerFiltersTestJSON-483030130 tempest-ListServerFiltersTestJSON-483030130-project-member] [instance: 896dd689-4810-4f23-af0a-d2f557a0796b] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 53407134-6b24-4ff7-b5d0-ef4cf1885933, please check neutron logs for more information. [ 606.647151] env[61923]: ERROR nova.compute.manager [instance: 896dd689-4810-4f23-af0a-d2f557a0796b] Traceback (most recent call last): [ 606.647151] env[61923]: ERROR nova.compute.manager [instance: 896dd689-4810-4f23-af0a-d2f557a0796b] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 606.647151] env[61923]: ERROR nova.compute.manager [instance: 896dd689-4810-4f23-af0a-d2f557a0796b] yield resources [ 606.647151] env[61923]: ERROR nova.compute.manager [instance: 896dd689-4810-4f23-af0a-d2f557a0796b] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 606.647151] env[61923]: ERROR nova.compute.manager [instance: 896dd689-4810-4f23-af0a-d2f557a0796b] self.driver.spawn(context, instance, image_meta, [ 606.647151] env[61923]: ERROR nova.compute.manager [instance: 896dd689-4810-4f23-af0a-d2f557a0796b] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 606.647151] env[61923]: ERROR nova.compute.manager [instance: 896dd689-4810-4f23-af0a-d2f557a0796b] self._vmops.spawn(context, instance, image_meta, injected_files, [ 606.647151] env[61923]: ERROR nova.compute.manager [instance: 896dd689-4810-4f23-af0a-d2f557a0796b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 606.647151] env[61923]: ERROR nova.compute.manager [instance: 896dd689-4810-4f23-af0a-d2f557a0796b] vm_ref = self.build_virtual_machine(instance, [ 606.647455] env[61923]: ERROR nova.compute.manager [instance: 896dd689-4810-4f23-af0a-d2f557a0796b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 606.647455] env[61923]: ERROR nova.compute.manager [instance: 896dd689-4810-4f23-af0a-d2f557a0796b] vif_infos = vmwarevif.get_vif_info(self._session, [ 606.647455] env[61923]: ERROR nova.compute.manager [instance: 896dd689-4810-4f23-af0a-d2f557a0796b] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 606.647455] env[61923]: ERROR nova.compute.manager [instance: 896dd689-4810-4f23-af0a-d2f557a0796b] for vif in network_info: [ 606.647455] env[61923]: ERROR nova.compute.manager [instance: 896dd689-4810-4f23-af0a-d2f557a0796b] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 606.647455] env[61923]: ERROR nova.compute.manager [instance: 896dd689-4810-4f23-af0a-d2f557a0796b] return self._sync_wrapper(fn, *args, **kwargs) [ 606.647455] env[61923]: ERROR nova.compute.manager [instance: 896dd689-4810-4f23-af0a-d2f557a0796b] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 606.647455] env[61923]: ERROR nova.compute.manager [instance: 896dd689-4810-4f23-af0a-d2f557a0796b] self.wait() [ 606.647455] env[61923]: ERROR nova.compute.manager [instance: 896dd689-4810-4f23-af0a-d2f557a0796b] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 606.647455] env[61923]: ERROR nova.compute.manager [instance: 896dd689-4810-4f23-af0a-d2f557a0796b] self[:] = self._gt.wait() [ 606.647455] env[61923]: ERROR nova.compute.manager [instance: 896dd689-4810-4f23-af0a-d2f557a0796b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 606.647455] env[61923]: ERROR nova.compute.manager [instance: 896dd689-4810-4f23-af0a-d2f557a0796b] return self._exit_event.wait() [ 606.647455] env[61923]: ERROR nova.compute.manager [instance: 896dd689-4810-4f23-af0a-d2f557a0796b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 606.647827] env[61923]: ERROR nova.compute.manager [instance: 896dd689-4810-4f23-af0a-d2f557a0796b] result = hub.switch() [ 606.647827] env[61923]: ERROR nova.compute.manager [instance: 896dd689-4810-4f23-af0a-d2f557a0796b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 606.647827] env[61923]: ERROR nova.compute.manager [instance: 896dd689-4810-4f23-af0a-d2f557a0796b] return self.greenlet.switch() [ 606.647827] env[61923]: ERROR nova.compute.manager [instance: 896dd689-4810-4f23-af0a-d2f557a0796b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 606.647827] env[61923]: ERROR nova.compute.manager [instance: 896dd689-4810-4f23-af0a-d2f557a0796b] result = function(*args, **kwargs) [ 606.647827] env[61923]: ERROR nova.compute.manager [instance: 896dd689-4810-4f23-af0a-d2f557a0796b] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 606.647827] env[61923]: ERROR nova.compute.manager [instance: 896dd689-4810-4f23-af0a-d2f557a0796b] return func(*args, **kwargs) [ 606.647827] env[61923]: ERROR nova.compute.manager [instance: 896dd689-4810-4f23-af0a-d2f557a0796b] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 606.647827] env[61923]: ERROR nova.compute.manager [instance: 896dd689-4810-4f23-af0a-d2f557a0796b] raise e [ 606.647827] env[61923]: ERROR nova.compute.manager [instance: 896dd689-4810-4f23-af0a-d2f557a0796b] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 606.647827] env[61923]: ERROR nova.compute.manager [instance: 896dd689-4810-4f23-af0a-d2f557a0796b] nwinfo = self.network_api.allocate_for_instance( [ 606.647827] env[61923]: ERROR nova.compute.manager [instance: 896dd689-4810-4f23-af0a-d2f557a0796b] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 606.647827] env[61923]: ERROR nova.compute.manager [instance: 896dd689-4810-4f23-af0a-d2f557a0796b] created_port_ids = self._update_ports_for_instance( [ 606.648171] env[61923]: ERROR nova.compute.manager [instance: 896dd689-4810-4f23-af0a-d2f557a0796b] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 606.648171] env[61923]: ERROR nova.compute.manager [instance: 896dd689-4810-4f23-af0a-d2f557a0796b] with excutils.save_and_reraise_exception(): [ 606.648171] env[61923]: ERROR nova.compute.manager [instance: 896dd689-4810-4f23-af0a-d2f557a0796b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 606.648171] env[61923]: ERROR nova.compute.manager [instance: 896dd689-4810-4f23-af0a-d2f557a0796b] self.force_reraise() [ 606.648171] env[61923]: ERROR nova.compute.manager [instance: 896dd689-4810-4f23-af0a-d2f557a0796b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 606.648171] env[61923]: ERROR nova.compute.manager [instance: 896dd689-4810-4f23-af0a-d2f557a0796b] raise self.value [ 606.648171] env[61923]: ERROR nova.compute.manager [instance: 896dd689-4810-4f23-af0a-d2f557a0796b] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 606.648171] env[61923]: ERROR nova.compute.manager [instance: 896dd689-4810-4f23-af0a-d2f557a0796b] updated_port = self._update_port( [ 606.648171] env[61923]: ERROR nova.compute.manager [instance: 896dd689-4810-4f23-af0a-d2f557a0796b] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 606.648171] env[61923]: ERROR nova.compute.manager [instance: 896dd689-4810-4f23-af0a-d2f557a0796b] _ensure_no_port_binding_failure(port) [ 606.648171] env[61923]: ERROR nova.compute.manager [instance: 896dd689-4810-4f23-af0a-d2f557a0796b] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 606.648171] env[61923]: ERROR nova.compute.manager [instance: 896dd689-4810-4f23-af0a-d2f557a0796b] raise exception.PortBindingFailed(port_id=port['id']) [ 606.648474] env[61923]: ERROR nova.compute.manager [instance: 896dd689-4810-4f23-af0a-d2f557a0796b] nova.exception.PortBindingFailed: Binding failed for port 53407134-6b24-4ff7-b5d0-ef4cf1885933, please check neutron logs for more information. [ 606.648474] env[61923]: ERROR nova.compute.manager [instance: 896dd689-4810-4f23-af0a-d2f557a0796b] [ 606.648474] env[61923]: INFO nova.compute.manager [None req-f3aa5656-98df-4c46-9436-b968a8770dbc tempest-ListServerFiltersTestJSON-483030130 tempest-ListServerFiltersTestJSON-483030130-project-member] [instance: 896dd689-4810-4f23-af0a-d2f557a0796b] Terminating instance [ 606.648474] env[61923]: DEBUG oslo_concurrency.lockutils [None req-f3aa5656-98df-4c46-9436-b968a8770dbc tempest-ListServerFiltersTestJSON-483030130 tempest-ListServerFiltersTestJSON-483030130-project-member] Acquiring lock "refresh_cache-896dd689-4810-4f23-af0a-d2f557a0796b" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 606.740281] env[61923]: DEBUG oslo_concurrency.lockutils [None req-47902537-98f9-4fe7-afdb-4b2d4d440d10 tempest-ServersAdmin275Test-1304626830 tempest-ServersAdmin275Test-1304626830-project-member] Lock "f7cc960c-e06a-4c58-9367-ec8771fe09d5" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 24.140s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 606.818658] env[61923]: DEBUG nova.network.neutron [req-141ed8ab-3127-4ad5-9869-de706c9f0f40 req-c6f1014c-ae25-4f9c-ab1b-e386e64a245c service nova] [instance: 896dd689-4810-4f23-af0a-d2f557a0796b] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 606.971728] env[61923]: DEBUG nova.network.neutron [req-141ed8ab-3127-4ad5-9869-de706c9f0f40 req-c6f1014c-ae25-4f9c-ab1b-e386e64a245c service nova] [instance: 896dd689-4810-4f23-af0a-d2f557a0796b] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 607.178532] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e826d2a-022a-4bd3-96e4-dd44566dc370 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 607.187406] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e64613d-e43a-4a65-8cb7-adf1224783cc {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 607.222844] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6eb43762-0e92-401e-9971-280d371e9aa0 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 607.233103] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2ca4871-01c5-4b4d-9ff5-2ee99526c700 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 607.247038] env[61923]: DEBUG nova.compute.provider_tree [None req-fba27a13-474d-4fe3-9187-c31143f40aa5 tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 607.477428] env[61923]: DEBUG oslo_concurrency.lockutils [req-141ed8ab-3127-4ad5-9869-de706c9f0f40 req-c6f1014c-ae25-4f9c-ab1b-e386e64a245c service nova] Releasing lock "refresh_cache-896dd689-4810-4f23-af0a-d2f557a0796b" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 607.478123] env[61923]: DEBUG oslo_concurrency.lockutils [None req-f3aa5656-98df-4c46-9436-b968a8770dbc tempest-ListServerFiltersTestJSON-483030130 tempest-ListServerFiltersTestJSON-483030130-project-member] Acquired lock "refresh_cache-896dd689-4810-4f23-af0a-d2f557a0796b" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 607.478327] env[61923]: DEBUG nova.network.neutron [None req-f3aa5656-98df-4c46-9436-b968a8770dbc tempest-ListServerFiltersTestJSON-483030130 tempest-ListServerFiltersTestJSON-483030130-project-member] [instance: 896dd689-4810-4f23-af0a-d2f557a0796b] Building network info cache for instance {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 607.750733] env[61923]: DEBUG nova.scheduler.client.report [None req-fba27a13-474d-4fe3-9187-c31143f40aa5 tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 608.011684] env[61923]: DEBUG nova.network.neutron [None req-f3aa5656-98df-4c46-9436-b968a8770dbc tempest-ListServerFiltersTestJSON-483030130 tempest-ListServerFiltersTestJSON-483030130-project-member] [instance: 896dd689-4810-4f23-af0a-d2f557a0796b] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 608.189879] env[61923]: DEBUG nova.network.neutron [None req-f3aa5656-98df-4c46-9436-b968a8770dbc tempest-ListServerFiltersTestJSON-483030130 tempest-ListServerFiltersTestJSON-483030130-project-member] [instance: 896dd689-4810-4f23-af0a-d2f557a0796b] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 608.258324] env[61923]: DEBUG oslo_concurrency.lockutils [None req-fba27a13-474d-4fe3-9187-c31143f40aa5 tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.067s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 608.259025] env[61923]: ERROR nova.compute.manager [None req-fba27a13-474d-4fe3-9187-c31143f40aa5 tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] [instance: 9c3e2f94-91da-4255-b4d6-1c01e100ff48] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port cf3f1ca2-6afa-4f12-99b7-24adee47529f, please check neutron logs for more information. [ 608.259025] env[61923]: ERROR nova.compute.manager [instance: 9c3e2f94-91da-4255-b4d6-1c01e100ff48] Traceback (most recent call last): [ 608.259025] env[61923]: ERROR nova.compute.manager [instance: 9c3e2f94-91da-4255-b4d6-1c01e100ff48] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 608.259025] env[61923]: ERROR nova.compute.manager [instance: 9c3e2f94-91da-4255-b4d6-1c01e100ff48] self.driver.spawn(context, instance, image_meta, [ 608.259025] env[61923]: ERROR nova.compute.manager [instance: 9c3e2f94-91da-4255-b4d6-1c01e100ff48] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 608.259025] env[61923]: ERROR nova.compute.manager [instance: 9c3e2f94-91da-4255-b4d6-1c01e100ff48] self._vmops.spawn(context, instance, image_meta, injected_files, [ 608.259025] env[61923]: ERROR nova.compute.manager [instance: 9c3e2f94-91da-4255-b4d6-1c01e100ff48] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 608.259025] env[61923]: ERROR nova.compute.manager [instance: 9c3e2f94-91da-4255-b4d6-1c01e100ff48] vm_ref = self.build_virtual_machine(instance, [ 608.259025] env[61923]: ERROR nova.compute.manager [instance: 9c3e2f94-91da-4255-b4d6-1c01e100ff48] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 608.259025] env[61923]: ERROR nova.compute.manager [instance: 9c3e2f94-91da-4255-b4d6-1c01e100ff48] vif_infos = vmwarevif.get_vif_info(self._session, [ 608.259025] env[61923]: ERROR nova.compute.manager [instance: 9c3e2f94-91da-4255-b4d6-1c01e100ff48] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 608.259428] env[61923]: ERROR nova.compute.manager [instance: 9c3e2f94-91da-4255-b4d6-1c01e100ff48] for vif in network_info: [ 608.259428] env[61923]: ERROR nova.compute.manager [instance: 9c3e2f94-91da-4255-b4d6-1c01e100ff48] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 608.259428] env[61923]: ERROR nova.compute.manager [instance: 9c3e2f94-91da-4255-b4d6-1c01e100ff48] return self._sync_wrapper(fn, *args, **kwargs) [ 608.259428] env[61923]: ERROR nova.compute.manager [instance: 9c3e2f94-91da-4255-b4d6-1c01e100ff48] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 608.259428] env[61923]: ERROR nova.compute.manager [instance: 9c3e2f94-91da-4255-b4d6-1c01e100ff48] self.wait() [ 608.259428] env[61923]: ERROR nova.compute.manager [instance: 9c3e2f94-91da-4255-b4d6-1c01e100ff48] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 608.259428] env[61923]: ERROR nova.compute.manager [instance: 9c3e2f94-91da-4255-b4d6-1c01e100ff48] self[:] = self._gt.wait() [ 608.259428] env[61923]: ERROR nova.compute.manager [instance: 9c3e2f94-91da-4255-b4d6-1c01e100ff48] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 608.259428] env[61923]: ERROR nova.compute.manager [instance: 9c3e2f94-91da-4255-b4d6-1c01e100ff48] return self._exit_event.wait() [ 608.259428] env[61923]: ERROR nova.compute.manager [instance: 9c3e2f94-91da-4255-b4d6-1c01e100ff48] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 608.259428] env[61923]: ERROR nova.compute.manager [instance: 9c3e2f94-91da-4255-b4d6-1c01e100ff48] result = hub.switch() [ 608.259428] env[61923]: ERROR nova.compute.manager [instance: 9c3e2f94-91da-4255-b4d6-1c01e100ff48] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 608.259428] env[61923]: ERROR nova.compute.manager [instance: 9c3e2f94-91da-4255-b4d6-1c01e100ff48] return self.greenlet.switch() [ 608.259789] env[61923]: ERROR nova.compute.manager [instance: 9c3e2f94-91da-4255-b4d6-1c01e100ff48] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 608.259789] env[61923]: ERROR nova.compute.manager [instance: 9c3e2f94-91da-4255-b4d6-1c01e100ff48] result = function(*args, **kwargs) [ 608.259789] env[61923]: ERROR nova.compute.manager [instance: 9c3e2f94-91da-4255-b4d6-1c01e100ff48] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 608.259789] env[61923]: ERROR nova.compute.manager [instance: 9c3e2f94-91da-4255-b4d6-1c01e100ff48] return func(*args, **kwargs) [ 608.259789] env[61923]: ERROR nova.compute.manager [instance: 9c3e2f94-91da-4255-b4d6-1c01e100ff48] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 608.259789] env[61923]: ERROR nova.compute.manager [instance: 9c3e2f94-91da-4255-b4d6-1c01e100ff48] raise e [ 608.259789] env[61923]: ERROR nova.compute.manager [instance: 9c3e2f94-91da-4255-b4d6-1c01e100ff48] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 608.259789] env[61923]: ERROR nova.compute.manager [instance: 9c3e2f94-91da-4255-b4d6-1c01e100ff48] nwinfo = self.network_api.allocate_for_instance( [ 608.259789] env[61923]: ERROR nova.compute.manager [instance: 9c3e2f94-91da-4255-b4d6-1c01e100ff48] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 608.259789] env[61923]: ERROR nova.compute.manager [instance: 9c3e2f94-91da-4255-b4d6-1c01e100ff48] created_port_ids = self._update_ports_for_instance( [ 608.259789] env[61923]: ERROR nova.compute.manager [instance: 9c3e2f94-91da-4255-b4d6-1c01e100ff48] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 608.259789] env[61923]: ERROR nova.compute.manager [instance: 9c3e2f94-91da-4255-b4d6-1c01e100ff48] with excutils.save_and_reraise_exception(): [ 608.259789] env[61923]: ERROR nova.compute.manager [instance: 9c3e2f94-91da-4255-b4d6-1c01e100ff48] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 608.260112] env[61923]: ERROR nova.compute.manager [instance: 9c3e2f94-91da-4255-b4d6-1c01e100ff48] self.force_reraise() [ 608.260112] env[61923]: ERROR nova.compute.manager [instance: 9c3e2f94-91da-4255-b4d6-1c01e100ff48] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 608.260112] env[61923]: ERROR nova.compute.manager [instance: 9c3e2f94-91da-4255-b4d6-1c01e100ff48] raise self.value [ 608.260112] env[61923]: ERROR nova.compute.manager [instance: 9c3e2f94-91da-4255-b4d6-1c01e100ff48] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 608.260112] env[61923]: ERROR nova.compute.manager [instance: 9c3e2f94-91da-4255-b4d6-1c01e100ff48] updated_port = self._update_port( [ 608.260112] env[61923]: ERROR nova.compute.manager [instance: 9c3e2f94-91da-4255-b4d6-1c01e100ff48] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 608.260112] env[61923]: ERROR nova.compute.manager [instance: 9c3e2f94-91da-4255-b4d6-1c01e100ff48] _ensure_no_port_binding_failure(port) [ 608.260112] env[61923]: ERROR nova.compute.manager [instance: 9c3e2f94-91da-4255-b4d6-1c01e100ff48] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 608.260112] env[61923]: ERROR nova.compute.manager [instance: 9c3e2f94-91da-4255-b4d6-1c01e100ff48] raise exception.PortBindingFailed(port_id=port['id']) [ 608.260112] env[61923]: ERROR nova.compute.manager [instance: 9c3e2f94-91da-4255-b4d6-1c01e100ff48] nova.exception.PortBindingFailed: Binding failed for port cf3f1ca2-6afa-4f12-99b7-24adee47529f, please check neutron logs for more information. [ 608.260112] env[61923]: ERROR nova.compute.manager [instance: 9c3e2f94-91da-4255-b4d6-1c01e100ff48] [ 608.260375] env[61923]: DEBUG nova.compute.utils [None req-fba27a13-474d-4fe3-9187-c31143f40aa5 tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] [instance: 9c3e2f94-91da-4255-b4d6-1c01e100ff48] Binding failed for port cf3f1ca2-6afa-4f12-99b7-24adee47529f, please check neutron logs for more information. {{(pid=61923) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 608.262034] env[61923]: DEBUG oslo_concurrency.lockutils [None req-fba27a13-474d-4fe3-9187-c31143f40aa5 tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 19.613s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 608.269136] env[61923]: DEBUG nova.compute.manager [None req-fba27a13-474d-4fe3-9187-c31143f40aa5 tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] [instance: 9c3e2f94-91da-4255-b4d6-1c01e100ff48] Build of instance 9c3e2f94-91da-4255-b4d6-1c01e100ff48 was re-scheduled: Binding failed for port cf3f1ca2-6afa-4f12-99b7-24adee47529f, please check neutron logs for more information. {{(pid=61923) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 608.269136] env[61923]: DEBUG nova.compute.manager [None req-fba27a13-474d-4fe3-9187-c31143f40aa5 tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] [instance: 9c3e2f94-91da-4255-b4d6-1c01e100ff48] Unplugging VIFs for instance {{(pid=61923) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 608.269136] env[61923]: DEBUG oslo_concurrency.lockutils [None req-fba27a13-474d-4fe3-9187-c31143f40aa5 tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] Acquiring lock "refresh_cache-9c3e2f94-91da-4255-b4d6-1c01e100ff48" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 608.269136] env[61923]: DEBUG oslo_concurrency.lockutils [None req-fba27a13-474d-4fe3-9187-c31143f40aa5 tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] Acquired lock "refresh_cache-9c3e2f94-91da-4255-b4d6-1c01e100ff48" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 608.269338] env[61923]: DEBUG nova.network.neutron [None req-fba27a13-474d-4fe3-9187-c31143f40aa5 tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] [instance: 9c3e2f94-91da-4255-b4d6-1c01e100ff48] Building network info cache for instance {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 608.504921] env[61923]: DEBUG nova.compute.manager [req-4a58a632-1718-4f89-8f72-338ba1bb8d0c req-6c85e9a2-12f7-47aa-9bff-2f9458dc3f97 service nova] [instance: 896dd689-4810-4f23-af0a-d2f557a0796b] Received event network-vif-deleted-53407134-6b24-4ff7-b5d0-ef4cf1885933 {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 608.693299] env[61923]: DEBUG oslo_concurrency.lockutils [None req-f3aa5656-98df-4c46-9436-b968a8770dbc tempest-ListServerFiltersTestJSON-483030130 tempest-ListServerFiltersTestJSON-483030130-project-member] Releasing lock "refresh_cache-896dd689-4810-4f23-af0a-d2f557a0796b" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 608.693720] env[61923]: DEBUG nova.compute.manager [None req-f3aa5656-98df-4c46-9436-b968a8770dbc tempest-ListServerFiltersTestJSON-483030130 tempest-ListServerFiltersTestJSON-483030130-project-member] [instance: 896dd689-4810-4f23-af0a-d2f557a0796b] Start destroying the instance on the hypervisor. {{(pid=61923) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 608.693936] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-f3aa5656-98df-4c46-9436-b968a8770dbc tempest-ListServerFiltersTestJSON-483030130 tempest-ListServerFiltersTestJSON-483030130-project-member] [instance: 896dd689-4810-4f23-af0a-d2f557a0796b] Destroying instance {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 608.694414] env[61923]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-4f93e708-c6df-4576-b8f5-a7075a2d377b {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.708324] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3046484c-d4c1-4f72-842c-4ad43ebd5157 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.735527] env[61923]: WARNING nova.virt.vmwareapi.vmops [None req-f3aa5656-98df-4c46-9436-b968a8770dbc tempest-ListServerFiltersTestJSON-483030130 tempest-ListServerFiltersTestJSON-483030130-project-member] [instance: 896dd689-4810-4f23-af0a-d2f557a0796b] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 896dd689-4810-4f23-af0a-d2f557a0796b could not be found. [ 608.735770] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-f3aa5656-98df-4c46-9436-b968a8770dbc tempest-ListServerFiltersTestJSON-483030130 tempest-ListServerFiltersTestJSON-483030130-project-member] [instance: 896dd689-4810-4f23-af0a-d2f557a0796b] Instance destroyed {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 608.736226] env[61923]: INFO nova.compute.manager [None req-f3aa5656-98df-4c46-9436-b968a8770dbc tempest-ListServerFiltersTestJSON-483030130 tempest-ListServerFiltersTestJSON-483030130-project-member] [instance: 896dd689-4810-4f23-af0a-d2f557a0796b] Took 0.04 seconds to destroy the instance on the hypervisor. [ 608.736830] env[61923]: DEBUG oslo.service.loopingcall [None req-f3aa5656-98df-4c46-9436-b968a8770dbc tempest-ListServerFiltersTestJSON-483030130 tempest-ListServerFiltersTestJSON-483030130-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61923) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 608.737506] env[61923]: DEBUG nova.compute.manager [-] [instance: 896dd689-4810-4f23-af0a-d2f557a0796b] Deallocating network for instance {{(pid=61923) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 608.737506] env[61923]: DEBUG nova.network.neutron [-] [instance: 896dd689-4810-4f23-af0a-d2f557a0796b] deallocate_for_instance() {{(pid=61923) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 608.755249] env[61923]: DEBUG nova.network.neutron [-] [instance: 896dd689-4810-4f23-af0a-d2f557a0796b] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 608.811171] env[61923]: DEBUG nova.network.neutron [None req-fba27a13-474d-4fe3-9187-c31143f40aa5 tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] [instance: 9c3e2f94-91da-4255-b4d6-1c01e100ff48] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 608.913145] env[61923]: DEBUG nova.network.neutron [None req-fba27a13-474d-4fe3-9187-c31143f40aa5 tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] [instance: 9c3e2f94-91da-4255-b4d6-1c01e100ff48] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 609.231815] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6fae8a36-70f6-459e-aaa8-1a197fe01460 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.244756] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c6dfd30-0398-48ab-920c-6ad35b1803cb {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.279762] env[61923]: DEBUG nova.network.neutron [-] [instance: 896dd689-4810-4f23-af0a-d2f557a0796b] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 609.281923] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6fcc30d2-10b9-477b-ac06-aaa681813f54 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.294019] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8433a2aa-d5bf-4968-ae81-5fb4955ea4ca {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.306800] env[61923]: DEBUG nova.compute.provider_tree [None req-fba27a13-474d-4fe3-9187-c31143f40aa5 tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 609.417882] env[61923]: DEBUG oslo_concurrency.lockutils [None req-fba27a13-474d-4fe3-9187-c31143f40aa5 tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] Releasing lock "refresh_cache-9c3e2f94-91da-4255-b4d6-1c01e100ff48" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 609.418538] env[61923]: DEBUG nova.compute.manager [None req-fba27a13-474d-4fe3-9187-c31143f40aa5 tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61923) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 609.419045] env[61923]: DEBUG nova.compute.manager [None req-fba27a13-474d-4fe3-9187-c31143f40aa5 tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] [instance: 9c3e2f94-91da-4255-b4d6-1c01e100ff48] Deallocating network for instance {{(pid=61923) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 609.419279] env[61923]: DEBUG nova.network.neutron [None req-fba27a13-474d-4fe3-9187-c31143f40aa5 tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] [instance: 9c3e2f94-91da-4255-b4d6-1c01e100ff48] deallocate_for_instance() {{(pid=61923) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 609.441205] env[61923]: DEBUG nova.network.neutron [None req-fba27a13-474d-4fe3-9187-c31143f40aa5 tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] [instance: 9c3e2f94-91da-4255-b4d6-1c01e100ff48] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 609.785451] env[61923]: INFO nova.compute.manager [-] [instance: 896dd689-4810-4f23-af0a-d2f557a0796b] Took 1.05 seconds to deallocate network for instance. [ 609.788828] env[61923]: DEBUG nova.compute.claims [None req-f3aa5656-98df-4c46-9436-b968a8770dbc tempest-ListServerFiltersTestJSON-483030130 tempest-ListServerFiltersTestJSON-483030130-project-member] [instance: 896dd689-4810-4f23-af0a-d2f557a0796b] Aborting claim: {{(pid=61923) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 609.789010] env[61923]: DEBUG oslo_concurrency.lockutils [None req-f3aa5656-98df-4c46-9436-b968a8770dbc tempest-ListServerFiltersTestJSON-483030130 tempest-ListServerFiltersTestJSON-483030130-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 609.810619] env[61923]: DEBUG nova.scheduler.client.report [None req-fba27a13-474d-4fe3-9187-c31143f40aa5 tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 609.944266] env[61923]: DEBUG nova.network.neutron [None req-fba27a13-474d-4fe3-9187-c31143f40aa5 tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] [instance: 9c3e2f94-91da-4255-b4d6-1c01e100ff48] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 610.317470] env[61923]: DEBUG oslo_concurrency.lockutils [None req-fba27a13-474d-4fe3-9187-c31143f40aa5 tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.055s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 610.317669] env[61923]: ERROR nova.compute.manager [None req-fba27a13-474d-4fe3-9187-c31143f40aa5 tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] [instance: 93818ffd-eb85-48f0-aebd-9774467b0a8f] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port e0c01f6c-546d-4532-b730-becd9e5dfcda, please check neutron logs for more information. [ 610.317669] env[61923]: ERROR nova.compute.manager [instance: 93818ffd-eb85-48f0-aebd-9774467b0a8f] Traceback (most recent call last): [ 610.317669] env[61923]: ERROR nova.compute.manager [instance: 93818ffd-eb85-48f0-aebd-9774467b0a8f] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 610.317669] env[61923]: ERROR nova.compute.manager [instance: 93818ffd-eb85-48f0-aebd-9774467b0a8f] self.driver.spawn(context, instance, image_meta, [ 610.317669] env[61923]: ERROR nova.compute.manager [instance: 93818ffd-eb85-48f0-aebd-9774467b0a8f] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 610.317669] env[61923]: ERROR nova.compute.manager [instance: 93818ffd-eb85-48f0-aebd-9774467b0a8f] self._vmops.spawn(context, instance, image_meta, injected_files, [ 610.317669] env[61923]: ERROR nova.compute.manager [instance: 93818ffd-eb85-48f0-aebd-9774467b0a8f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 610.317669] env[61923]: ERROR nova.compute.manager [instance: 93818ffd-eb85-48f0-aebd-9774467b0a8f] vm_ref = self.build_virtual_machine(instance, [ 610.317669] env[61923]: ERROR nova.compute.manager [instance: 93818ffd-eb85-48f0-aebd-9774467b0a8f] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 610.317669] env[61923]: ERROR nova.compute.manager [instance: 93818ffd-eb85-48f0-aebd-9774467b0a8f] vif_infos = vmwarevif.get_vif_info(self._session, [ 610.317669] env[61923]: ERROR nova.compute.manager [instance: 93818ffd-eb85-48f0-aebd-9774467b0a8f] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 610.317962] env[61923]: ERROR nova.compute.manager [instance: 93818ffd-eb85-48f0-aebd-9774467b0a8f] for vif in network_info: [ 610.317962] env[61923]: ERROR nova.compute.manager [instance: 93818ffd-eb85-48f0-aebd-9774467b0a8f] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 610.317962] env[61923]: ERROR nova.compute.manager [instance: 93818ffd-eb85-48f0-aebd-9774467b0a8f] return self._sync_wrapper(fn, *args, **kwargs) [ 610.317962] env[61923]: ERROR nova.compute.manager [instance: 93818ffd-eb85-48f0-aebd-9774467b0a8f] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 610.317962] env[61923]: ERROR nova.compute.manager [instance: 93818ffd-eb85-48f0-aebd-9774467b0a8f] self.wait() [ 610.317962] env[61923]: ERROR nova.compute.manager [instance: 93818ffd-eb85-48f0-aebd-9774467b0a8f] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 610.317962] env[61923]: ERROR nova.compute.manager [instance: 93818ffd-eb85-48f0-aebd-9774467b0a8f] self[:] = self._gt.wait() [ 610.317962] env[61923]: ERROR nova.compute.manager [instance: 93818ffd-eb85-48f0-aebd-9774467b0a8f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 610.317962] env[61923]: ERROR nova.compute.manager [instance: 93818ffd-eb85-48f0-aebd-9774467b0a8f] return self._exit_event.wait() [ 610.317962] env[61923]: ERROR nova.compute.manager [instance: 93818ffd-eb85-48f0-aebd-9774467b0a8f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 610.317962] env[61923]: ERROR nova.compute.manager [instance: 93818ffd-eb85-48f0-aebd-9774467b0a8f] result = hub.switch() [ 610.317962] env[61923]: ERROR nova.compute.manager [instance: 93818ffd-eb85-48f0-aebd-9774467b0a8f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 610.317962] env[61923]: ERROR nova.compute.manager [instance: 93818ffd-eb85-48f0-aebd-9774467b0a8f] return self.greenlet.switch() [ 610.318266] env[61923]: ERROR nova.compute.manager [instance: 93818ffd-eb85-48f0-aebd-9774467b0a8f] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 610.318266] env[61923]: ERROR nova.compute.manager [instance: 93818ffd-eb85-48f0-aebd-9774467b0a8f] result = function(*args, **kwargs) [ 610.318266] env[61923]: ERROR nova.compute.manager [instance: 93818ffd-eb85-48f0-aebd-9774467b0a8f] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 610.318266] env[61923]: ERROR nova.compute.manager [instance: 93818ffd-eb85-48f0-aebd-9774467b0a8f] return func(*args, **kwargs) [ 610.318266] env[61923]: ERROR nova.compute.manager [instance: 93818ffd-eb85-48f0-aebd-9774467b0a8f] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 610.318266] env[61923]: ERROR nova.compute.manager [instance: 93818ffd-eb85-48f0-aebd-9774467b0a8f] raise e [ 610.318266] env[61923]: ERROR nova.compute.manager [instance: 93818ffd-eb85-48f0-aebd-9774467b0a8f] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 610.318266] env[61923]: ERROR nova.compute.manager [instance: 93818ffd-eb85-48f0-aebd-9774467b0a8f] nwinfo = self.network_api.allocate_for_instance( [ 610.318266] env[61923]: ERROR nova.compute.manager [instance: 93818ffd-eb85-48f0-aebd-9774467b0a8f] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 610.318266] env[61923]: ERROR nova.compute.manager [instance: 93818ffd-eb85-48f0-aebd-9774467b0a8f] created_port_ids = self._update_ports_for_instance( [ 610.318266] env[61923]: ERROR nova.compute.manager [instance: 93818ffd-eb85-48f0-aebd-9774467b0a8f] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 610.318266] env[61923]: ERROR nova.compute.manager [instance: 93818ffd-eb85-48f0-aebd-9774467b0a8f] with excutils.save_and_reraise_exception(): [ 610.318266] env[61923]: ERROR nova.compute.manager [instance: 93818ffd-eb85-48f0-aebd-9774467b0a8f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 610.318568] env[61923]: ERROR nova.compute.manager [instance: 93818ffd-eb85-48f0-aebd-9774467b0a8f] self.force_reraise() [ 610.318568] env[61923]: ERROR nova.compute.manager [instance: 93818ffd-eb85-48f0-aebd-9774467b0a8f] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 610.318568] env[61923]: ERROR nova.compute.manager [instance: 93818ffd-eb85-48f0-aebd-9774467b0a8f] raise self.value [ 610.318568] env[61923]: ERROR nova.compute.manager [instance: 93818ffd-eb85-48f0-aebd-9774467b0a8f] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 610.318568] env[61923]: ERROR nova.compute.manager [instance: 93818ffd-eb85-48f0-aebd-9774467b0a8f] updated_port = self._update_port( [ 610.318568] env[61923]: ERROR nova.compute.manager [instance: 93818ffd-eb85-48f0-aebd-9774467b0a8f] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 610.318568] env[61923]: ERROR nova.compute.manager [instance: 93818ffd-eb85-48f0-aebd-9774467b0a8f] _ensure_no_port_binding_failure(port) [ 610.318568] env[61923]: ERROR nova.compute.manager [instance: 93818ffd-eb85-48f0-aebd-9774467b0a8f] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 610.318568] env[61923]: ERROR nova.compute.manager [instance: 93818ffd-eb85-48f0-aebd-9774467b0a8f] raise exception.PortBindingFailed(port_id=port['id']) [ 610.318568] env[61923]: ERROR nova.compute.manager [instance: 93818ffd-eb85-48f0-aebd-9774467b0a8f] nova.exception.PortBindingFailed: Binding failed for port e0c01f6c-546d-4532-b730-becd9e5dfcda, please check neutron logs for more information. [ 610.318568] env[61923]: ERROR nova.compute.manager [instance: 93818ffd-eb85-48f0-aebd-9774467b0a8f] [ 610.318932] env[61923]: DEBUG nova.compute.utils [None req-fba27a13-474d-4fe3-9187-c31143f40aa5 tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] [instance: 93818ffd-eb85-48f0-aebd-9774467b0a8f] Binding failed for port e0c01f6c-546d-4532-b730-becd9e5dfcda, please check neutron logs for more information. {{(pid=61923) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 610.319611] env[61923]: DEBUG oslo_concurrency.lockutils [None req-543d2693-878a-4ade-aaa2-339016c76eca tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 18.281s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 610.322672] env[61923]: DEBUG nova.compute.manager [None req-fba27a13-474d-4fe3-9187-c31143f40aa5 tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] [instance: 93818ffd-eb85-48f0-aebd-9774467b0a8f] Build of instance 93818ffd-eb85-48f0-aebd-9774467b0a8f was re-scheduled: Binding failed for port e0c01f6c-546d-4532-b730-becd9e5dfcda, please check neutron logs for more information. {{(pid=61923) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 610.323023] env[61923]: DEBUG nova.compute.manager [None req-fba27a13-474d-4fe3-9187-c31143f40aa5 tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] [instance: 93818ffd-eb85-48f0-aebd-9774467b0a8f] Unplugging VIFs for instance {{(pid=61923) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 610.323311] env[61923]: DEBUG oslo_concurrency.lockutils [None req-fba27a13-474d-4fe3-9187-c31143f40aa5 tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] Acquiring lock "refresh_cache-93818ffd-eb85-48f0-aebd-9774467b0a8f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 610.323453] env[61923]: DEBUG oslo_concurrency.lockutils [None req-fba27a13-474d-4fe3-9187-c31143f40aa5 tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] Acquired lock "refresh_cache-93818ffd-eb85-48f0-aebd-9774467b0a8f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 610.323547] env[61923]: DEBUG nova.network.neutron [None req-fba27a13-474d-4fe3-9187-c31143f40aa5 tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] [instance: 93818ffd-eb85-48f0-aebd-9774467b0a8f] Building network info cache for instance {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 610.447528] env[61923]: INFO nova.compute.manager [None req-fba27a13-474d-4fe3-9187-c31143f40aa5 tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] [instance: 9c3e2f94-91da-4255-b4d6-1c01e100ff48] Took 1.03 seconds to deallocate network for instance. [ 610.854778] env[61923]: DEBUG nova.network.neutron [None req-fba27a13-474d-4fe3-9187-c31143f40aa5 tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] [instance: 93818ffd-eb85-48f0-aebd-9774467b0a8f] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 610.888326] env[61923]: DEBUG oslo_concurrency.lockutils [None req-13202777-73e8-40d3-a6ad-7bb49e61267a tempest-ImagesNegativeTestJSON-1202128281 tempest-ImagesNegativeTestJSON-1202128281-project-member] Acquiring lock "f7e1d517-3e39-4381-83be-0152368b8120" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 610.888607] env[61923]: DEBUG oslo_concurrency.lockutils [None req-13202777-73e8-40d3-a6ad-7bb49e61267a tempest-ImagesNegativeTestJSON-1202128281 tempest-ImagesNegativeTestJSON-1202128281-project-member] Lock "f7e1d517-3e39-4381-83be-0152368b8120" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 610.971140] env[61923]: DEBUG nova.network.neutron [None req-fba27a13-474d-4fe3-9187-c31143f40aa5 tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] [instance: 93818ffd-eb85-48f0-aebd-9774467b0a8f] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 611.285739] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9dc46abd-f4f2-4b51-ad3d-c74a7e972490 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.296175] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8154136a-debb-459a-b1e9-549c8000a66b {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.329297] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b15f6618-35d3-42c6-9379-c70ad20951fe {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.338127] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6addea6-9d22-4ab5-97f9-2cacf269e032 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.353081] env[61923]: DEBUG nova.compute.provider_tree [None req-543d2693-878a-4ade-aaa2-339016c76eca tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 611.476019] env[61923]: DEBUG oslo_concurrency.lockutils [None req-fba27a13-474d-4fe3-9187-c31143f40aa5 tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] Releasing lock "refresh_cache-93818ffd-eb85-48f0-aebd-9774467b0a8f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 611.476277] env[61923]: DEBUG nova.compute.manager [None req-fba27a13-474d-4fe3-9187-c31143f40aa5 tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61923) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 611.476473] env[61923]: DEBUG nova.compute.manager [None req-fba27a13-474d-4fe3-9187-c31143f40aa5 tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] [instance: 93818ffd-eb85-48f0-aebd-9774467b0a8f] Deallocating network for instance {{(pid=61923) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 611.476663] env[61923]: DEBUG nova.network.neutron [None req-fba27a13-474d-4fe3-9187-c31143f40aa5 tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] [instance: 93818ffd-eb85-48f0-aebd-9774467b0a8f] deallocate_for_instance() {{(pid=61923) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 611.491346] env[61923]: INFO nova.scheduler.client.report [None req-fba27a13-474d-4fe3-9187-c31143f40aa5 tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] Deleted allocations for instance 9c3e2f94-91da-4255-b4d6-1c01e100ff48 [ 611.500023] env[61923]: DEBUG nova.network.neutron [None req-fba27a13-474d-4fe3-9187-c31143f40aa5 tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] [instance: 93818ffd-eb85-48f0-aebd-9774467b0a8f] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 611.857898] env[61923]: DEBUG nova.scheduler.client.report [None req-543d2693-878a-4ade-aaa2-339016c76eca tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 612.002760] env[61923]: DEBUG nova.network.neutron [None req-fba27a13-474d-4fe3-9187-c31143f40aa5 tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] [instance: 93818ffd-eb85-48f0-aebd-9774467b0a8f] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 612.004181] env[61923]: DEBUG oslo_concurrency.lockutils [None req-fba27a13-474d-4fe3-9187-c31143f40aa5 tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] Lock "9c3e2f94-91da-4255-b4d6-1c01e100ff48" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 73.007s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 612.005825] env[61923]: DEBUG oslo_concurrency.lockutils [None req-cfefea89-2815-4df2-96e2-c7370c791e5f tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] Lock "9c3e2f94-91da-4255-b4d6-1c01e100ff48" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 5.393s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 612.005956] env[61923]: DEBUG oslo_concurrency.lockutils [None req-cfefea89-2815-4df2-96e2-c7370c791e5f tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] Acquiring lock "9c3e2f94-91da-4255-b4d6-1c01e100ff48-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 612.006180] env[61923]: DEBUG oslo_concurrency.lockutils [None req-cfefea89-2815-4df2-96e2-c7370c791e5f tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] Lock "9c3e2f94-91da-4255-b4d6-1c01e100ff48-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 612.006381] env[61923]: DEBUG oslo_concurrency.lockutils [None req-cfefea89-2815-4df2-96e2-c7370c791e5f tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] Lock "9c3e2f94-91da-4255-b4d6-1c01e100ff48-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 612.008227] env[61923]: INFO nova.compute.manager [None req-cfefea89-2815-4df2-96e2-c7370c791e5f tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] [instance: 9c3e2f94-91da-4255-b4d6-1c01e100ff48] Terminating instance [ 612.009843] env[61923]: DEBUG oslo_concurrency.lockutils [None req-cfefea89-2815-4df2-96e2-c7370c791e5f tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] Acquiring lock "refresh_cache-9c3e2f94-91da-4255-b4d6-1c01e100ff48" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 612.010012] env[61923]: DEBUG oslo_concurrency.lockutils [None req-cfefea89-2815-4df2-96e2-c7370c791e5f tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] Acquired lock "refresh_cache-9c3e2f94-91da-4255-b4d6-1c01e100ff48" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 612.010200] env[61923]: DEBUG nova.network.neutron [None req-cfefea89-2815-4df2-96e2-c7370c791e5f tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] [instance: 9c3e2f94-91da-4255-b4d6-1c01e100ff48] Building network info cache for instance {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 612.362584] env[61923]: DEBUG oslo_concurrency.lockutils [None req-543d2693-878a-4ade-aaa2-339016c76eca tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.043s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 612.363296] env[61923]: ERROR nova.compute.manager [None req-543d2693-878a-4ade-aaa2-339016c76eca tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] [instance: 045c0b05-333e-4f95-94b0-8f51d87dca0a] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 3d9f5735-20ed-42f4-9565-073437eef966, please check neutron logs for more information. [ 612.363296] env[61923]: ERROR nova.compute.manager [instance: 045c0b05-333e-4f95-94b0-8f51d87dca0a] Traceback (most recent call last): [ 612.363296] env[61923]: ERROR nova.compute.manager [instance: 045c0b05-333e-4f95-94b0-8f51d87dca0a] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 612.363296] env[61923]: ERROR nova.compute.manager [instance: 045c0b05-333e-4f95-94b0-8f51d87dca0a] self.driver.spawn(context, instance, image_meta, [ 612.363296] env[61923]: ERROR nova.compute.manager [instance: 045c0b05-333e-4f95-94b0-8f51d87dca0a] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 612.363296] env[61923]: ERROR nova.compute.manager [instance: 045c0b05-333e-4f95-94b0-8f51d87dca0a] self._vmops.spawn(context, instance, image_meta, injected_files, [ 612.363296] env[61923]: ERROR nova.compute.manager [instance: 045c0b05-333e-4f95-94b0-8f51d87dca0a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 612.363296] env[61923]: ERROR nova.compute.manager [instance: 045c0b05-333e-4f95-94b0-8f51d87dca0a] vm_ref = self.build_virtual_machine(instance, [ 612.363296] env[61923]: ERROR nova.compute.manager [instance: 045c0b05-333e-4f95-94b0-8f51d87dca0a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 612.363296] env[61923]: ERROR nova.compute.manager [instance: 045c0b05-333e-4f95-94b0-8f51d87dca0a] vif_infos = vmwarevif.get_vif_info(self._session, [ 612.363296] env[61923]: ERROR nova.compute.manager [instance: 045c0b05-333e-4f95-94b0-8f51d87dca0a] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 612.363591] env[61923]: ERROR nova.compute.manager [instance: 045c0b05-333e-4f95-94b0-8f51d87dca0a] for vif in network_info: [ 612.363591] env[61923]: ERROR nova.compute.manager [instance: 045c0b05-333e-4f95-94b0-8f51d87dca0a] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 612.363591] env[61923]: ERROR nova.compute.manager [instance: 045c0b05-333e-4f95-94b0-8f51d87dca0a] return self._sync_wrapper(fn, *args, **kwargs) [ 612.363591] env[61923]: ERROR nova.compute.manager [instance: 045c0b05-333e-4f95-94b0-8f51d87dca0a] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 612.363591] env[61923]: ERROR nova.compute.manager [instance: 045c0b05-333e-4f95-94b0-8f51d87dca0a] self.wait() [ 612.363591] env[61923]: ERROR nova.compute.manager [instance: 045c0b05-333e-4f95-94b0-8f51d87dca0a] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 612.363591] env[61923]: ERROR nova.compute.manager [instance: 045c0b05-333e-4f95-94b0-8f51d87dca0a] self[:] = self._gt.wait() [ 612.363591] env[61923]: ERROR nova.compute.manager [instance: 045c0b05-333e-4f95-94b0-8f51d87dca0a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 612.363591] env[61923]: ERROR nova.compute.manager [instance: 045c0b05-333e-4f95-94b0-8f51d87dca0a] return self._exit_event.wait() [ 612.363591] env[61923]: ERROR nova.compute.manager [instance: 045c0b05-333e-4f95-94b0-8f51d87dca0a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 612.363591] env[61923]: ERROR nova.compute.manager [instance: 045c0b05-333e-4f95-94b0-8f51d87dca0a] result = hub.switch() [ 612.363591] env[61923]: ERROR nova.compute.manager [instance: 045c0b05-333e-4f95-94b0-8f51d87dca0a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 612.363591] env[61923]: ERROR nova.compute.manager [instance: 045c0b05-333e-4f95-94b0-8f51d87dca0a] return self.greenlet.switch() [ 612.363869] env[61923]: ERROR nova.compute.manager [instance: 045c0b05-333e-4f95-94b0-8f51d87dca0a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 612.363869] env[61923]: ERROR nova.compute.manager [instance: 045c0b05-333e-4f95-94b0-8f51d87dca0a] result = function(*args, **kwargs) [ 612.363869] env[61923]: ERROR nova.compute.manager [instance: 045c0b05-333e-4f95-94b0-8f51d87dca0a] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 612.363869] env[61923]: ERROR nova.compute.manager [instance: 045c0b05-333e-4f95-94b0-8f51d87dca0a] return func(*args, **kwargs) [ 612.363869] env[61923]: ERROR nova.compute.manager [instance: 045c0b05-333e-4f95-94b0-8f51d87dca0a] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 612.363869] env[61923]: ERROR nova.compute.manager [instance: 045c0b05-333e-4f95-94b0-8f51d87dca0a] raise e [ 612.363869] env[61923]: ERROR nova.compute.manager [instance: 045c0b05-333e-4f95-94b0-8f51d87dca0a] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 612.363869] env[61923]: ERROR nova.compute.manager [instance: 045c0b05-333e-4f95-94b0-8f51d87dca0a] nwinfo = self.network_api.allocate_for_instance( [ 612.363869] env[61923]: ERROR nova.compute.manager [instance: 045c0b05-333e-4f95-94b0-8f51d87dca0a] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 612.363869] env[61923]: ERROR nova.compute.manager [instance: 045c0b05-333e-4f95-94b0-8f51d87dca0a] created_port_ids = self._update_ports_for_instance( [ 612.363869] env[61923]: ERROR nova.compute.manager [instance: 045c0b05-333e-4f95-94b0-8f51d87dca0a] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 612.363869] env[61923]: ERROR nova.compute.manager [instance: 045c0b05-333e-4f95-94b0-8f51d87dca0a] with excutils.save_and_reraise_exception(): [ 612.363869] env[61923]: ERROR nova.compute.manager [instance: 045c0b05-333e-4f95-94b0-8f51d87dca0a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 612.364164] env[61923]: ERROR nova.compute.manager [instance: 045c0b05-333e-4f95-94b0-8f51d87dca0a] self.force_reraise() [ 612.364164] env[61923]: ERROR nova.compute.manager [instance: 045c0b05-333e-4f95-94b0-8f51d87dca0a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 612.364164] env[61923]: ERROR nova.compute.manager [instance: 045c0b05-333e-4f95-94b0-8f51d87dca0a] raise self.value [ 612.364164] env[61923]: ERROR nova.compute.manager [instance: 045c0b05-333e-4f95-94b0-8f51d87dca0a] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 612.364164] env[61923]: ERROR nova.compute.manager [instance: 045c0b05-333e-4f95-94b0-8f51d87dca0a] updated_port = self._update_port( [ 612.364164] env[61923]: ERROR nova.compute.manager [instance: 045c0b05-333e-4f95-94b0-8f51d87dca0a] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 612.364164] env[61923]: ERROR nova.compute.manager [instance: 045c0b05-333e-4f95-94b0-8f51d87dca0a] _ensure_no_port_binding_failure(port) [ 612.364164] env[61923]: ERROR nova.compute.manager [instance: 045c0b05-333e-4f95-94b0-8f51d87dca0a] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 612.364164] env[61923]: ERROR nova.compute.manager [instance: 045c0b05-333e-4f95-94b0-8f51d87dca0a] raise exception.PortBindingFailed(port_id=port['id']) [ 612.364164] env[61923]: ERROR nova.compute.manager [instance: 045c0b05-333e-4f95-94b0-8f51d87dca0a] nova.exception.PortBindingFailed: Binding failed for port 3d9f5735-20ed-42f4-9565-073437eef966, please check neutron logs for more information. [ 612.364164] env[61923]: ERROR nova.compute.manager [instance: 045c0b05-333e-4f95-94b0-8f51d87dca0a] [ 612.364415] env[61923]: DEBUG nova.compute.utils [None req-543d2693-878a-4ade-aaa2-339016c76eca tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] [instance: 045c0b05-333e-4f95-94b0-8f51d87dca0a] Binding failed for port 3d9f5735-20ed-42f4-9565-073437eef966, please check neutron logs for more information. {{(pid=61923) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 612.365515] env[61923]: DEBUG oslo_concurrency.lockutils [None req-b2fe22fb-da8f-41ba-8845-95fc6177390e tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.686s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 612.367598] env[61923]: INFO nova.compute.claims [None req-b2fe22fb-da8f-41ba-8845-95fc6177390e tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] [instance: 119bb4a7-2d71-44d7-ae81-9238512937c8] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 612.372601] env[61923]: DEBUG nova.compute.manager [None req-543d2693-878a-4ade-aaa2-339016c76eca tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] [instance: 045c0b05-333e-4f95-94b0-8f51d87dca0a] Build of instance 045c0b05-333e-4f95-94b0-8f51d87dca0a was re-scheduled: Binding failed for port 3d9f5735-20ed-42f4-9565-073437eef966, please check neutron logs for more information. {{(pid=61923) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 612.372601] env[61923]: DEBUG nova.compute.manager [None req-543d2693-878a-4ade-aaa2-339016c76eca tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] [instance: 045c0b05-333e-4f95-94b0-8f51d87dca0a] Unplugging VIFs for instance {{(pid=61923) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 612.372601] env[61923]: DEBUG oslo_concurrency.lockutils [None req-543d2693-878a-4ade-aaa2-339016c76eca tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Acquiring lock "refresh_cache-045c0b05-333e-4f95-94b0-8f51d87dca0a" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 612.372601] env[61923]: DEBUG oslo_concurrency.lockutils [None req-543d2693-878a-4ade-aaa2-339016c76eca tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Acquired lock "refresh_cache-045c0b05-333e-4f95-94b0-8f51d87dca0a" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 612.372789] env[61923]: DEBUG nova.network.neutron [None req-543d2693-878a-4ade-aaa2-339016c76eca tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] [instance: 045c0b05-333e-4f95-94b0-8f51d87dca0a] Building network info cache for instance {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 612.510773] env[61923]: INFO nova.compute.manager [None req-fba27a13-474d-4fe3-9187-c31143f40aa5 tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] [instance: 93818ffd-eb85-48f0-aebd-9774467b0a8f] Took 1.03 seconds to deallocate network for instance. [ 612.513574] env[61923]: DEBUG nova.compute.manager [None req-95b9b70a-a835-4901-9a6d-c0fa1b2217f3 tempest-MigrationsAdminTest-1311988845 tempest-MigrationsAdminTest-1311988845-project-member] [instance: a4b54a7a-3f3d-49ae-ba43-5a10c262d4e5] Starting instance... {{(pid=61923) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 612.542669] env[61923]: DEBUG nova.network.neutron [None req-cfefea89-2815-4df2-96e2-c7370c791e5f tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] [instance: 9c3e2f94-91da-4255-b4d6-1c01e100ff48] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 612.741199] env[61923]: DEBUG nova.network.neutron [None req-cfefea89-2815-4df2-96e2-c7370c791e5f tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] [instance: 9c3e2f94-91da-4255-b4d6-1c01e100ff48] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 612.898365] env[61923]: DEBUG nova.network.neutron [None req-543d2693-878a-4ade-aaa2-339016c76eca tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] [instance: 045c0b05-333e-4f95-94b0-8f51d87dca0a] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 613.054265] env[61923]: DEBUG oslo_concurrency.lockutils [None req-95b9b70a-a835-4901-9a6d-c0fa1b2217f3 tempest-MigrationsAdminTest-1311988845 tempest-MigrationsAdminTest-1311988845-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 613.055804] env[61923]: DEBUG nova.network.neutron [None req-543d2693-878a-4ade-aaa2-339016c76eca tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] [instance: 045c0b05-333e-4f95-94b0-8f51d87dca0a] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 613.247020] env[61923]: DEBUG oslo_concurrency.lockutils [None req-cfefea89-2815-4df2-96e2-c7370c791e5f tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] Releasing lock "refresh_cache-9c3e2f94-91da-4255-b4d6-1c01e100ff48" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 613.247020] env[61923]: DEBUG nova.compute.manager [None req-cfefea89-2815-4df2-96e2-c7370c791e5f tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] [instance: 9c3e2f94-91da-4255-b4d6-1c01e100ff48] Start destroying the instance on the hypervisor. {{(pid=61923) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 613.247020] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-cfefea89-2815-4df2-96e2-c7370c791e5f tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] [instance: 9c3e2f94-91da-4255-b4d6-1c01e100ff48] Destroying instance {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 613.247020] env[61923]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-0c59ded2-a91d-49f7-a0d3-6c2c3929dcc7 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.257543] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4358808d-4e28-48a7-94a1-1bc8498d43d6 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.282870] env[61923]: WARNING nova.virt.vmwareapi.vmops [None req-cfefea89-2815-4df2-96e2-c7370c791e5f tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] [instance: 9c3e2f94-91da-4255-b4d6-1c01e100ff48] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 9c3e2f94-91da-4255-b4d6-1c01e100ff48 could not be found. [ 613.283124] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-cfefea89-2815-4df2-96e2-c7370c791e5f tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] [instance: 9c3e2f94-91da-4255-b4d6-1c01e100ff48] Instance destroyed {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 613.283453] env[61923]: INFO nova.compute.manager [None req-cfefea89-2815-4df2-96e2-c7370c791e5f tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] [instance: 9c3e2f94-91da-4255-b4d6-1c01e100ff48] Took 0.04 seconds to destroy the instance on the hypervisor. [ 613.283728] env[61923]: DEBUG oslo.service.loopingcall [None req-cfefea89-2815-4df2-96e2-c7370c791e5f tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61923) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 613.283948] env[61923]: DEBUG nova.compute.manager [-] [instance: 9c3e2f94-91da-4255-b4d6-1c01e100ff48] Deallocating network for instance {{(pid=61923) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 613.284058] env[61923]: DEBUG nova.network.neutron [-] [instance: 9c3e2f94-91da-4255-b4d6-1c01e100ff48] deallocate_for_instance() {{(pid=61923) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 613.316144] env[61923]: DEBUG nova.network.neutron [-] [instance: 9c3e2f94-91da-4255-b4d6-1c01e100ff48] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 613.545398] env[61923]: INFO nova.scheduler.client.report [None req-fba27a13-474d-4fe3-9187-c31143f40aa5 tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] Deleted allocations for instance 93818ffd-eb85-48f0-aebd-9774467b0a8f [ 613.558307] env[61923]: DEBUG oslo_concurrency.lockutils [None req-543d2693-878a-4ade-aaa2-339016c76eca tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Releasing lock "refresh_cache-045c0b05-333e-4f95-94b0-8f51d87dca0a" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 613.558820] env[61923]: DEBUG nova.compute.manager [None req-543d2693-878a-4ade-aaa2-339016c76eca tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61923) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 613.558820] env[61923]: DEBUG nova.compute.manager [None req-543d2693-878a-4ade-aaa2-339016c76eca tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] [instance: 045c0b05-333e-4f95-94b0-8f51d87dca0a] Deallocating network for instance {{(pid=61923) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 613.558820] env[61923]: DEBUG nova.network.neutron [None req-543d2693-878a-4ade-aaa2-339016c76eca tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] [instance: 045c0b05-333e-4f95-94b0-8f51d87dca0a] deallocate_for_instance() {{(pid=61923) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 613.649063] env[61923]: DEBUG nova.network.neutron [None req-543d2693-878a-4ade-aaa2-339016c76eca tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] [instance: 045c0b05-333e-4f95-94b0-8f51d87dca0a] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 613.815377] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-607c01eb-6a10-4a6b-b2b7-ab49912a814b {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.821388] env[61923]: DEBUG nova.network.neutron [-] [instance: 9c3e2f94-91da-4255-b4d6-1c01e100ff48] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 613.823410] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d1efe5a-5631-4242-9c32-4070c0c0a8e7 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.858131] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d2f03f4-cdeb-4066-84d6-01bdd67cb3d9 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.865700] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89914ba2-71d4-465f-bc4f-0534d852a282 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.880521] env[61923]: DEBUG nova.compute.provider_tree [None req-b2fe22fb-da8f-41ba-8845-95fc6177390e tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 614.056405] env[61923]: DEBUG oslo_concurrency.lockutils [None req-fba27a13-474d-4fe3-9187-c31143f40aa5 tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] Lock "93818ffd-eb85-48f0-aebd-9774467b0a8f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 75.005s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 614.058237] env[61923]: DEBUG oslo_concurrency.lockutils [None req-ae039c28-826e-4a1e-bb21-b1102865618f tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] Lock "93818ffd-eb85-48f0-aebd-9774467b0a8f" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 7.628s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 614.058237] env[61923]: DEBUG oslo_concurrency.lockutils [None req-ae039c28-826e-4a1e-bb21-b1102865618f tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] Acquiring lock "93818ffd-eb85-48f0-aebd-9774467b0a8f-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 614.058362] env[61923]: DEBUG oslo_concurrency.lockutils [None req-ae039c28-826e-4a1e-bb21-b1102865618f tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] Lock "93818ffd-eb85-48f0-aebd-9774467b0a8f-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 614.058458] env[61923]: DEBUG oslo_concurrency.lockutils [None req-ae039c28-826e-4a1e-bb21-b1102865618f tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] Lock "93818ffd-eb85-48f0-aebd-9774467b0a8f-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 614.060727] env[61923]: INFO nova.compute.manager [None req-ae039c28-826e-4a1e-bb21-b1102865618f tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] [instance: 93818ffd-eb85-48f0-aebd-9774467b0a8f] Terminating instance [ 614.066103] env[61923]: DEBUG oslo_concurrency.lockutils [None req-ae039c28-826e-4a1e-bb21-b1102865618f tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] Acquiring lock "refresh_cache-93818ffd-eb85-48f0-aebd-9774467b0a8f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 614.066103] env[61923]: DEBUG oslo_concurrency.lockutils [None req-ae039c28-826e-4a1e-bb21-b1102865618f tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] Acquired lock "refresh_cache-93818ffd-eb85-48f0-aebd-9774467b0a8f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 614.066103] env[61923]: DEBUG nova.network.neutron [None req-ae039c28-826e-4a1e-bb21-b1102865618f tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] [instance: 93818ffd-eb85-48f0-aebd-9774467b0a8f] Building network info cache for instance {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 614.154184] env[61923]: DEBUG nova.network.neutron [None req-543d2693-878a-4ade-aaa2-339016c76eca tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] [instance: 045c0b05-333e-4f95-94b0-8f51d87dca0a] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 614.328428] env[61923]: INFO nova.compute.manager [-] [instance: 9c3e2f94-91da-4255-b4d6-1c01e100ff48] Took 1.04 seconds to deallocate network for instance. [ 614.383241] env[61923]: DEBUG nova.scheduler.client.report [None req-b2fe22fb-da8f-41ba-8845-95fc6177390e tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 614.561084] env[61923]: DEBUG nova.compute.manager [None req-cb394400-3d4b-4149-a290-413641bb6f18 tempest-InstanceActionsV221TestJSON-1730893270 tempest-InstanceActionsV221TestJSON-1730893270-project-member] [instance: dc8cc22d-aba9-4dba-b456-1b323013d234] Starting instance... {{(pid=61923) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 614.590657] env[61923]: DEBUG nova.network.neutron [None req-ae039c28-826e-4a1e-bb21-b1102865618f tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] [instance: 93818ffd-eb85-48f0-aebd-9774467b0a8f] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 614.658534] env[61923]: INFO nova.compute.manager [None req-543d2693-878a-4ade-aaa2-339016c76eca tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] [instance: 045c0b05-333e-4f95-94b0-8f51d87dca0a] Took 1.10 seconds to deallocate network for instance. [ 614.689837] env[61923]: DEBUG nova.network.neutron [None req-ae039c28-826e-4a1e-bb21-b1102865618f tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] [instance: 93818ffd-eb85-48f0-aebd-9774467b0a8f] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 614.888196] env[61923]: DEBUG oslo_concurrency.lockutils [None req-b2fe22fb-da8f-41ba-8845-95fc6177390e tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.522s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 614.888755] env[61923]: DEBUG nova.compute.manager [None req-b2fe22fb-da8f-41ba-8845-95fc6177390e tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] [instance: 119bb4a7-2d71-44d7-ae81-9238512937c8] Start building networks asynchronously for instance. {{(pid=61923) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 614.891521] env[61923]: DEBUG oslo_concurrency.lockutils [None req-58d8b214-0b32-4993-bdd8-d652de7f9397 tempest-ServerActionsTestOtherA-1581037736 tempest-ServerActionsTestOtherA-1581037736-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 20.122s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 614.960400] env[61923]: DEBUG oslo_concurrency.lockutils [None req-f7b52f4c-b2e3-49f8-b056-b7254f72bb20 tempest-AttachInterfacesUnderV243Test-498638506 tempest-AttachInterfacesUnderV243Test-498638506-project-member] Acquiring lock "0eab2200-2528-4a89-821a-cdd6a3c99297" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 614.960618] env[61923]: DEBUG oslo_concurrency.lockutils [None req-f7b52f4c-b2e3-49f8-b056-b7254f72bb20 tempest-AttachInterfacesUnderV243Test-498638506 tempest-AttachInterfacesUnderV243Test-498638506-project-member] Lock "0eab2200-2528-4a89-821a-cdd6a3c99297" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 615.082424] env[61923]: DEBUG oslo_concurrency.lockutils [None req-cb394400-3d4b-4149-a290-413641bb6f18 tempest-InstanceActionsV221TestJSON-1730893270 tempest-InstanceActionsV221TestJSON-1730893270-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 615.194324] env[61923]: DEBUG oslo_concurrency.lockutils [None req-ae039c28-826e-4a1e-bb21-b1102865618f tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] Releasing lock "refresh_cache-93818ffd-eb85-48f0-aebd-9774467b0a8f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 615.194471] env[61923]: DEBUG nova.compute.manager [None req-ae039c28-826e-4a1e-bb21-b1102865618f tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] [instance: 93818ffd-eb85-48f0-aebd-9774467b0a8f] Start destroying the instance on the hypervisor. {{(pid=61923) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 615.194660] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-ae039c28-826e-4a1e-bb21-b1102865618f tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] [instance: 93818ffd-eb85-48f0-aebd-9774467b0a8f] Destroying instance {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 615.194966] env[61923]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-993d34da-da34-4faa-9f2a-4ade41cd48fa {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.204983] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da994924-32ae-40bf-8279-ffb2c6823b27 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.228965] env[61923]: WARNING nova.virt.vmwareapi.vmops [None req-ae039c28-826e-4a1e-bb21-b1102865618f tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] [instance: 93818ffd-eb85-48f0-aebd-9774467b0a8f] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 93818ffd-eb85-48f0-aebd-9774467b0a8f could not be found. [ 615.229813] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-ae039c28-826e-4a1e-bb21-b1102865618f tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] [instance: 93818ffd-eb85-48f0-aebd-9774467b0a8f] Instance destroyed {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 615.229813] env[61923]: INFO nova.compute.manager [None req-ae039c28-826e-4a1e-bb21-b1102865618f tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] [instance: 93818ffd-eb85-48f0-aebd-9774467b0a8f] Took 0.03 seconds to destroy the instance on the hypervisor. [ 615.229813] env[61923]: DEBUG oslo.service.loopingcall [None req-ae039c28-826e-4a1e-bb21-b1102865618f tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61923) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 615.230096] env[61923]: DEBUG nova.compute.manager [-] [instance: 93818ffd-eb85-48f0-aebd-9774467b0a8f] Deallocating network for instance {{(pid=61923) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 615.230193] env[61923]: DEBUG nova.network.neutron [-] [instance: 93818ffd-eb85-48f0-aebd-9774467b0a8f] deallocate_for_instance() {{(pid=61923) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 615.246784] env[61923]: DEBUG nova.network.neutron [-] [instance: 93818ffd-eb85-48f0-aebd-9774467b0a8f] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 615.356191] env[61923]: DEBUG oslo_concurrency.lockutils [None req-cfefea89-2815-4df2-96e2-c7370c791e5f tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] Lock "9c3e2f94-91da-4255-b4d6-1c01e100ff48" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 3.350s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 615.396662] env[61923]: DEBUG nova.compute.utils [None req-b2fe22fb-da8f-41ba-8845-95fc6177390e tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Using /dev/sd instead of None {{(pid=61923) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 615.401615] env[61923]: DEBUG nova.compute.manager [None req-b2fe22fb-da8f-41ba-8845-95fc6177390e tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] [instance: 119bb4a7-2d71-44d7-ae81-9238512937c8] Allocating IP information in the background. {{(pid=61923) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 615.401615] env[61923]: DEBUG nova.network.neutron [None req-b2fe22fb-da8f-41ba-8845-95fc6177390e tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] [instance: 119bb4a7-2d71-44d7-ae81-9238512937c8] allocate_for_instance() {{(pid=61923) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 615.467238] env[61923]: DEBUG nova.policy [None req-b2fe22fb-da8f-41ba-8845-95fc6177390e tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'eb00f8aff92b44e9a0e739366a1d2662', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7abfe95660904cd4b8824257795d8e8a', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61923) authorize /opt/stack/nova/nova/policy.py:201}} [ 615.692404] env[61923]: INFO nova.scheduler.client.report [None req-543d2693-878a-4ade-aaa2-339016c76eca tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Deleted allocations for instance 045c0b05-333e-4f95-94b0-8f51d87dca0a [ 615.748990] env[61923]: DEBUG nova.network.neutron [-] [instance: 93818ffd-eb85-48f0-aebd-9774467b0a8f] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 615.787203] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b4cc32b-f1b4-4afc-92b1-46efac2369dc {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.402302] env[61923]: DEBUG nova.network.neutron [None req-b2fe22fb-da8f-41ba-8845-95fc6177390e tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] [instance: 119bb4a7-2d71-44d7-ae81-9238512937c8] Successfully created port: 37c91d0a-5c78-46e0-800e-9c9760eb9fbc {{(pid=61923) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 616.404878] env[61923]: DEBUG nova.compute.manager [None req-b2fe22fb-da8f-41ba-8845-95fc6177390e tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] [instance: 119bb4a7-2d71-44d7-ae81-9238512937c8] Start building block device mappings for instance. {{(pid=61923) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 616.408578] env[61923]: DEBUG oslo_concurrency.lockutils [None req-543d2693-878a-4ade-aaa2-339016c76eca tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Lock "045c0b05-333e-4f95-94b0-8f51d87dca0a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 77.304s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 616.408937] env[61923]: INFO nova.compute.manager [-] [instance: 93818ffd-eb85-48f0-aebd-9774467b0a8f] Took 1.18 seconds to deallocate network for instance. [ 616.417404] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28a2080b-c39b-49fd-82ea-49b20ee582f0 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.450925] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5e4beb6-e6ec-48fd-a2bd-14968be3bb15 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.459522] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4810cddc-f241-44a8-a1e4-71920188bdc2 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.473240] env[61923]: DEBUG nova.compute.provider_tree [None req-58d8b214-0b32-4993-bdd8-d652de7f9397 tempest-ServerActionsTestOtherA-1581037736 tempest-ServerActionsTestOtherA-1581037736-project-member] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 616.914480] env[61923]: DEBUG nova.compute.manager [None req-753531fa-39a1-4c82-9915-d195bf0fe1b2 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] [instance: 03100310-86ce-4739-8636-cd2aa1b542ff] Starting instance... {{(pid=61923) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 616.975842] env[61923]: DEBUG nova.scheduler.client.report [None req-58d8b214-0b32-4993-bdd8-d652de7f9397 tempest-ServerActionsTestOtherA-1581037736 tempest-ServerActionsTestOtherA-1581037736-project-member] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 617.416734] env[61923]: DEBUG nova.compute.manager [None req-b2fe22fb-da8f-41ba-8845-95fc6177390e tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] [instance: 119bb4a7-2d71-44d7-ae81-9238512937c8] Start spawning the instance on the hypervisor. {{(pid=61923) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 617.446562] env[61923]: DEBUG oslo_concurrency.lockutils [None req-ae039c28-826e-4a1e-bb21-b1102865618f tempest-ListServersNegativeTestJSON-1751562555 tempest-ListServersNegativeTestJSON-1751562555-project-member] Lock "93818ffd-eb85-48f0-aebd-9774467b0a8f" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 3.388s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 617.455155] env[61923]: DEBUG oslo_concurrency.lockutils [None req-753531fa-39a1-4c82-9915-d195bf0fe1b2 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 617.457920] env[61923]: DEBUG nova.virt.hardware [None req-b2fe22fb-da8f-41ba-8845-95fc6177390e tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-29T20:07:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-29T20:07:35Z,direct_url=,disk_format='vmdk',id=0f153f63-ae0a-45b1-b7f5-7f9b673c947f,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='4e7b5e3b3c3443c8bd5c70f8a15739f5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-29T20:07:36Z,virtual_size=,visibility=), allow threads: False {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 617.458184] env[61923]: DEBUG nova.virt.hardware [None req-b2fe22fb-da8f-41ba-8845-95fc6177390e tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Flavor limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 617.458339] env[61923]: DEBUG nova.virt.hardware [None req-b2fe22fb-da8f-41ba-8845-95fc6177390e tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Image limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 617.458514] env[61923]: DEBUG nova.virt.hardware [None req-b2fe22fb-da8f-41ba-8845-95fc6177390e tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Flavor pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 617.458757] env[61923]: DEBUG nova.virt.hardware [None req-b2fe22fb-da8f-41ba-8845-95fc6177390e tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Image pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 617.458847] env[61923]: DEBUG nova.virt.hardware [None req-b2fe22fb-da8f-41ba-8845-95fc6177390e tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 617.459011] env[61923]: DEBUG nova.virt.hardware [None req-b2fe22fb-da8f-41ba-8845-95fc6177390e tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 617.459178] env[61923]: DEBUG nova.virt.hardware [None req-b2fe22fb-da8f-41ba-8845-95fc6177390e tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 617.459415] env[61923]: DEBUG nova.virt.hardware [None req-b2fe22fb-da8f-41ba-8845-95fc6177390e tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Got 1 possible topologies {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 617.459614] env[61923]: DEBUG nova.virt.hardware [None req-b2fe22fb-da8f-41ba-8845-95fc6177390e tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 617.459806] env[61923]: DEBUG nova.virt.hardware [None req-b2fe22fb-da8f-41ba-8845-95fc6177390e tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 617.460667] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e33210da-9683-431e-9a50-a12603ca4bcc {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 617.471616] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5c13590-722b-4371-b0cb-59cf5b8d312a {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 617.489258] env[61923]: DEBUG oslo_concurrency.lockutils [None req-58d8b214-0b32-4993-bdd8-d652de7f9397 tempest-ServerActionsTestOtherA-1581037736 tempest-ServerActionsTestOtherA-1581037736-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.597s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 617.489963] env[61923]: ERROR nova.compute.manager [None req-58d8b214-0b32-4993-bdd8-d652de7f9397 tempest-ServerActionsTestOtherA-1581037736 tempest-ServerActionsTestOtherA-1581037736-project-member] [instance: af5feda7-204a-47f1-b7fa-afe44832b9a2] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 1b61be5f-12b3-4ed1-a87e-41558aedeada, please check neutron logs for more information. [ 617.489963] env[61923]: ERROR nova.compute.manager [instance: af5feda7-204a-47f1-b7fa-afe44832b9a2] Traceback (most recent call last): [ 617.489963] env[61923]: ERROR nova.compute.manager [instance: af5feda7-204a-47f1-b7fa-afe44832b9a2] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 617.489963] env[61923]: ERROR nova.compute.manager [instance: af5feda7-204a-47f1-b7fa-afe44832b9a2] self.driver.spawn(context, instance, image_meta, [ 617.489963] env[61923]: ERROR nova.compute.manager [instance: af5feda7-204a-47f1-b7fa-afe44832b9a2] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 617.489963] env[61923]: ERROR nova.compute.manager [instance: af5feda7-204a-47f1-b7fa-afe44832b9a2] self._vmops.spawn(context, instance, image_meta, injected_files, [ 617.489963] env[61923]: ERROR nova.compute.manager [instance: af5feda7-204a-47f1-b7fa-afe44832b9a2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 617.489963] env[61923]: ERROR nova.compute.manager [instance: af5feda7-204a-47f1-b7fa-afe44832b9a2] vm_ref = self.build_virtual_machine(instance, [ 617.489963] env[61923]: ERROR nova.compute.manager [instance: af5feda7-204a-47f1-b7fa-afe44832b9a2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 617.489963] env[61923]: ERROR nova.compute.manager [instance: af5feda7-204a-47f1-b7fa-afe44832b9a2] vif_infos = vmwarevif.get_vif_info(self._session, [ 617.489963] env[61923]: ERROR nova.compute.manager [instance: af5feda7-204a-47f1-b7fa-afe44832b9a2] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 617.490296] env[61923]: ERROR nova.compute.manager [instance: af5feda7-204a-47f1-b7fa-afe44832b9a2] for vif in network_info: [ 617.490296] env[61923]: ERROR nova.compute.manager [instance: af5feda7-204a-47f1-b7fa-afe44832b9a2] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 617.490296] env[61923]: ERROR nova.compute.manager [instance: af5feda7-204a-47f1-b7fa-afe44832b9a2] return self._sync_wrapper(fn, *args, **kwargs) [ 617.490296] env[61923]: ERROR nova.compute.manager [instance: af5feda7-204a-47f1-b7fa-afe44832b9a2] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 617.490296] env[61923]: ERROR nova.compute.manager [instance: af5feda7-204a-47f1-b7fa-afe44832b9a2] self.wait() [ 617.490296] env[61923]: ERROR nova.compute.manager [instance: af5feda7-204a-47f1-b7fa-afe44832b9a2] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 617.490296] env[61923]: ERROR nova.compute.manager [instance: af5feda7-204a-47f1-b7fa-afe44832b9a2] self[:] = self._gt.wait() [ 617.490296] env[61923]: ERROR nova.compute.manager [instance: af5feda7-204a-47f1-b7fa-afe44832b9a2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 617.490296] env[61923]: ERROR nova.compute.manager [instance: af5feda7-204a-47f1-b7fa-afe44832b9a2] return self._exit_event.wait() [ 617.490296] env[61923]: ERROR nova.compute.manager [instance: af5feda7-204a-47f1-b7fa-afe44832b9a2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 617.490296] env[61923]: ERROR nova.compute.manager [instance: af5feda7-204a-47f1-b7fa-afe44832b9a2] result = hub.switch() [ 617.490296] env[61923]: ERROR nova.compute.manager [instance: af5feda7-204a-47f1-b7fa-afe44832b9a2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 617.490296] env[61923]: ERROR nova.compute.manager [instance: af5feda7-204a-47f1-b7fa-afe44832b9a2] return self.greenlet.switch() [ 617.490585] env[61923]: ERROR nova.compute.manager [instance: af5feda7-204a-47f1-b7fa-afe44832b9a2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 617.490585] env[61923]: ERROR nova.compute.manager [instance: af5feda7-204a-47f1-b7fa-afe44832b9a2] result = function(*args, **kwargs) [ 617.490585] env[61923]: ERROR nova.compute.manager [instance: af5feda7-204a-47f1-b7fa-afe44832b9a2] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 617.490585] env[61923]: ERROR nova.compute.manager [instance: af5feda7-204a-47f1-b7fa-afe44832b9a2] return func(*args, **kwargs) [ 617.490585] env[61923]: ERROR nova.compute.manager [instance: af5feda7-204a-47f1-b7fa-afe44832b9a2] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 617.490585] env[61923]: ERROR nova.compute.manager [instance: af5feda7-204a-47f1-b7fa-afe44832b9a2] raise e [ 617.490585] env[61923]: ERROR nova.compute.manager [instance: af5feda7-204a-47f1-b7fa-afe44832b9a2] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 617.490585] env[61923]: ERROR nova.compute.manager [instance: af5feda7-204a-47f1-b7fa-afe44832b9a2] nwinfo = self.network_api.allocate_for_instance( [ 617.490585] env[61923]: ERROR nova.compute.manager [instance: af5feda7-204a-47f1-b7fa-afe44832b9a2] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 617.490585] env[61923]: ERROR nova.compute.manager [instance: af5feda7-204a-47f1-b7fa-afe44832b9a2] created_port_ids = self._update_ports_for_instance( [ 617.490585] env[61923]: ERROR nova.compute.manager [instance: af5feda7-204a-47f1-b7fa-afe44832b9a2] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 617.490585] env[61923]: ERROR nova.compute.manager [instance: af5feda7-204a-47f1-b7fa-afe44832b9a2] with excutils.save_and_reraise_exception(): [ 617.490585] env[61923]: ERROR nova.compute.manager [instance: af5feda7-204a-47f1-b7fa-afe44832b9a2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 617.491044] env[61923]: ERROR nova.compute.manager [instance: af5feda7-204a-47f1-b7fa-afe44832b9a2] self.force_reraise() [ 617.491044] env[61923]: ERROR nova.compute.manager [instance: af5feda7-204a-47f1-b7fa-afe44832b9a2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 617.491044] env[61923]: ERROR nova.compute.manager [instance: af5feda7-204a-47f1-b7fa-afe44832b9a2] raise self.value [ 617.491044] env[61923]: ERROR nova.compute.manager [instance: af5feda7-204a-47f1-b7fa-afe44832b9a2] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 617.491044] env[61923]: ERROR nova.compute.manager [instance: af5feda7-204a-47f1-b7fa-afe44832b9a2] updated_port = self._update_port( [ 617.491044] env[61923]: ERROR nova.compute.manager [instance: af5feda7-204a-47f1-b7fa-afe44832b9a2] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 617.491044] env[61923]: ERROR nova.compute.manager [instance: af5feda7-204a-47f1-b7fa-afe44832b9a2] _ensure_no_port_binding_failure(port) [ 617.491044] env[61923]: ERROR nova.compute.manager [instance: af5feda7-204a-47f1-b7fa-afe44832b9a2] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 617.491044] env[61923]: ERROR nova.compute.manager [instance: af5feda7-204a-47f1-b7fa-afe44832b9a2] raise exception.PortBindingFailed(port_id=port['id']) [ 617.491044] env[61923]: ERROR nova.compute.manager [instance: af5feda7-204a-47f1-b7fa-afe44832b9a2] nova.exception.PortBindingFailed: Binding failed for port 1b61be5f-12b3-4ed1-a87e-41558aedeada, please check neutron logs for more information. [ 617.491044] env[61923]: ERROR nova.compute.manager [instance: af5feda7-204a-47f1-b7fa-afe44832b9a2] [ 617.491306] env[61923]: DEBUG nova.compute.utils [None req-58d8b214-0b32-4993-bdd8-d652de7f9397 tempest-ServerActionsTestOtherA-1581037736 tempest-ServerActionsTestOtherA-1581037736-project-member] [instance: af5feda7-204a-47f1-b7fa-afe44832b9a2] Binding failed for port 1b61be5f-12b3-4ed1-a87e-41558aedeada, please check neutron logs for more information. {{(pid=61923) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 617.492713] env[61923]: DEBUG nova.compute.manager [None req-58d8b214-0b32-4993-bdd8-d652de7f9397 tempest-ServerActionsTestOtherA-1581037736 tempest-ServerActionsTestOtherA-1581037736-project-member] [instance: af5feda7-204a-47f1-b7fa-afe44832b9a2] Build of instance af5feda7-204a-47f1-b7fa-afe44832b9a2 was re-scheduled: Binding failed for port 1b61be5f-12b3-4ed1-a87e-41558aedeada, please check neutron logs for more information. {{(pid=61923) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 617.493242] env[61923]: DEBUG nova.compute.manager [None req-58d8b214-0b32-4993-bdd8-d652de7f9397 tempest-ServerActionsTestOtherA-1581037736 tempest-ServerActionsTestOtherA-1581037736-project-member] [instance: af5feda7-204a-47f1-b7fa-afe44832b9a2] Unplugging VIFs for instance {{(pid=61923) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 617.493242] env[61923]: DEBUG oslo_concurrency.lockutils [None req-58d8b214-0b32-4993-bdd8-d652de7f9397 tempest-ServerActionsTestOtherA-1581037736 tempest-ServerActionsTestOtherA-1581037736-project-member] Acquiring lock "refresh_cache-af5feda7-204a-47f1-b7fa-afe44832b9a2" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 617.493242] env[61923]: DEBUG oslo_concurrency.lockutils [None req-58d8b214-0b32-4993-bdd8-d652de7f9397 tempest-ServerActionsTestOtherA-1581037736 tempest-ServerActionsTestOtherA-1581037736-project-member] Acquired lock "refresh_cache-af5feda7-204a-47f1-b7fa-afe44832b9a2" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 617.493941] env[61923]: DEBUG nova.network.neutron [None req-58d8b214-0b32-4993-bdd8-d652de7f9397 tempest-ServerActionsTestOtherA-1581037736 tempest-ServerActionsTestOtherA-1581037736-project-member] [instance: af5feda7-204a-47f1-b7fa-afe44832b9a2] Building network info cache for instance {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 617.494359] env[61923]: DEBUG oslo_concurrency.lockutils [None req-595ebb22-536b-4375-9cbe-6fc472a5a17d tempest-ListServerFiltersTestJSON-483030130 tempest-ListServerFiltersTestJSON-483030130-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 18.251s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 617.624827] env[61923]: DEBUG nova.compute.manager [req-656a14ee-4ad8-4146-a4a3-743750683b24 req-9fc26fb8-05d2-4ac6-978a-3edad9a9c4ee service nova] [instance: 119bb4a7-2d71-44d7-ae81-9238512937c8] Received event network-changed-37c91d0a-5c78-46e0-800e-9c9760eb9fbc {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 617.624827] env[61923]: DEBUG nova.compute.manager [req-656a14ee-4ad8-4146-a4a3-743750683b24 req-9fc26fb8-05d2-4ac6-978a-3edad9a9c4ee service nova] [instance: 119bb4a7-2d71-44d7-ae81-9238512937c8] Refreshing instance network info cache due to event network-changed-37c91d0a-5c78-46e0-800e-9c9760eb9fbc. {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 617.625034] env[61923]: DEBUG oslo_concurrency.lockutils [req-656a14ee-4ad8-4146-a4a3-743750683b24 req-9fc26fb8-05d2-4ac6-978a-3edad9a9c4ee service nova] Acquiring lock "refresh_cache-119bb4a7-2d71-44d7-ae81-9238512937c8" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 617.625272] env[61923]: DEBUG oslo_concurrency.lockutils [req-656a14ee-4ad8-4146-a4a3-743750683b24 req-9fc26fb8-05d2-4ac6-978a-3edad9a9c4ee service nova] Acquired lock "refresh_cache-119bb4a7-2d71-44d7-ae81-9238512937c8" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 617.625444] env[61923]: DEBUG nova.network.neutron [req-656a14ee-4ad8-4146-a4a3-743750683b24 req-9fc26fb8-05d2-4ac6-978a-3edad9a9c4ee service nova] [instance: 119bb4a7-2d71-44d7-ae81-9238512937c8] Refreshing network info cache for port 37c91d0a-5c78-46e0-800e-9c9760eb9fbc {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 617.933527] env[61923]: ERROR nova.compute.manager [None req-b2fe22fb-da8f-41ba-8845-95fc6177390e tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 37c91d0a-5c78-46e0-800e-9c9760eb9fbc, please check neutron logs for more information. [ 617.933527] env[61923]: ERROR nova.compute.manager Traceback (most recent call last): [ 617.933527] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 617.933527] env[61923]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 617.933527] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 617.933527] env[61923]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 617.933527] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 617.933527] env[61923]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 617.933527] env[61923]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 617.933527] env[61923]: ERROR nova.compute.manager self.force_reraise() [ 617.933527] env[61923]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 617.933527] env[61923]: ERROR nova.compute.manager raise self.value [ 617.933527] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 617.933527] env[61923]: ERROR nova.compute.manager updated_port = self._update_port( [ 617.933527] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 617.933527] env[61923]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 617.934193] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 617.934193] env[61923]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 617.934193] env[61923]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 37c91d0a-5c78-46e0-800e-9c9760eb9fbc, please check neutron logs for more information. [ 617.934193] env[61923]: ERROR nova.compute.manager [ 617.934193] env[61923]: Traceback (most recent call last): [ 617.934193] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 617.934193] env[61923]: listener.cb(fileno) [ 617.934193] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 617.934193] env[61923]: result = function(*args, **kwargs) [ 617.934193] env[61923]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 617.934193] env[61923]: return func(*args, **kwargs) [ 617.934193] env[61923]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 617.934193] env[61923]: raise e [ 617.934193] env[61923]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 617.934193] env[61923]: nwinfo = self.network_api.allocate_for_instance( [ 617.934193] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 617.934193] env[61923]: created_port_ids = self._update_ports_for_instance( [ 617.934193] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 617.934193] env[61923]: with excutils.save_and_reraise_exception(): [ 617.934193] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 617.934193] env[61923]: self.force_reraise() [ 617.934193] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 617.934193] env[61923]: raise self.value [ 617.934193] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 617.934193] env[61923]: updated_port = self._update_port( [ 617.934193] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 617.934193] env[61923]: _ensure_no_port_binding_failure(port) [ 617.934193] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 617.934193] env[61923]: raise exception.PortBindingFailed(port_id=port['id']) [ 617.935370] env[61923]: nova.exception.PortBindingFailed: Binding failed for port 37c91d0a-5c78-46e0-800e-9c9760eb9fbc, please check neutron logs for more information. [ 617.935370] env[61923]: Removing descriptor: 14 [ 617.935370] env[61923]: ERROR nova.compute.manager [None req-b2fe22fb-da8f-41ba-8845-95fc6177390e tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] [instance: 119bb4a7-2d71-44d7-ae81-9238512937c8] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 37c91d0a-5c78-46e0-800e-9c9760eb9fbc, please check neutron logs for more information. [ 617.935370] env[61923]: ERROR nova.compute.manager [instance: 119bb4a7-2d71-44d7-ae81-9238512937c8] Traceback (most recent call last): [ 617.935370] env[61923]: ERROR nova.compute.manager [instance: 119bb4a7-2d71-44d7-ae81-9238512937c8] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 617.935370] env[61923]: ERROR nova.compute.manager [instance: 119bb4a7-2d71-44d7-ae81-9238512937c8] yield resources [ 617.935370] env[61923]: ERROR nova.compute.manager [instance: 119bb4a7-2d71-44d7-ae81-9238512937c8] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 617.935370] env[61923]: ERROR nova.compute.manager [instance: 119bb4a7-2d71-44d7-ae81-9238512937c8] self.driver.spawn(context, instance, image_meta, [ 617.935370] env[61923]: ERROR nova.compute.manager [instance: 119bb4a7-2d71-44d7-ae81-9238512937c8] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 617.935370] env[61923]: ERROR nova.compute.manager [instance: 119bb4a7-2d71-44d7-ae81-9238512937c8] self._vmops.spawn(context, instance, image_meta, injected_files, [ 617.935370] env[61923]: ERROR nova.compute.manager [instance: 119bb4a7-2d71-44d7-ae81-9238512937c8] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 617.935370] env[61923]: ERROR nova.compute.manager [instance: 119bb4a7-2d71-44d7-ae81-9238512937c8] vm_ref = self.build_virtual_machine(instance, [ 617.935875] env[61923]: ERROR nova.compute.manager [instance: 119bb4a7-2d71-44d7-ae81-9238512937c8] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 617.935875] env[61923]: ERROR nova.compute.manager [instance: 119bb4a7-2d71-44d7-ae81-9238512937c8] vif_infos = vmwarevif.get_vif_info(self._session, [ 617.935875] env[61923]: ERROR nova.compute.manager [instance: 119bb4a7-2d71-44d7-ae81-9238512937c8] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 617.935875] env[61923]: ERROR nova.compute.manager [instance: 119bb4a7-2d71-44d7-ae81-9238512937c8] for vif in network_info: [ 617.935875] env[61923]: ERROR nova.compute.manager [instance: 119bb4a7-2d71-44d7-ae81-9238512937c8] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 617.935875] env[61923]: ERROR nova.compute.manager [instance: 119bb4a7-2d71-44d7-ae81-9238512937c8] return self._sync_wrapper(fn, *args, **kwargs) [ 617.935875] env[61923]: ERROR nova.compute.manager [instance: 119bb4a7-2d71-44d7-ae81-9238512937c8] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 617.935875] env[61923]: ERROR nova.compute.manager [instance: 119bb4a7-2d71-44d7-ae81-9238512937c8] self.wait() [ 617.935875] env[61923]: ERROR nova.compute.manager [instance: 119bb4a7-2d71-44d7-ae81-9238512937c8] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 617.935875] env[61923]: ERROR nova.compute.manager [instance: 119bb4a7-2d71-44d7-ae81-9238512937c8] self[:] = self._gt.wait() [ 617.935875] env[61923]: ERROR nova.compute.manager [instance: 119bb4a7-2d71-44d7-ae81-9238512937c8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 617.935875] env[61923]: ERROR nova.compute.manager [instance: 119bb4a7-2d71-44d7-ae81-9238512937c8] return self._exit_event.wait() [ 617.935875] env[61923]: ERROR nova.compute.manager [instance: 119bb4a7-2d71-44d7-ae81-9238512937c8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 617.936890] env[61923]: ERROR nova.compute.manager [instance: 119bb4a7-2d71-44d7-ae81-9238512937c8] result = hub.switch() [ 617.936890] env[61923]: ERROR nova.compute.manager [instance: 119bb4a7-2d71-44d7-ae81-9238512937c8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 617.936890] env[61923]: ERROR nova.compute.manager [instance: 119bb4a7-2d71-44d7-ae81-9238512937c8] return self.greenlet.switch() [ 617.936890] env[61923]: ERROR nova.compute.manager [instance: 119bb4a7-2d71-44d7-ae81-9238512937c8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 617.936890] env[61923]: ERROR nova.compute.manager [instance: 119bb4a7-2d71-44d7-ae81-9238512937c8] result = function(*args, **kwargs) [ 617.936890] env[61923]: ERROR nova.compute.manager [instance: 119bb4a7-2d71-44d7-ae81-9238512937c8] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 617.936890] env[61923]: ERROR nova.compute.manager [instance: 119bb4a7-2d71-44d7-ae81-9238512937c8] return func(*args, **kwargs) [ 617.936890] env[61923]: ERROR nova.compute.manager [instance: 119bb4a7-2d71-44d7-ae81-9238512937c8] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 617.936890] env[61923]: ERROR nova.compute.manager [instance: 119bb4a7-2d71-44d7-ae81-9238512937c8] raise e [ 617.936890] env[61923]: ERROR nova.compute.manager [instance: 119bb4a7-2d71-44d7-ae81-9238512937c8] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 617.936890] env[61923]: ERROR nova.compute.manager [instance: 119bb4a7-2d71-44d7-ae81-9238512937c8] nwinfo = self.network_api.allocate_for_instance( [ 617.936890] env[61923]: ERROR nova.compute.manager [instance: 119bb4a7-2d71-44d7-ae81-9238512937c8] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 617.936890] env[61923]: ERROR nova.compute.manager [instance: 119bb4a7-2d71-44d7-ae81-9238512937c8] created_port_ids = self._update_ports_for_instance( [ 617.937372] env[61923]: ERROR nova.compute.manager [instance: 119bb4a7-2d71-44d7-ae81-9238512937c8] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 617.937372] env[61923]: ERROR nova.compute.manager [instance: 119bb4a7-2d71-44d7-ae81-9238512937c8] with excutils.save_and_reraise_exception(): [ 617.937372] env[61923]: ERROR nova.compute.manager [instance: 119bb4a7-2d71-44d7-ae81-9238512937c8] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 617.937372] env[61923]: ERROR nova.compute.manager [instance: 119bb4a7-2d71-44d7-ae81-9238512937c8] self.force_reraise() [ 617.937372] env[61923]: ERROR nova.compute.manager [instance: 119bb4a7-2d71-44d7-ae81-9238512937c8] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 617.937372] env[61923]: ERROR nova.compute.manager [instance: 119bb4a7-2d71-44d7-ae81-9238512937c8] raise self.value [ 617.937372] env[61923]: ERROR nova.compute.manager [instance: 119bb4a7-2d71-44d7-ae81-9238512937c8] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 617.937372] env[61923]: ERROR nova.compute.manager [instance: 119bb4a7-2d71-44d7-ae81-9238512937c8] updated_port = self._update_port( [ 617.937372] env[61923]: ERROR nova.compute.manager [instance: 119bb4a7-2d71-44d7-ae81-9238512937c8] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 617.937372] env[61923]: ERROR nova.compute.manager [instance: 119bb4a7-2d71-44d7-ae81-9238512937c8] _ensure_no_port_binding_failure(port) [ 617.937372] env[61923]: ERROR nova.compute.manager [instance: 119bb4a7-2d71-44d7-ae81-9238512937c8] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 617.937372] env[61923]: ERROR nova.compute.manager [instance: 119bb4a7-2d71-44d7-ae81-9238512937c8] raise exception.PortBindingFailed(port_id=port['id']) [ 617.937675] env[61923]: ERROR nova.compute.manager [instance: 119bb4a7-2d71-44d7-ae81-9238512937c8] nova.exception.PortBindingFailed: Binding failed for port 37c91d0a-5c78-46e0-800e-9c9760eb9fbc, please check neutron logs for more information. [ 617.937675] env[61923]: ERROR nova.compute.manager [instance: 119bb4a7-2d71-44d7-ae81-9238512937c8] [ 617.937675] env[61923]: INFO nova.compute.manager [None req-b2fe22fb-da8f-41ba-8845-95fc6177390e tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] [instance: 119bb4a7-2d71-44d7-ae81-9238512937c8] Terminating instance [ 617.937675] env[61923]: DEBUG oslo_concurrency.lockutils [None req-b2fe22fb-da8f-41ba-8845-95fc6177390e tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Acquiring lock "refresh_cache-119bb4a7-2d71-44d7-ae81-9238512937c8" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 618.022659] env[61923]: DEBUG nova.network.neutron [None req-58d8b214-0b32-4993-bdd8-d652de7f9397 tempest-ServerActionsTestOtherA-1581037736 tempest-ServerActionsTestOtherA-1581037736-project-member] [instance: af5feda7-204a-47f1-b7fa-afe44832b9a2] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 618.158614] env[61923]: DEBUG nova.network.neutron [None req-58d8b214-0b32-4993-bdd8-d652de7f9397 tempest-ServerActionsTestOtherA-1581037736 tempest-ServerActionsTestOtherA-1581037736-project-member] [instance: af5feda7-204a-47f1-b7fa-afe44832b9a2] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 618.167242] env[61923]: DEBUG nova.network.neutron [req-656a14ee-4ad8-4146-a4a3-743750683b24 req-9fc26fb8-05d2-4ac6-978a-3edad9a9c4ee service nova] [instance: 119bb4a7-2d71-44d7-ae81-9238512937c8] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 618.276335] env[61923]: DEBUG nova.network.neutron [req-656a14ee-4ad8-4146-a4a3-743750683b24 req-9fc26fb8-05d2-4ac6-978a-3edad9a9c4ee service nova] [instance: 119bb4a7-2d71-44d7-ae81-9238512937c8] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 618.438075] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8aaa6626-87da-4cc7-b17e-2c32a91c0ac9 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.445811] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a079bdc4-b63f-40c0-9d59-d596880f29a4 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.478131] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-152addcd-8d10-4063-ab76-a0394af2a7de {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.485681] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63cae65b-47f6-425f-b00d-64f268e6430a {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.499369] env[61923]: DEBUG nova.compute.provider_tree [None req-595ebb22-536b-4375-9cbe-6fc472a5a17d tempest-ListServerFiltersTestJSON-483030130 tempest-ListServerFiltersTestJSON-483030130-project-member] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 618.661689] env[61923]: DEBUG oslo_concurrency.lockutils [None req-58d8b214-0b32-4993-bdd8-d652de7f9397 tempest-ServerActionsTestOtherA-1581037736 tempest-ServerActionsTestOtherA-1581037736-project-member] Releasing lock "refresh_cache-af5feda7-204a-47f1-b7fa-afe44832b9a2" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 618.661940] env[61923]: DEBUG nova.compute.manager [None req-58d8b214-0b32-4993-bdd8-d652de7f9397 tempest-ServerActionsTestOtherA-1581037736 tempest-ServerActionsTestOtherA-1581037736-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61923) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 618.662134] env[61923]: DEBUG nova.compute.manager [None req-58d8b214-0b32-4993-bdd8-d652de7f9397 tempest-ServerActionsTestOtherA-1581037736 tempest-ServerActionsTestOtherA-1581037736-project-member] [instance: af5feda7-204a-47f1-b7fa-afe44832b9a2] Deallocating network for instance {{(pid=61923) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 618.662304] env[61923]: DEBUG nova.network.neutron [None req-58d8b214-0b32-4993-bdd8-d652de7f9397 tempest-ServerActionsTestOtherA-1581037736 tempest-ServerActionsTestOtherA-1581037736-project-member] [instance: af5feda7-204a-47f1-b7fa-afe44832b9a2] deallocate_for_instance() {{(pid=61923) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 618.685030] env[61923]: DEBUG nova.network.neutron [None req-58d8b214-0b32-4993-bdd8-d652de7f9397 tempest-ServerActionsTestOtherA-1581037736 tempest-ServerActionsTestOtherA-1581037736-project-member] [instance: af5feda7-204a-47f1-b7fa-afe44832b9a2] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 618.779025] env[61923]: DEBUG oslo_concurrency.lockutils [req-656a14ee-4ad8-4146-a4a3-743750683b24 req-9fc26fb8-05d2-4ac6-978a-3edad9a9c4ee service nova] Releasing lock "refresh_cache-119bb4a7-2d71-44d7-ae81-9238512937c8" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 618.779422] env[61923]: DEBUG oslo_concurrency.lockutils [None req-b2fe22fb-da8f-41ba-8845-95fc6177390e tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Acquired lock "refresh_cache-119bb4a7-2d71-44d7-ae81-9238512937c8" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 618.779611] env[61923]: DEBUG nova.network.neutron [None req-b2fe22fb-da8f-41ba-8845-95fc6177390e tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] [instance: 119bb4a7-2d71-44d7-ae81-9238512937c8] Building network info cache for instance {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 619.003979] env[61923]: DEBUG nova.scheduler.client.report [None req-595ebb22-536b-4375-9cbe-6fc472a5a17d tempest-ListServerFiltersTestJSON-483030130 tempest-ListServerFiltersTestJSON-483030130-project-member] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 619.188458] env[61923]: DEBUG nova.network.neutron [None req-58d8b214-0b32-4993-bdd8-d652de7f9397 tempest-ServerActionsTestOtherA-1581037736 tempest-ServerActionsTestOtherA-1581037736-project-member] [instance: af5feda7-204a-47f1-b7fa-afe44832b9a2] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 619.314044] env[61923]: DEBUG nova.network.neutron [None req-b2fe22fb-da8f-41ba-8845-95fc6177390e tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] [instance: 119bb4a7-2d71-44d7-ae81-9238512937c8] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 619.435344] env[61923]: DEBUG nova.network.neutron [None req-b2fe22fb-da8f-41ba-8845-95fc6177390e tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] [instance: 119bb4a7-2d71-44d7-ae81-9238512937c8] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 619.508290] env[61923]: DEBUG oslo_concurrency.lockutils [None req-595ebb22-536b-4375-9cbe-6fc472a5a17d tempest-ListServerFiltersTestJSON-483030130 tempest-ListServerFiltersTestJSON-483030130-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.014s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 619.508934] env[61923]: ERROR nova.compute.manager [None req-595ebb22-536b-4375-9cbe-6fc472a5a17d tempest-ListServerFiltersTestJSON-483030130 tempest-ListServerFiltersTestJSON-483030130-project-member] [instance: 71d9eccc-3b90-4b0a-a59a-d6a8e08a9417] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port b42828d3-f036-4a41-949a-3505532203d9, please check neutron logs for more information. [ 619.508934] env[61923]: ERROR nova.compute.manager [instance: 71d9eccc-3b90-4b0a-a59a-d6a8e08a9417] Traceback (most recent call last): [ 619.508934] env[61923]: ERROR nova.compute.manager [instance: 71d9eccc-3b90-4b0a-a59a-d6a8e08a9417] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 619.508934] env[61923]: ERROR nova.compute.manager [instance: 71d9eccc-3b90-4b0a-a59a-d6a8e08a9417] self.driver.spawn(context, instance, image_meta, [ 619.508934] env[61923]: ERROR nova.compute.manager [instance: 71d9eccc-3b90-4b0a-a59a-d6a8e08a9417] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 619.508934] env[61923]: ERROR nova.compute.manager [instance: 71d9eccc-3b90-4b0a-a59a-d6a8e08a9417] self._vmops.spawn(context, instance, image_meta, injected_files, [ 619.508934] env[61923]: ERROR nova.compute.manager [instance: 71d9eccc-3b90-4b0a-a59a-d6a8e08a9417] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 619.508934] env[61923]: ERROR nova.compute.manager [instance: 71d9eccc-3b90-4b0a-a59a-d6a8e08a9417] vm_ref = self.build_virtual_machine(instance, [ 619.508934] env[61923]: ERROR nova.compute.manager [instance: 71d9eccc-3b90-4b0a-a59a-d6a8e08a9417] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 619.508934] env[61923]: ERROR nova.compute.manager [instance: 71d9eccc-3b90-4b0a-a59a-d6a8e08a9417] vif_infos = vmwarevif.get_vif_info(self._session, [ 619.508934] env[61923]: ERROR nova.compute.manager [instance: 71d9eccc-3b90-4b0a-a59a-d6a8e08a9417] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 619.509288] env[61923]: ERROR nova.compute.manager [instance: 71d9eccc-3b90-4b0a-a59a-d6a8e08a9417] for vif in network_info: [ 619.509288] env[61923]: ERROR nova.compute.manager [instance: 71d9eccc-3b90-4b0a-a59a-d6a8e08a9417] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 619.509288] env[61923]: ERROR nova.compute.manager [instance: 71d9eccc-3b90-4b0a-a59a-d6a8e08a9417] return self._sync_wrapper(fn, *args, **kwargs) [ 619.509288] env[61923]: ERROR nova.compute.manager [instance: 71d9eccc-3b90-4b0a-a59a-d6a8e08a9417] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 619.509288] env[61923]: ERROR nova.compute.manager [instance: 71d9eccc-3b90-4b0a-a59a-d6a8e08a9417] self.wait() [ 619.509288] env[61923]: ERROR nova.compute.manager [instance: 71d9eccc-3b90-4b0a-a59a-d6a8e08a9417] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 619.509288] env[61923]: ERROR nova.compute.manager [instance: 71d9eccc-3b90-4b0a-a59a-d6a8e08a9417] self[:] = self._gt.wait() [ 619.509288] env[61923]: ERROR nova.compute.manager [instance: 71d9eccc-3b90-4b0a-a59a-d6a8e08a9417] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 619.509288] env[61923]: ERROR nova.compute.manager [instance: 71d9eccc-3b90-4b0a-a59a-d6a8e08a9417] return self._exit_event.wait() [ 619.509288] env[61923]: ERROR nova.compute.manager [instance: 71d9eccc-3b90-4b0a-a59a-d6a8e08a9417] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 619.509288] env[61923]: ERROR nova.compute.manager [instance: 71d9eccc-3b90-4b0a-a59a-d6a8e08a9417] result = hub.switch() [ 619.509288] env[61923]: ERROR nova.compute.manager [instance: 71d9eccc-3b90-4b0a-a59a-d6a8e08a9417] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 619.509288] env[61923]: ERROR nova.compute.manager [instance: 71d9eccc-3b90-4b0a-a59a-d6a8e08a9417] return self.greenlet.switch() [ 619.509635] env[61923]: ERROR nova.compute.manager [instance: 71d9eccc-3b90-4b0a-a59a-d6a8e08a9417] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 619.509635] env[61923]: ERROR nova.compute.manager [instance: 71d9eccc-3b90-4b0a-a59a-d6a8e08a9417] result = function(*args, **kwargs) [ 619.509635] env[61923]: ERROR nova.compute.manager [instance: 71d9eccc-3b90-4b0a-a59a-d6a8e08a9417] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 619.509635] env[61923]: ERROR nova.compute.manager [instance: 71d9eccc-3b90-4b0a-a59a-d6a8e08a9417] return func(*args, **kwargs) [ 619.509635] env[61923]: ERROR nova.compute.manager [instance: 71d9eccc-3b90-4b0a-a59a-d6a8e08a9417] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 619.509635] env[61923]: ERROR nova.compute.manager [instance: 71d9eccc-3b90-4b0a-a59a-d6a8e08a9417] raise e [ 619.509635] env[61923]: ERROR nova.compute.manager [instance: 71d9eccc-3b90-4b0a-a59a-d6a8e08a9417] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 619.509635] env[61923]: ERROR nova.compute.manager [instance: 71d9eccc-3b90-4b0a-a59a-d6a8e08a9417] nwinfo = self.network_api.allocate_for_instance( [ 619.509635] env[61923]: ERROR nova.compute.manager [instance: 71d9eccc-3b90-4b0a-a59a-d6a8e08a9417] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 619.509635] env[61923]: ERROR nova.compute.manager [instance: 71d9eccc-3b90-4b0a-a59a-d6a8e08a9417] created_port_ids = self._update_ports_for_instance( [ 619.509635] env[61923]: ERROR nova.compute.manager [instance: 71d9eccc-3b90-4b0a-a59a-d6a8e08a9417] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 619.509635] env[61923]: ERROR nova.compute.manager [instance: 71d9eccc-3b90-4b0a-a59a-d6a8e08a9417] with excutils.save_and_reraise_exception(): [ 619.509635] env[61923]: ERROR nova.compute.manager [instance: 71d9eccc-3b90-4b0a-a59a-d6a8e08a9417] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 619.509957] env[61923]: ERROR nova.compute.manager [instance: 71d9eccc-3b90-4b0a-a59a-d6a8e08a9417] self.force_reraise() [ 619.509957] env[61923]: ERROR nova.compute.manager [instance: 71d9eccc-3b90-4b0a-a59a-d6a8e08a9417] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 619.509957] env[61923]: ERROR nova.compute.manager [instance: 71d9eccc-3b90-4b0a-a59a-d6a8e08a9417] raise self.value [ 619.509957] env[61923]: ERROR nova.compute.manager [instance: 71d9eccc-3b90-4b0a-a59a-d6a8e08a9417] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 619.509957] env[61923]: ERROR nova.compute.manager [instance: 71d9eccc-3b90-4b0a-a59a-d6a8e08a9417] updated_port = self._update_port( [ 619.509957] env[61923]: ERROR nova.compute.manager [instance: 71d9eccc-3b90-4b0a-a59a-d6a8e08a9417] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 619.509957] env[61923]: ERROR nova.compute.manager [instance: 71d9eccc-3b90-4b0a-a59a-d6a8e08a9417] _ensure_no_port_binding_failure(port) [ 619.509957] env[61923]: ERROR nova.compute.manager [instance: 71d9eccc-3b90-4b0a-a59a-d6a8e08a9417] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 619.509957] env[61923]: ERROR nova.compute.manager [instance: 71d9eccc-3b90-4b0a-a59a-d6a8e08a9417] raise exception.PortBindingFailed(port_id=port['id']) [ 619.509957] env[61923]: ERROR nova.compute.manager [instance: 71d9eccc-3b90-4b0a-a59a-d6a8e08a9417] nova.exception.PortBindingFailed: Binding failed for port b42828d3-f036-4a41-949a-3505532203d9, please check neutron logs for more information. [ 619.509957] env[61923]: ERROR nova.compute.manager [instance: 71d9eccc-3b90-4b0a-a59a-d6a8e08a9417] [ 619.510273] env[61923]: DEBUG nova.compute.utils [None req-595ebb22-536b-4375-9cbe-6fc472a5a17d tempest-ListServerFiltersTestJSON-483030130 tempest-ListServerFiltersTestJSON-483030130-project-member] [instance: 71d9eccc-3b90-4b0a-a59a-d6a8e08a9417] Binding failed for port b42828d3-f036-4a41-949a-3505532203d9, please check neutron logs for more information. {{(pid=61923) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 619.511016] env[61923]: DEBUG oslo_concurrency.lockutils [None req-03cb7270-244b-47b9-9dd5-e358ebdfa946 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.518s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 619.516755] env[61923]: DEBUG nova.compute.manager [None req-595ebb22-536b-4375-9cbe-6fc472a5a17d tempest-ListServerFiltersTestJSON-483030130 tempest-ListServerFiltersTestJSON-483030130-project-member] [instance: 71d9eccc-3b90-4b0a-a59a-d6a8e08a9417] Build of instance 71d9eccc-3b90-4b0a-a59a-d6a8e08a9417 was re-scheduled: Binding failed for port b42828d3-f036-4a41-949a-3505532203d9, please check neutron logs for more information. {{(pid=61923) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 619.519995] env[61923]: DEBUG nova.compute.manager [None req-595ebb22-536b-4375-9cbe-6fc472a5a17d tempest-ListServerFiltersTestJSON-483030130 tempest-ListServerFiltersTestJSON-483030130-project-member] [instance: 71d9eccc-3b90-4b0a-a59a-d6a8e08a9417] Unplugging VIFs for instance {{(pid=61923) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 619.520251] env[61923]: DEBUG oslo_concurrency.lockutils [None req-595ebb22-536b-4375-9cbe-6fc472a5a17d tempest-ListServerFiltersTestJSON-483030130 tempest-ListServerFiltersTestJSON-483030130-project-member] Acquiring lock "refresh_cache-71d9eccc-3b90-4b0a-a59a-d6a8e08a9417" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 619.520399] env[61923]: DEBUG oslo_concurrency.lockutils [None req-595ebb22-536b-4375-9cbe-6fc472a5a17d tempest-ListServerFiltersTestJSON-483030130 tempest-ListServerFiltersTestJSON-483030130-project-member] Acquired lock "refresh_cache-71d9eccc-3b90-4b0a-a59a-d6a8e08a9417" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 619.520557] env[61923]: DEBUG nova.network.neutron [None req-595ebb22-536b-4375-9cbe-6fc472a5a17d tempest-ListServerFiltersTestJSON-483030130 tempest-ListServerFiltersTestJSON-483030130-project-member] [instance: 71d9eccc-3b90-4b0a-a59a-d6a8e08a9417] Building network info cache for instance {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 619.656205] env[61923]: DEBUG oslo_concurrency.lockutils [None req-4a9f7dc2-7221-4911-9bc8-b39b7a22ddda tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Acquiring lock "22d1e8ed-75c5-42e9-ae55-12d59e565914" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 619.656518] env[61923]: DEBUG oslo_concurrency.lockutils [None req-4a9f7dc2-7221-4911-9bc8-b39b7a22ddda tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Lock "22d1e8ed-75c5-42e9-ae55-12d59e565914" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 619.663240] env[61923]: DEBUG nova.compute.manager [req-be09d243-ce07-470c-8fae-90653b2f45e0 req-a6ad39a8-f588-4ea0-83b5-cdf8ea160336 service nova] [instance: 119bb4a7-2d71-44d7-ae81-9238512937c8] Received event network-vif-deleted-37c91d0a-5c78-46e0-800e-9c9760eb9fbc {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 619.691699] env[61923]: INFO nova.compute.manager [None req-58d8b214-0b32-4993-bdd8-d652de7f9397 tempest-ServerActionsTestOtherA-1581037736 tempest-ServerActionsTestOtherA-1581037736-project-member] [instance: af5feda7-204a-47f1-b7fa-afe44832b9a2] Took 1.03 seconds to deallocate network for instance. [ 619.940217] env[61923]: DEBUG oslo_concurrency.lockutils [None req-b2fe22fb-da8f-41ba-8845-95fc6177390e tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Releasing lock "refresh_cache-119bb4a7-2d71-44d7-ae81-9238512937c8" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 619.940656] env[61923]: DEBUG nova.compute.manager [None req-b2fe22fb-da8f-41ba-8845-95fc6177390e tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] [instance: 119bb4a7-2d71-44d7-ae81-9238512937c8] Start destroying the instance on the hypervisor. {{(pid=61923) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 619.940843] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-b2fe22fb-da8f-41ba-8845-95fc6177390e tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] [instance: 119bb4a7-2d71-44d7-ae81-9238512937c8] Destroying instance {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 619.941167] env[61923]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-947aee10-60d0-409f-bc34-55429eaa6531 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.949870] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25830bd2-b13a-4bdb-b37f-07fcb83c39b2 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.972777] env[61923]: WARNING nova.virt.vmwareapi.vmops [None req-b2fe22fb-da8f-41ba-8845-95fc6177390e tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] [instance: 119bb4a7-2d71-44d7-ae81-9238512937c8] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 119bb4a7-2d71-44d7-ae81-9238512937c8 could not be found. [ 619.972991] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-b2fe22fb-da8f-41ba-8845-95fc6177390e tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] [instance: 119bb4a7-2d71-44d7-ae81-9238512937c8] Instance destroyed {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 619.973182] env[61923]: INFO nova.compute.manager [None req-b2fe22fb-da8f-41ba-8845-95fc6177390e tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] [instance: 119bb4a7-2d71-44d7-ae81-9238512937c8] Took 0.03 seconds to destroy the instance on the hypervisor. [ 619.973410] env[61923]: DEBUG oslo.service.loopingcall [None req-b2fe22fb-da8f-41ba-8845-95fc6177390e tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61923) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 619.973845] env[61923]: DEBUG nova.compute.manager [-] [instance: 119bb4a7-2d71-44d7-ae81-9238512937c8] Deallocating network for instance {{(pid=61923) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 619.973845] env[61923]: DEBUG nova.network.neutron [-] [instance: 119bb4a7-2d71-44d7-ae81-9238512937c8] deallocate_for_instance() {{(pid=61923) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 619.996147] env[61923]: DEBUG nova.network.neutron [-] [instance: 119bb4a7-2d71-44d7-ae81-9238512937c8] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 620.050832] env[61923]: DEBUG nova.network.neutron [None req-595ebb22-536b-4375-9cbe-6fc472a5a17d tempest-ListServerFiltersTestJSON-483030130 tempest-ListServerFiltersTestJSON-483030130-project-member] [instance: 71d9eccc-3b90-4b0a-a59a-d6a8e08a9417] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 620.173518] env[61923]: DEBUG nova.network.neutron [None req-595ebb22-536b-4375-9cbe-6fc472a5a17d tempest-ListServerFiltersTestJSON-483030130 tempest-ListServerFiltersTestJSON-483030130-project-member] [instance: 71d9eccc-3b90-4b0a-a59a-d6a8e08a9417] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 620.449548] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b785c1f-b478-4ee9-bb5c-6a76895e0373 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.457920] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de40b2ab-2613-499f-828e-31f4ca5550db {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.486580] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-395f9e99-e538-49d5-9562-65151773cbb4 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.493935] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3d5e86f-ebd4-426e-b388-edbf57afd146 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.499148] env[61923]: DEBUG nova.network.neutron [-] [instance: 119bb4a7-2d71-44d7-ae81-9238512937c8] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 620.513074] env[61923]: DEBUG nova.compute.provider_tree [None req-03cb7270-244b-47b9-9dd5-e358ebdfa946 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 620.679144] env[61923]: DEBUG oslo_concurrency.lockutils [None req-595ebb22-536b-4375-9cbe-6fc472a5a17d tempest-ListServerFiltersTestJSON-483030130 tempest-ListServerFiltersTestJSON-483030130-project-member] Releasing lock "refresh_cache-71d9eccc-3b90-4b0a-a59a-d6a8e08a9417" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 620.679144] env[61923]: DEBUG nova.compute.manager [None req-595ebb22-536b-4375-9cbe-6fc472a5a17d tempest-ListServerFiltersTestJSON-483030130 tempest-ListServerFiltersTestJSON-483030130-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61923) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 620.679144] env[61923]: DEBUG nova.compute.manager [None req-595ebb22-536b-4375-9cbe-6fc472a5a17d tempest-ListServerFiltersTestJSON-483030130 tempest-ListServerFiltersTestJSON-483030130-project-member] [instance: 71d9eccc-3b90-4b0a-a59a-d6a8e08a9417] Deallocating network for instance {{(pid=61923) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 620.679144] env[61923]: DEBUG nova.network.neutron [None req-595ebb22-536b-4375-9cbe-6fc472a5a17d tempest-ListServerFiltersTestJSON-483030130 tempest-ListServerFiltersTestJSON-483030130-project-member] [instance: 71d9eccc-3b90-4b0a-a59a-d6a8e08a9417] deallocate_for_instance() {{(pid=61923) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 620.701854] env[61923]: DEBUG nova.network.neutron [None req-595ebb22-536b-4375-9cbe-6fc472a5a17d tempest-ListServerFiltersTestJSON-483030130 tempest-ListServerFiltersTestJSON-483030130-project-member] [instance: 71d9eccc-3b90-4b0a-a59a-d6a8e08a9417] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 620.733188] env[61923]: INFO nova.scheduler.client.report [None req-58d8b214-0b32-4993-bdd8-d652de7f9397 tempest-ServerActionsTestOtherA-1581037736 tempest-ServerActionsTestOtherA-1581037736-project-member] Deleted allocations for instance af5feda7-204a-47f1-b7fa-afe44832b9a2 [ 621.004055] env[61923]: INFO nova.compute.manager [-] [instance: 119bb4a7-2d71-44d7-ae81-9238512937c8] Took 1.03 seconds to deallocate network for instance. [ 621.006776] env[61923]: DEBUG nova.compute.claims [None req-b2fe22fb-da8f-41ba-8845-95fc6177390e tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] [instance: 119bb4a7-2d71-44d7-ae81-9238512937c8] Aborting claim: {{(pid=61923) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 621.007024] env[61923]: DEBUG oslo_concurrency.lockutils [None req-b2fe22fb-da8f-41ba-8845-95fc6177390e tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 621.015825] env[61923]: DEBUG nova.scheduler.client.report [None req-03cb7270-244b-47b9-9dd5-e358ebdfa946 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 621.206219] env[61923]: DEBUG nova.network.neutron [None req-595ebb22-536b-4375-9cbe-6fc472a5a17d tempest-ListServerFiltersTestJSON-483030130 tempest-ListServerFiltersTestJSON-483030130-project-member] [instance: 71d9eccc-3b90-4b0a-a59a-d6a8e08a9417] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 621.241386] env[61923]: DEBUG oslo_concurrency.lockutils [None req-58d8b214-0b32-4993-bdd8-d652de7f9397 tempest-ServerActionsTestOtherA-1581037736 tempest-ServerActionsTestOtherA-1581037736-project-member] Lock "af5feda7-204a-47f1-b7fa-afe44832b9a2" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 81.238s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 621.523199] env[61923]: DEBUG oslo_concurrency.lockutils [None req-03cb7270-244b-47b9-9dd5-e358ebdfa946 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.012s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 621.523875] env[61923]: ERROR nova.compute.manager [None req-03cb7270-244b-47b9-9dd5-e358ebdfa946 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: cb998da7-8db0-4725-bed2-657d7df748b5] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 225a6542-d2aa-4f13-a9d6-7c2e6c048241, please check neutron logs for more information. [ 621.523875] env[61923]: ERROR nova.compute.manager [instance: cb998da7-8db0-4725-bed2-657d7df748b5] Traceback (most recent call last): [ 621.523875] env[61923]: ERROR nova.compute.manager [instance: cb998da7-8db0-4725-bed2-657d7df748b5] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 621.523875] env[61923]: ERROR nova.compute.manager [instance: cb998da7-8db0-4725-bed2-657d7df748b5] self.driver.spawn(context, instance, image_meta, [ 621.523875] env[61923]: ERROR nova.compute.manager [instance: cb998da7-8db0-4725-bed2-657d7df748b5] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 621.523875] env[61923]: ERROR nova.compute.manager [instance: cb998da7-8db0-4725-bed2-657d7df748b5] self._vmops.spawn(context, instance, image_meta, injected_files, [ 621.523875] env[61923]: ERROR nova.compute.manager [instance: cb998da7-8db0-4725-bed2-657d7df748b5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 621.523875] env[61923]: ERROR nova.compute.manager [instance: cb998da7-8db0-4725-bed2-657d7df748b5] vm_ref = self.build_virtual_machine(instance, [ 621.523875] env[61923]: ERROR nova.compute.manager [instance: cb998da7-8db0-4725-bed2-657d7df748b5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 621.523875] env[61923]: ERROR nova.compute.manager [instance: cb998da7-8db0-4725-bed2-657d7df748b5] vif_infos = vmwarevif.get_vif_info(self._session, [ 621.523875] env[61923]: ERROR nova.compute.manager [instance: cb998da7-8db0-4725-bed2-657d7df748b5] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 621.524296] env[61923]: ERROR nova.compute.manager [instance: cb998da7-8db0-4725-bed2-657d7df748b5] for vif in network_info: [ 621.524296] env[61923]: ERROR nova.compute.manager [instance: cb998da7-8db0-4725-bed2-657d7df748b5] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 621.524296] env[61923]: ERROR nova.compute.manager [instance: cb998da7-8db0-4725-bed2-657d7df748b5] return self._sync_wrapper(fn, *args, **kwargs) [ 621.524296] env[61923]: ERROR nova.compute.manager [instance: cb998da7-8db0-4725-bed2-657d7df748b5] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 621.524296] env[61923]: ERROR nova.compute.manager [instance: cb998da7-8db0-4725-bed2-657d7df748b5] self.wait() [ 621.524296] env[61923]: ERROR nova.compute.manager [instance: cb998da7-8db0-4725-bed2-657d7df748b5] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 621.524296] env[61923]: ERROR nova.compute.manager [instance: cb998da7-8db0-4725-bed2-657d7df748b5] self[:] = self._gt.wait() [ 621.524296] env[61923]: ERROR nova.compute.manager [instance: cb998da7-8db0-4725-bed2-657d7df748b5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 621.524296] env[61923]: ERROR nova.compute.manager [instance: cb998da7-8db0-4725-bed2-657d7df748b5] return self._exit_event.wait() [ 621.524296] env[61923]: ERROR nova.compute.manager [instance: cb998da7-8db0-4725-bed2-657d7df748b5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 621.524296] env[61923]: ERROR nova.compute.manager [instance: cb998da7-8db0-4725-bed2-657d7df748b5] result = hub.switch() [ 621.524296] env[61923]: ERROR nova.compute.manager [instance: cb998da7-8db0-4725-bed2-657d7df748b5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 621.524296] env[61923]: ERROR nova.compute.manager [instance: cb998da7-8db0-4725-bed2-657d7df748b5] return self.greenlet.switch() [ 621.524707] env[61923]: ERROR nova.compute.manager [instance: cb998da7-8db0-4725-bed2-657d7df748b5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 621.524707] env[61923]: ERROR nova.compute.manager [instance: cb998da7-8db0-4725-bed2-657d7df748b5] result = function(*args, **kwargs) [ 621.524707] env[61923]: ERROR nova.compute.manager [instance: cb998da7-8db0-4725-bed2-657d7df748b5] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 621.524707] env[61923]: ERROR nova.compute.manager [instance: cb998da7-8db0-4725-bed2-657d7df748b5] return func(*args, **kwargs) [ 621.524707] env[61923]: ERROR nova.compute.manager [instance: cb998da7-8db0-4725-bed2-657d7df748b5] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 621.524707] env[61923]: ERROR nova.compute.manager [instance: cb998da7-8db0-4725-bed2-657d7df748b5] raise e [ 621.524707] env[61923]: ERROR nova.compute.manager [instance: cb998da7-8db0-4725-bed2-657d7df748b5] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 621.524707] env[61923]: ERROR nova.compute.manager [instance: cb998da7-8db0-4725-bed2-657d7df748b5] nwinfo = self.network_api.allocate_for_instance( [ 621.524707] env[61923]: ERROR nova.compute.manager [instance: cb998da7-8db0-4725-bed2-657d7df748b5] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 621.524707] env[61923]: ERROR nova.compute.manager [instance: cb998da7-8db0-4725-bed2-657d7df748b5] created_port_ids = self._update_ports_for_instance( [ 621.524707] env[61923]: ERROR nova.compute.manager [instance: cb998da7-8db0-4725-bed2-657d7df748b5] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 621.524707] env[61923]: ERROR nova.compute.manager [instance: cb998da7-8db0-4725-bed2-657d7df748b5] with excutils.save_and_reraise_exception(): [ 621.524707] env[61923]: ERROR nova.compute.manager [instance: cb998da7-8db0-4725-bed2-657d7df748b5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 621.525127] env[61923]: ERROR nova.compute.manager [instance: cb998da7-8db0-4725-bed2-657d7df748b5] self.force_reraise() [ 621.525127] env[61923]: ERROR nova.compute.manager [instance: cb998da7-8db0-4725-bed2-657d7df748b5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 621.525127] env[61923]: ERROR nova.compute.manager [instance: cb998da7-8db0-4725-bed2-657d7df748b5] raise self.value [ 621.525127] env[61923]: ERROR nova.compute.manager [instance: cb998da7-8db0-4725-bed2-657d7df748b5] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 621.525127] env[61923]: ERROR nova.compute.manager [instance: cb998da7-8db0-4725-bed2-657d7df748b5] updated_port = self._update_port( [ 621.525127] env[61923]: ERROR nova.compute.manager [instance: cb998da7-8db0-4725-bed2-657d7df748b5] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 621.525127] env[61923]: ERROR nova.compute.manager [instance: cb998da7-8db0-4725-bed2-657d7df748b5] _ensure_no_port_binding_failure(port) [ 621.525127] env[61923]: ERROR nova.compute.manager [instance: cb998da7-8db0-4725-bed2-657d7df748b5] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 621.525127] env[61923]: ERROR nova.compute.manager [instance: cb998da7-8db0-4725-bed2-657d7df748b5] raise exception.PortBindingFailed(port_id=port['id']) [ 621.525127] env[61923]: ERROR nova.compute.manager [instance: cb998da7-8db0-4725-bed2-657d7df748b5] nova.exception.PortBindingFailed: Binding failed for port 225a6542-d2aa-4f13-a9d6-7c2e6c048241, please check neutron logs for more information. [ 621.525127] env[61923]: ERROR nova.compute.manager [instance: cb998da7-8db0-4725-bed2-657d7df748b5] [ 621.525480] env[61923]: DEBUG nova.compute.utils [None req-03cb7270-244b-47b9-9dd5-e358ebdfa946 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: cb998da7-8db0-4725-bed2-657d7df748b5] Binding failed for port 225a6542-d2aa-4f13-a9d6-7c2e6c048241, please check neutron logs for more information. {{(pid=61923) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 621.525932] env[61923]: DEBUG oslo_concurrency.lockutils [None req-aa9c464b-ac6b-44fe-9384-28cd9bdfecbc tempest-ImagesOneServerTestJSON-63657725 tempest-ImagesOneServerTestJSON-63657725-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.251s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 621.529092] env[61923]: INFO nova.compute.claims [None req-aa9c464b-ac6b-44fe-9384-28cd9bdfecbc tempest-ImagesOneServerTestJSON-63657725 tempest-ImagesOneServerTestJSON-63657725-project-member] [instance: 0c1b8117-199f-4101-ab81-4ae7ef0d1251] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 621.530375] env[61923]: DEBUG nova.compute.manager [None req-03cb7270-244b-47b9-9dd5-e358ebdfa946 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: cb998da7-8db0-4725-bed2-657d7df748b5] Build of instance cb998da7-8db0-4725-bed2-657d7df748b5 was re-scheduled: Binding failed for port 225a6542-d2aa-4f13-a9d6-7c2e6c048241, please check neutron logs for more information. {{(pid=61923) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 621.530831] env[61923]: DEBUG nova.compute.manager [None req-03cb7270-244b-47b9-9dd5-e358ebdfa946 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: cb998da7-8db0-4725-bed2-657d7df748b5] Unplugging VIFs for instance {{(pid=61923) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 621.531101] env[61923]: DEBUG oslo_concurrency.lockutils [None req-03cb7270-244b-47b9-9dd5-e358ebdfa946 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Acquiring lock "refresh_cache-cb998da7-8db0-4725-bed2-657d7df748b5" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 621.533351] env[61923]: DEBUG oslo_concurrency.lockutils [None req-03cb7270-244b-47b9-9dd5-e358ebdfa946 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Acquired lock "refresh_cache-cb998da7-8db0-4725-bed2-657d7df748b5" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 621.533527] env[61923]: DEBUG nova.network.neutron [None req-03cb7270-244b-47b9-9dd5-e358ebdfa946 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: cb998da7-8db0-4725-bed2-657d7df748b5] Building network info cache for instance {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 621.711878] env[61923]: INFO nova.compute.manager [None req-595ebb22-536b-4375-9cbe-6fc472a5a17d tempest-ListServerFiltersTestJSON-483030130 tempest-ListServerFiltersTestJSON-483030130-project-member] [instance: 71d9eccc-3b90-4b0a-a59a-d6a8e08a9417] Took 1.03 seconds to deallocate network for instance. [ 621.744232] env[61923]: DEBUG nova.compute.manager [None req-76010439-9fb6-4296-9b5a-107842db76db tempest-ServerExternalEventsTest-1479796458 tempest-ServerExternalEventsTest-1479796458-project-member] [instance: c0f87aa5-447b-4920-8251-354d89a14fbb] Starting instance... {{(pid=61923) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 622.072570] env[61923]: DEBUG nova.network.neutron [None req-03cb7270-244b-47b9-9dd5-e358ebdfa946 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: cb998da7-8db0-4725-bed2-657d7df748b5] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 622.217348] env[61923]: DEBUG nova.network.neutron [None req-03cb7270-244b-47b9-9dd5-e358ebdfa946 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: cb998da7-8db0-4725-bed2-657d7df748b5] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 622.274511] env[61923]: DEBUG oslo_concurrency.lockutils [None req-76010439-9fb6-4296-9b5a-107842db76db tempest-ServerExternalEventsTest-1479796458 tempest-ServerExternalEventsTest-1479796458-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 622.725382] env[61923]: DEBUG oslo_concurrency.lockutils [None req-03cb7270-244b-47b9-9dd5-e358ebdfa946 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Releasing lock "refresh_cache-cb998da7-8db0-4725-bed2-657d7df748b5" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 622.725657] env[61923]: DEBUG nova.compute.manager [None req-03cb7270-244b-47b9-9dd5-e358ebdfa946 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61923) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 622.725807] env[61923]: DEBUG nova.compute.manager [None req-03cb7270-244b-47b9-9dd5-e358ebdfa946 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: cb998da7-8db0-4725-bed2-657d7df748b5] Deallocating network for instance {{(pid=61923) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 622.726411] env[61923]: DEBUG nova.network.neutron [None req-03cb7270-244b-47b9-9dd5-e358ebdfa946 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: cb998da7-8db0-4725-bed2-657d7df748b5] deallocate_for_instance() {{(pid=61923) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 622.750731] env[61923]: INFO nova.scheduler.client.report [None req-595ebb22-536b-4375-9cbe-6fc472a5a17d tempest-ListServerFiltersTestJSON-483030130 tempest-ListServerFiltersTestJSON-483030130-project-member] Deleted allocations for instance 71d9eccc-3b90-4b0a-a59a-d6a8e08a9417 [ 622.757753] env[61923]: DEBUG nova.network.neutron [None req-03cb7270-244b-47b9-9dd5-e358ebdfa946 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: cb998da7-8db0-4725-bed2-657d7df748b5] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 622.968222] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a15879a2-64c2-4a6d-bb44-c705c5961780 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.976982] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9c6c3c8-5c2b-4f34-83b6-4afd5c046247 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.009895] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b0ab42a-aa42-4ceb-b75e-11063faa7a94 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.018125] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bbaa1c7a-0aa2-42fb-915d-d026c68f5494 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 623.037317] env[61923]: DEBUG nova.compute.provider_tree [None req-aa9c464b-ac6b-44fe-9384-28cd9bdfecbc tempest-ImagesOneServerTestJSON-63657725 tempest-ImagesOneServerTestJSON-63657725-project-member] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 623.263575] env[61923]: DEBUG nova.network.neutron [None req-03cb7270-244b-47b9-9dd5-e358ebdfa946 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: cb998da7-8db0-4725-bed2-657d7df748b5] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 623.264936] env[61923]: DEBUG oslo_concurrency.lockutils [None req-595ebb22-536b-4375-9cbe-6fc472a5a17d tempest-ListServerFiltersTestJSON-483030130 tempest-ListServerFiltersTestJSON-483030130-project-member] Lock "71d9eccc-3b90-4b0a-a59a-d6a8e08a9417" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 83.182s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 623.537760] env[61923]: DEBUG nova.scheduler.client.report [None req-aa9c464b-ac6b-44fe-9384-28cd9bdfecbc tempest-ImagesOneServerTestJSON-63657725 tempest-ImagesOneServerTestJSON-63657725-project-member] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 623.767450] env[61923]: INFO nova.compute.manager [None req-03cb7270-244b-47b9-9dd5-e358ebdfa946 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: cb998da7-8db0-4725-bed2-657d7df748b5] Took 1.04 seconds to deallocate network for instance. [ 623.770284] env[61923]: DEBUG nova.compute.manager [None req-85ac5bbf-a6d0-481b-b0bf-1428265b44a9 tempest-AttachInterfacesV270Test-411696101 tempest-AttachInterfacesV270Test-411696101-project-member] [instance: 525f37d5-2262-40c8-a339-fc262c53ce6e] Starting instance... {{(pid=61923) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 624.043964] env[61923]: DEBUG oslo_concurrency.lockutils [None req-aa9c464b-ac6b-44fe-9384-28cd9bdfecbc tempest-ImagesOneServerTestJSON-63657725 tempest-ImagesOneServerTestJSON-63657725-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.518s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 624.044813] env[61923]: DEBUG nova.compute.manager [None req-aa9c464b-ac6b-44fe-9384-28cd9bdfecbc tempest-ImagesOneServerTestJSON-63657725 tempest-ImagesOneServerTestJSON-63657725-project-member] [instance: 0c1b8117-199f-4101-ab81-4ae7ef0d1251] Start building networks asynchronously for instance. {{(pid=61923) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 624.047986] env[61923]: DEBUG oslo_concurrency.lockutils [None req-0c55849d-766e-4fac-b1c4-e4e97ba688a6 tempest-DeleteServersAdminTestJSON-590116201 tempest-DeleteServersAdminTestJSON-590116201-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.745s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 624.049956] env[61923]: INFO nova.compute.claims [None req-0c55849d-766e-4fac-b1c4-e4e97ba688a6 tempest-DeleteServersAdminTestJSON-590116201 tempest-DeleteServersAdminTestJSON-590116201-project-member] [instance: 1b509288-2df3-4927-afcf-cb76918d3da3] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 624.305103] env[61923]: DEBUG oslo_concurrency.lockutils [None req-85ac5bbf-a6d0-481b-b0bf-1428265b44a9 tempest-AttachInterfacesV270Test-411696101 tempest-AttachInterfacesV270Test-411696101-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 624.554106] env[61923]: DEBUG nova.compute.utils [None req-aa9c464b-ac6b-44fe-9384-28cd9bdfecbc tempest-ImagesOneServerTestJSON-63657725 tempest-ImagesOneServerTestJSON-63657725-project-member] Using /dev/sd instead of None {{(pid=61923) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 624.558183] env[61923]: DEBUG nova.compute.manager [None req-aa9c464b-ac6b-44fe-9384-28cd9bdfecbc tempest-ImagesOneServerTestJSON-63657725 tempest-ImagesOneServerTestJSON-63657725-project-member] [instance: 0c1b8117-199f-4101-ab81-4ae7ef0d1251] Allocating IP information in the background. {{(pid=61923) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 624.558183] env[61923]: DEBUG nova.network.neutron [None req-aa9c464b-ac6b-44fe-9384-28cd9bdfecbc tempest-ImagesOneServerTestJSON-63657725 tempest-ImagesOneServerTestJSON-63657725-project-member] [instance: 0c1b8117-199f-4101-ab81-4ae7ef0d1251] allocate_for_instance() {{(pid=61923) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 624.601073] env[61923]: DEBUG nova.policy [None req-aa9c464b-ac6b-44fe-9384-28cd9bdfecbc tempest-ImagesOneServerTestJSON-63657725 tempest-ImagesOneServerTestJSON-63657725-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0a4851142de446eeb5b546a12698cd7b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ff8171d634664e70b58bd29b7e73546a', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61923) authorize /opt/stack/nova/nova/policy.py:201}} [ 624.813984] env[61923]: INFO nova.scheduler.client.report [None req-03cb7270-244b-47b9-9dd5-e358ebdfa946 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Deleted allocations for instance cb998da7-8db0-4725-bed2-657d7df748b5 [ 624.956648] env[61923]: DEBUG nova.network.neutron [None req-aa9c464b-ac6b-44fe-9384-28cd9bdfecbc tempest-ImagesOneServerTestJSON-63657725 tempest-ImagesOneServerTestJSON-63657725-project-member] [instance: 0c1b8117-199f-4101-ab81-4ae7ef0d1251] Successfully created port: 9ba992af-f024-4728-b3b6-b6cf6af4e319 {{(pid=61923) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 625.061140] env[61923]: DEBUG nova.compute.manager [None req-aa9c464b-ac6b-44fe-9384-28cd9bdfecbc tempest-ImagesOneServerTestJSON-63657725 tempest-ImagesOneServerTestJSON-63657725-project-member] [instance: 0c1b8117-199f-4101-ab81-4ae7ef0d1251] Start building block device mappings for instance. {{(pid=61923) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 625.326228] env[61923]: DEBUG oslo_concurrency.lockutils [None req-03cb7270-244b-47b9-9dd5-e358ebdfa946 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Lock "cb998da7-8db0-4725-bed2-657d7df748b5" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 83.980s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 625.453375] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1d66df1-7ea0-46c2-85dd-1cb87ea67637 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.464310] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee708be7-a7b4-4a50-b627-ed7b54df0b9f {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.501073] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-559cf5e2-95bc-4e3f-a3a9-4aa46fa5589d {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.508752] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba0bb184-6897-4ef4-aed9-819469353c4e {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 625.523226] env[61923]: DEBUG nova.compute.provider_tree [None req-0c55849d-766e-4fac-b1c4-e4e97ba688a6 tempest-DeleteServersAdminTestJSON-590116201 tempest-DeleteServersAdminTestJSON-590116201-project-member] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 625.829998] env[61923]: DEBUG nova.compute.manager [None req-4bd5c12b-d52a-4c54-93f7-87894f09013a tempest-ServersTestJSON-483293076 tempest-ServersTestJSON-483293076-project-member] [instance: 55b60fc2-6484-4fad-aa40-3e73e020841a] Starting instance... {{(pid=61923) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 625.976997] env[61923]: DEBUG nova.compute.manager [req-c0f2ceaf-a105-4a18-aa8f-88675b047967 req-1b6f47bc-6097-42c5-a20c-03cd0a551b12 service nova] [instance: 0c1b8117-199f-4101-ab81-4ae7ef0d1251] Received event network-changed-9ba992af-f024-4728-b3b6-b6cf6af4e319 {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 625.977243] env[61923]: DEBUG nova.compute.manager [req-c0f2ceaf-a105-4a18-aa8f-88675b047967 req-1b6f47bc-6097-42c5-a20c-03cd0a551b12 service nova] [instance: 0c1b8117-199f-4101-ab81-4ae7ef0d1251] Refreshing instance network info cache due to event network-changed-9ba992af-f024-4728-b3b6-b6cf6af4e319. {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 625.977465] env[61923]: DEBUG oslo_concurrency.lockutils [req-c0f2ceaf-a105-4a18-aa8f-88675b047967 req-1b6f47bc-6097-42c5-a20c-03cd0a551b12 service nova] Acquiring lock "refresh_cache-0c1b8117-199f-4101-ab81-4ae7ef0d1251" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 625.977609] env[61923]: DEBUG oslo_concurrency.lockutils [req-c0f2ceaf-a105-4a18-aa8f-88675b047967 req-1b6f47bc-6097-42c5-a20c-03cd0a551b12 service nova] Acquired lock "refresh_cache-0c1b8117-199f-4101-ab81-4ae7ef0d1251" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 625.977765] env[61923]: DEBUG nova.network.neutron [req-c0f2ceaf-a105-4a18-aa8f-88675b047967 req-1b6f47bc-6097-42c5-a20c-03cd0a551b12 service nova] [instance: 0c1b8117-199f-4101-ab81-4ae7ef0d1251] Refreshing network info cache for port 9ba992af-f024-4728-b3b6-b6cf6af4e319 {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 626.026412] env[61923]: DEBUG nova.scheduler.client.report [None req-0c55849d-766e-4fac-b1c4-e4e97ba688a6 tempest-DeleteServersAdminTestJSON-590116201 tempest-DeleteServersAdminTestJSON-590116201-project-member] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 626.051822] env[61923]: DEBUG oslo_service.periodic_task [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61923) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 626.052119] env[61923]: DEBUG oslo_service.periodic_task [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=61923) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 626.075848] env[61923]: DEBUG nova.compute.manager [None req-aa9c464b-ac6b-44fe-9384-28cd9bdfecbc tempest-ImagesOneServerTestJSON-63657725 tempest-ImagesOneServerTestJSON-63657725-project-member] [instance: 0c1b8117-199f-4101-ab81-4ae7ef0d1251] Start spawning the instance on the hypervisor. {{(pid=61923) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 626.100824] env[61923]: DEBUG nova.virt.hardware [None req-aa9c464b-ac6b-44fe-9384-28cd9bdfecbc tempest-ImagesOneServerTestJSON-63657725 tempest-ImagesOneServerTestJSON-63657725-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-29T20:07:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-29T20:07:35Z,direct_url=,disk_format='vmdk',id=0f153f63-ae0a-45b1-b7f5-7f9b673c947f,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='4e7b5e3b3c3443c8bd5c70f8a15739f5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-29T20:07:36Z,virtual_size=,visibility=), allow threads: False {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 626.101086] env[61923]: DEBUG nova.virt.hardware [None req-aa9c464b-ac6b-44fe-9384-28cd9bdfecbc tempest-ImagesOneServerTestJSON-63657725 tempest-ImagesOneServerTestJSON-63657725-project-member] Flavor limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 626.101237] env[61923]: DEBUG nova.virt.hardware [None req-aa9c464b-ac6b-44fe-9384-28cd9bdfecbc tempest-ImagesOneServerTestJSON-63657725 tempest-ImagesOneServerTestJSON-63657725-project-member] Image limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 626.101422] env[61923]: DEBUG nova.virt.hardware [None req-aa9c464b-ac6b-44fe-9384-28cd9bdfecbc tempest-ImagesOneServerTestJSON-63657725 tempest-ImagesOneServerTestJSON-63657725-project-member] Flavor pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 626.102337] env[61923]: DEBUG nova.virt.hardware [None req-aa9c464b-ac6b-44fe-9384-28cd9bdfecbc tempest-ImagesOneServerTestJSON-63657725 tempest-ImagesOneServerTestJSON-63657725-project-member] Image pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 626.102337] env[61923]: DEBUG nova.virt.hardware [None req-aa9c464b-ac6b-44fe-9384-28cd9bdfecbc tempest-ImagesOneServerTestJSON-63657725 tempest-ImagesOneServerTestJSON-63657725-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 626.102337] env[61923]: DEBUG nova.virt.hardware [None req-aa9c464b-ac6b-44fe-9384-28cd9bdfecbc tempest-ImagesOneServerTestJSON-63657725 tempest-ImagesOneServerTestJSON-63657725-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 626.102337] env[61923]: DEBUG nova.virt.hardware [None req-aa9c464b-ac6b-44fe-9384-28cd9bdfecbc tempest-ImagesOneServerTestJSON-63657725 tempest-ImagesOneServerTestJSON-63657725-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 626.102466] env[61923]: DEBUG nova.virt.hardware [None req-aa9c464b-ac6b-44fe-9384-28cd9bdfecbc tempest-ImagesOneServerTestJSON-63657725 tempest-ImagesOneServerTestJSON-63657725-project-member] Got 1 possible topologies {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 626.102733] env[61923]: DEBUG nova.virt.hardware [None req-aa9c464b-ac6b-44fe-9384-28cd9bdfecbc tempest-ImagesOneServerTestJSON-63657725 tempest-ImagesOneServerTestJSON-63657725-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 626.103815] env[61923]: DEBUG nova.virt.hardware [None req-aa9c464b-ac6b-44fe-9384-28cd9bdfecbc tempest-ImagesOneServerTestJSON-63657725 tempest-ImagesOneServerTestJSON-63657725-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 626.105982] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc5d46e7-8823-4370-94da-24f5155d56de {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.112450] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba0fc724-5d45-4383-b9f5-8216a925ba05 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.313648] env[61923]: ERROR nova.compute.manager [None req-aa9c464b-ac6b-44fe-9384-28cd9bdfecbc tempest-ImagesOneServerTestJSON-63657725 tempest-ImagesOneServerTestJSON-63657725-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 9ba992af-f024-4728-b3b6-b6cf6af4e319, please check neutron logs for more information. [ 626.313648] env[61923]: ERROR nova.compute.manager Traceback (most recent call last): [ 626.313648] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 626.313648] env[61923]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 626.313648] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 626.313648] env[61923]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 626.313648] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 626.313648] env[61923]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 626.313648] env[61923]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 626.313648] env[61923]: ERROR nova.compute.manager self.force_reraise() [ 626.313648] env[61923]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 626.313648] env[61923]: ERROR nova.compute.manager raise self.value [ 626.313648] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 626.313648] env[61923]: ERROR nova.compute.manager updated_port = self._update_port( [ 626.313648] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 626.313648] env[61923]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 626.315386] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 626.315386] env[61923]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 626.315386] env[61923]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 9ba992af-f024-4728-b3b6-b6cf6af4e319, please check neutron logs for more information. [ 626.315386] env[61923]: ERROR nova.compute.manager [ 626.315386] env[61923]: Traceback (most recent call last): [ 626.315386] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 626.315386] env[61923]: listener.cb(fileno) [ 626.315386] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 626.315386] env[61923]: result = function(*args, **kwargs) [ 626.315386] env[61923]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 626.315386] env[61923]: return func(*args, **kwargs) [ 626.315386] env[61923]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 626.315386] env[61923]: raise e [ 626.315386] env[61923]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 626.315386] env[61923]: nwinfo = self.network_api.allocate_for_instance( [ 626.315386] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 626.315386] env[61923]: created_port_ids = self._update_ports_for_instance( [ 626.315386] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 626.315386] env[61923]: with excutils.save_and_reraise_exception(): [ 626.315386] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 626.315386] env[61923]: self.force_reraise() [ 626.315386] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 626.315386] env[61923]: raise self.value [ 626.315386] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 626.315386] env[61923]: updated_port = self._update_port( [ 626.315386] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 626.315386] env[61923]: _ensure_no_port_binding_failure(port) [ 626.315386] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 626.315386] env[61923]: raise exception.PortBindingFailed(port_id=port['id']) [ 626.316254] env[61923]: nova.exception.PortBindingFailed: Binding failed for port 9ba992af-f024-4728-b3b6-b6cf6af4e319, please check neutron logs for more information. [ 626.316254] env[61923]: Removing descriptor: 17 [ 626.316254] env[61923]: ERROR nova.compute.manager [None req-aa9c464b-ac6b-44fe-9384-28cd9bdfecbc tempest-ImagesOneServerTestJSON-63657725 tempest-ImagesOneServerTestJSON-63657725-project-member] [instance: 0c1b8117-199f-4101-ab81-4ae7ef0d1251] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 9ba992af-f024-4728-b3b6-b6cf6af4e319, please check neutron logs for more information. [ 626.316254] env[61923]: ERROR nova.compute.manager [instance: 0c1b8117-199f-4101-ab81-4ae7ef0d1251] Traceback (most recent call last): [ 626.316254] env[61923]: ERROR nova.compute.manager [instance: 0c1b8117-199f-4101-ab81-4ae7ef0d1251] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 626.316254] env[61923]: ERROR nova.compute.manager [instance: 0c1b8117-199f-4101-ab81-4ae7ef0d1251] yield resources [ 626.316254] env[61923]: ERROR nova.compute.manager [instance: 0c1b8117-199f-4101-ab81-4ae7ef0d1251] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 626.316254] env[61923]: ERROR nova.compute.manager [instance: 0c1b8117-199f-4101-ab81-4ae7ef0d1251] self.driver.spawn(context, instance, image_meta, [ 626.316254] env[61923]: ERROR nova.compute.manager [instance: 0c1b8117-199f-4101-ab81-4ae7ef0d1251] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 626.316254] env[61923]: ERROR nova.compute.manager [instance: 0c1b8117-199f-4101-ab81-4ae7ef0d1251] self._vmops.spawn(context, instance, image_meta, injected_files, [ 626.316254] env[61923]: ERROR nova.compute.manager [instance: 0c1b8117-199f-4101-ab81-4ae7ef0d1251] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 626.316254] env[61923]: ERROR nova.compute.manager [instance: 0c1b8117-199f-4101-ab81-4ae7ef0d1251] vm_ref = self.build_virtual_machine(instance, [ 626.316582] env[61923]: ERROR nova.compute.manager [instance: 0c1b8117-199f-4101-ab81-4ae7ef0d1251] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 626.316582] env[61923]: ERROR nova.compute.manager [instance: 0c1b8117-199f-4101-ab81-4ae7ef0d1251] vif_infos = vmwarevif.get_vif_info(self._session, [ 626.316582] env[61923]: ERROR nova.compute.manager [instance: 0c1b8117-199f-4101-ab81-4ae7ef0d1251] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 626.316582] env[61923]: ERROR nova.compute.manager [instance: 0c1b8117-199f-4101-ab81-4ae7ef0d1251] for vif in network_info: [ 626.316582] env[61923]: ERROR nova.compute.manager [instance: 0c1b8117-199f-4101-ab81-4ae7ef0d1251] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 626.316582] env[61923]: ERROR nova.compute.manager [instance: 0c1b8117-199f-4101-ab81-4ae7ef0d1251] return self._sync_wrapper(fn, *args, **kwargs) [ 626.316582] env[61923]: ERROR nova.compute.manager [instance: 0c1b8117-199f-4101-ab81-4ae7ef0d1251] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 626.316582] env[61923]: ERROR nova.compute.manager [instance: 0c1b8117-199f-4101-ab81-4ae7ef0d1251] self.wait() [ 626.316582] env[61923]: ERROR nova.compute.manager [instance: 0c1b8117-199f-4101-ab81-4ae7ef0d1251] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 626.316582] env[61923]: ERROR nova.compute.manager [instance: 0c1b8117-199f-4101-ab81-4ae7ef0d1251] self[:] = self._gt.wait() [ 626.316582] env[61923]: ERROR nova.compute.manager [instance: 0c1b8117-199f-4101-ab81-4ae7ef0d1251] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 626.316582] env[61923]: ERROR nova.compute.manager [instance: 0c1b8117-199f-4101-ab81-4ae7ef0d1251] return self._exit_event.wait() [ 626.316582] env[61923]: ERROR nova.compute.manager [instance: 0c1b8117-199f-4101-ab81-4ae7ef0d1251] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 626.316879] env[61923]: ERROR nova.compute.manager [instance: 0c1b8117-199f-4101-ab81-4ae7ef0d1251] result = hub.switch() [ 626.316879] env[61923]: ERROR nova.compute.manager [instance: 0c1b8117-199f-4101-ab81-4ae7ef0d1251] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 626.316879] env[61923]: ERROR nova.compute.manager [instance: 0c1b8117-199f-4101-ab81-4ae7ef0d1251] return self.greenlet.switch() [ 626.316879] env[61923]: ERROR nova.compute.manager [instance: 0c1b8117-199f-4101-ab81-4ae7ef0d1251] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 626.316879] env[61923]: ERROR nova.compute.manager [instance: 0c1b8117-199f-4101-ab81-4ae7ef0d1251] result = function(*args, **kwargs) [ 626.316879] env[61923]: ERROR nova.compute.manager [instance: 0c1b8117-199f-4101-ab81-4ae7ef0d1251] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 626.316879] env[61923]: ERROR nova.compute.manager [instance: 0c1b8117-199f-4101-ab81-4ae7ef0d1251] return func(*args, **kwargs) [ 626.316879] env[61923]: ERROR nova.compute.manager [instance: 0c1b8117-199f-4101-ab81-4ae7ef0d1251] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 626.316879] env[61923]: ERROR nova.compute.manager [instance: 0c1b8117-199f-4101-ab81-4ae7ef0d1251] raise e [ 626.316879] env[61923]: ERROR nova.compute.manager [instance: 0c1b8117-199f-4101-ab81-4ae7ef0d1251] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 626.316879] env[61923]: ERROR nova.compute.manager [instance: 0c1b8117-199f-4101-ab81-4ae7ef0d1251] nwinfo = self.network_api.allocate_for_instance( [ 626.316879] env[61923]: ERROR nova.compute.manager [instance: 0c1b8117-199f-4101-ab81-4ae7ef0d1251] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 626.316879] env[61923]: ERROR nova.compute.manager [instance: 0c1b8117-199f-4101-ab81-4ae7ef0d1251] created_port_ids = self._update_ports_for_instance( [ 626.317218] env[61923]: ERROR nova.compute.manager [instance: 0c1b8117-199f-4101-ab81-4ae7ef0d1251] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 626.317218] env[61923]: ERROR nova.compute.manager [instance: 0c1b8117-199f-4101-ab81-4ae7ef0d1251] with excutils.save_and_reraise_exception(): [ 626.317218] env[61923]: ERROR nova.compute.manager [instance: 0c1b8117-199f-4101-ab81-4ae7ef0d1251] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 626.317218] env[61923]: ERROR nova.compute.manager [instance: 0c1b8117-199f-4101-ab81-4ae7ef0d1251] self.force_reraise() [ 626.317218] env[61923]: ERROR nova.compute.manager [instance: 0c1b8117-199f-4101-ab81-4ae7ef0d1251] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 626.317218] env[61923]: ERROR nova.compute.manager [instance: 0c1b8117-199f-4101-ab81-4ae7ef0d1251] raise self.value [ 626.317218] env[61923]: ERROR nova.compute.manager [instance: 0c1b8117-199f-4101-ab81-4ae7ef0d1251] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 626.317218] env[61923]: ERROR nova.compute.manager [instance: 0c1b8117-199f-4101-ab81-4ae7ef0d1251] updated_port = self._update_port( [ 626.317218] env[61923]: ERROR nova.compute.manager [instance: 0c1b8117-199f-4101-ab81-4ae7ef0d1251] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 626.317218] env[61923]: ERROR nova.compute.manager [instance: 0c1b8117-199f-4101-ab81-4ae7ef0d1251] _ensure_no_port_binding_failure(port) [ 626.317218] env[61923]: ERROR nova.compute.manager [instance: 0c1b8117-199f-4101-ab81-4ae7ef0d1251] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 626.317218] env[61923]: ERROR nova.compute.manager [instance: 0c1b8117-199f-4101-ab81-4ae7ef0d1251] raise exception.PortBindingFailed(port_id=port['id']) [ 626.317489] env[61923]: ERROR nova.compute.manager [instance: 0c1b8117-199f-4101-ab81-4ae7ef0d1251] nova.exception.PortBindingFailed: Binding failed for port 9ba992af-f024-4728-b3b6-b6cf6af4e319, please check neutron logs for more information. [ 626.317489] env[61923]: ERROR nova.compute.manager [instance: 0c1b8117-199f-4101-ab81-4ae7ef0d1251] [ 626.317489] env[61923]: INFO nova.compute.manager [None req-aa9c464b-ac6b-44fe-9384-28cd9bdfecbc tempest-ImagesOneServerTestJSON-63657725 tempest-ImagesOneServerTestJSON-63657725-project-member] [instance: 0c1b8117-199f-4101-ab81-4ae7ef0d1251] Terminating instance [ 626.317489] env[61923]: DEBUG oslo_concurrency.lockutils [None req-aa9c464b-ac6b-44fe-9384-28cd9bdfecbc tempest-ImagesOneServerTestJSON-63657725 tempest-ImagesOneServerTestJSON-63657725-project-member] Acquiring lock "refresh_cache-0c1b8117-199f-4101-ab81-4ae7ef0d1251" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 626.353630] env[61923]: DEBUG oslo_concurrency.lockutils [None req-4bd5c12b-d52a-4c54-93f7-87894f09013a tempest-ServersTestJSON-483293076 tempest-ServersTestJSON-483293076-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 626.495469] env[61923]: DEBUG nova.network.neutron [req-c0f2ceaf-a105-4a18-aa8f-88675b047967 req-1b6f47bc-6097-42c5-a20c-03cd0a551b12 service nova] [instance: 0c1b8117-199f-4101-ab81-4ae7ef0d1251] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 626.532241] env[61923]: DEBUG oslo_concurrency.lockutils [None req-0c55849d-766e-4fac-b1c4-e4e97ba688a6 tempest-DeleteServersAdminTestJSON-590116201 tempest-DeleteServersAdminTestJSON-590116201-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.484s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 626.534309] env[61923]: DEBUG nova.compute.manager [None req-0c55849d-766e-4fac-b1c4-e4e97ba688a6 tempest-DeleteServersAdminTestJSON-590116201 tempest-DeleteServersAdminTestJSON-590116201-project-member] [instance: 1b509288-2df3-4927-afcf-cb76918d3da3] Start building networks asynchronously for instance. {{(pid=61923) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 626.535377] env[61923]: DEBUG oslo_concurrency.lockutils [None req-f3aa5656-98df-4c46-9436-b968a8770dbc tempest-ListServerFiltersTestJSON-483030130 tempest-ListServerFiltersTestJSON-483030130-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.746s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 626.558027] env[61923]: DEBUG oslo_service.periodic_task [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61923) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 626.558027] env[61923]: DEBUG nova.compute.manager [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Starting heal instance info cache {{(pid=61923) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 626.558229] env[61923]: DEBUG nova.compute.manager [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Rebuilding the list of instances to heal {{(pid=61923) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 626.601551] env[61923]: DEBUG nova.network.neutron [req-c0f2ceaf-a105-4a18-aa8f-88675b047967 req-1b6f47bc-6097-42c5-a20c-03cd0a551b12 service nova] [instance: 0c1b8117-199f-4101-ab81-4ae7ef0d1251] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 627.040964] env[61923]: DEBUG nova.compute.utils [None req-0c55849d-766e-4fac-b1c4-e4e97ba688a6 tempest-DeleteServersAdminTestJSON-590116201 tempest-DeleteServersAdminTestJSON-590116201-project-member] Using /dev/sd instead of None {{(pid=61923) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 627.045985] env[61923]: DEBUG nova.compute.manager [None req-0c55849d-766e-4fac-b1c4-e4e97ba688a6 tempest-DeleteServersAdminTestJSON-590116201 tempest-DeleteServersAdminTestJSON-590116201-project-member] [instance: 1b509288-2df3-4927-afcf-cb76918d3da3] Allocating IP information in the background. {{(pid=61923) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 627.046180] env[61923]: DEBUG nova.network.neutron [None req-0c55849d-766e-4fac-b1c4-e4e97ba688a6 tempest-DeleteServersAdminTestJSON-590116201 tempest-DeleteServersAdminTestJSON-590116201-project-member] [instance: 1b509288-2df3-4927-afcf-cb76918d3da3] allocate_for_instance() {{(pid=61923) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 627.062290] env[61923]: DEBUG nova.compute.manager [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] [instance: 896dd689-4810-4f23-af0a-d2f557a0796b] Skipping network cache update for instance because it is Building. {{(pid=61923) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 627.062448] env[61923]: DEBUG nova.compute.manager [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] [instance: 119bb4a7-2d71-44d7-ae81-9238512937c8] Skipping network cache update for instance because it is Building. {{(pid=61923) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 627.062578] env[61923]: DEBUG nova.compute.manager [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] [instance: 0c1b8117-199f-4101-ab81-4ae7ef0d1251] Skipping network cache update for instance because it is Building. {{(pid=61923) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 627.062703] env[61923]: DEBUG nova.compute.manager [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] [instance: 1b509288-2df3-4927-afcf-cb76918d3da3] Skipping network cache update for instance because it is Building. {{(pid=61923) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 627.062824] env[61923]: DEBUG nova.compute.manager [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Didn't find any instances for network info cache update. {{(pid=61923) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10014}} [ 627.065785] env[61923]: DEBUG oslo_service.periodic_task [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61923) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 627.066390] env[61923]: DEBUG oslo_service.periodic_task [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61923) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 627.066539] env[61923]: DEBUG oslo_service.periodic_task [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61923) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 627.066690] env[61923]: DEBUG oslo_service.periodic_task [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61923) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 627.066854] env[61923]: DEBUG oslo_service.periodic_task [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61923) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 627.066965] env[61923]: DEBUG oslo_service.periodic_task [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61923) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 627.067163] env[61923]: DEBUG nova.compute.manager [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61923) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 627.067319] env[61923]: DEBUG oslo_service.periodic_task [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Running periodic task ComputeManager.update_available_resource {{(pid=61923) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 627.103852] env[61923]: DEBUG oslo_concurrency.lockutils [req-c0f2ceaf-a105-4a18-aa8f-88675b047967 req-1b6f47bc-6097-42c5-a20c-03cd0a551b12 service nova] Releasing lock "refresh_cache-0c1b8117-199f-4101-ab81-4ae7ef0d1251" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 627.104259] env[61923]: DEBUG oslo_concurrency.lockutils [None req-aa9c464b-ac6b-44fe-9384-28cd9bdfecbc tempest-ImagesOneServerTestJSON-63657725 tempest-ImagesOneServerTestJSON-63657725-project-member] Acquired lock "refresh_cache-0c1b8117-199f-4101-ab81-4ae7ef0d1251" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 627.104441] env[61923]: DEBUG nova.network.neutron [None req-aa9c464b-ac6b-44fe-9384-28cd9bdfecbc tempest-ImagesOneServerTestJSON-63657725 tempest-ImagesOneServerTestJSON-63657725-project-member] [instance: 0c1b8117-199f-4101-ab81-4ae7ef0d1251] Building network info cache for instance {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 627.160732] env[61923]: DEBUG nova.policy [None req-0c55849d-766e-4fac-b1c4-e4e97ba688a6 tempest-DeleteServersAdminTestJSON-590116201 tempest-DeleteServersAdminTestJSON-590116201-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '379d4ef2827847eeadcb2728bd5859f4', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8786415c53184755b01f324d83b72b9a', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61923) authorize /opt/stack/nova/nova/policy.py:201}} [ 627.389290] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af1e501e-68e4-4698-bac4-ee280c380014 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.403174] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74336389-38d8-45f0-a478-42bd9592bb0b {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.438641] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01444024-b113-477c-8b7d-43dcacee9189 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.447954] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bde664df-316f-47e1-b871-7711f3a7a1ab {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 627.461908] env[61923]: DEBUG nova.compute.provider_tree [None req-f3aa5656-98df-4c46-9436-b968a8770dbc tempest-ListServerFiltersTestJSON-483030130 tempest-ListServerFiltersTestJSON-483030130-project-member] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 627.523154] env[61923]: DEBUG nova.network.neutron [None req-0c55849d-766e-4fac-b1c4-e4e97ba688a6 tempest-DeleteServersAdminTestJSON-590116201 tempest-DeleteServersAdminTestJSON-590116201-project-member] [instance: 1b509288-2df3-4927-afcf-cb76918d3da3] Successfully created port: dbf4bfa8-cc4a-4be3-a70c-7f709e040750 {{(pid=61923) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 627.546389] env[61923]: DEBUG nova.compute.manager [None req-0c55849d-766e-4fac-b1c4-e4e97ba688a6 tempest-DeleteServersAdminTestJSON-590116201 tempest-DeleteServersAdminTestJSON-590116201-project-member] [instance: 1b509288-2df3-4927-afcf-cb76918d3da3] Start building block device mappings for instance. {{(pid=61923) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 627.574806] env[61923]: DEBUG oslo_concurrency.lockutils [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 627.636951] env[61923]: DEBUG nova.network.neutron [None req-aa9c464b-ac6b-44fe-9384-28cd9bdfecbc tempest-ImagesOneServerTestJSON-63657725 tempest-ImagesOneServerTestJSON-63657725-project-member] [instance: 0c1b8117-199f-4101-ab81-4ae7ef0d1251] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 627.828736] env[61923]: DEBUG nova.network.neutron [None req-aa9c464b-ac6b-44fe-9384-28cd9bdfecbc tempest-ImagesOneServerTestJSON-63657725 tempest-ImagesOneServerTestJSON-63657725-project-member] [instance: 0c1b8117-199f-4101-ab81-4ae7ef0d1251] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 627.838945] env[61923]: DEBUG oslo_concurrency.lockutils [None req-5bb1ca17-5442-4564-86ab-53f866a97e1d tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Acquiring lock "15eaaa7c-77be-4b8b-89d0-0b2b11a0c49e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 627.839359] env[61923]: DEBUG oslo_concurrency.lockutils [None req-5bb1ca17-5442-4564-86ab-53f866a97e1d tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Lock "15eaaa7c-77be-4b8b-89d0-0b2b11a0c49e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 627.966015] env[61923]: DEBUG nova.scheduler.client.report [None req-f3aa5656-98df-4c46-9436-b968a8770dbc tempest-ListServerFiltersTestJSON-483030130 tempest-ListServerFiltersTestJSON-483030130-project-member] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 628.042193] env[61923]: DEBUG nova.compute.manager [req-1368bbce-e220-47e5-8532-6b52a77e8cbd req-85996647-913e-416e-9980-e7bbf26e2118 service nova] [instance: 0c1b8117-199f-4101-ab81-4ae7ef0d1251] Received event network-vif-deleted-9ba992af-f024-4728-b3b6-b6cf6af4e319 {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 628.331778] env[61923]: DEBUG oslo_concurrency.lockutils [None req-aa9c464b-ac6b-44fe-9384-28cd9bdfecbc tempest-ImagesOneServerTestJSON-63657725 tempest-ImagesOneServerTestJSON-63657725-project-member] Releasing lock "refresh_cache-0c1b8117-199f-4101-ab81-4ae7ef0d1251" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 628.332228] env[61923]: DEBUG nova.compute.manager [None req-aa9c464b-ac6b-44fe-9384-28cd9bdfecbc tempest-ImagesOneServerTestJSON-63657725 tempest-ImagesOneServerTestJSON-63657725-project-member] [instance: 0c1b8117-199f-4101-ab81-4ae7ef0d1251] Start destroying the instance on the hypervisor. {{(pid=61923) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 628.332683] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-aa9c464b-ac6b-44fe-9384-28cd9bdfecbc tempest-ImagesOneServerTestJSON-63657725 tempest-ImagesOneServerTestJSON-63657725-project-member] [instance: 0c1b8117-199f-4101-ab81-4ae7ef0d1251] Destroying instance {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 628.332769] env[61923]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-5dfb0aa7-3e2c-45af-b9eb-c66e818cbb18 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.341244] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a78f0580-8f4e-4973-8f57-f81e87fda030 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.363872] env[61923]: WARNING nova.virt.vmwareapi.vmops [None req-aa9c464b-ac6b-44fe-9384-28cd9bdfecbc tempest-ImagesOneServerTestJSON-63657725 tempest-ImagesOneServerTestJSON-63657725-project-member] [instance: 0c1b8117-199f-4101-ab81-4ae7ef0d1251] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 0c1b8117-199f-4101-ab81-4ae7ef0d1251 could not be found. [ 628.364107] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-aa9c464b-ac6b-44fe-9384-28cd9bdfecbc tempest-ImagesOneServerTestJSON-63657725 tempest-ImagesOneServerTestJSON-63657725-project-member] [instance: 0c1b8117-199f-4101-ab81-4ae7ef0d1251] Instance destroyed {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 628.364291] env[61923]: INFO nova.compute.manager [None req-aa9c464b-ac6b-44fe-9384-28cd9bdfecbc tempest-ImagesOneServerTestJSON-63657725 tempest-ImagesOneServerTestJSON-63657725-project-member] [instance: 0c1b8117-199f-4101-ab81-4ae7ef0d1251] Took 0.03 seconds to destroy the instance on the hypervisor. [ 628.364529] env[61923]: DEBUG oslo.service.loopingcall [None req-aa9c464b-ac6b-44fe-9384-28cd9bdfecbc tempest-ImagesOneServerTestJSON-63657725 tempest-ImagesOneServerTestJSON-63657725-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61923) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 628.364757] env[61923]: DEBUG nova.compute.manager [-] [instance: 0c1b8117-199f-4101-ab81-4ae7ef0d1251] Deallocating network for instance {{(pid=61923) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 628.364845] env[61923]: DEBUG nova.network.neutron [-] [instance: 0c1b8117-199f-4101-ab81-4ae7ef0d1251] deallocate_for_instance() {{(pid=61923) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 628.390019] env[61923]: DEBUG nova.network.neutron [-] [instance: 0c1b8117-199f-4101-ab81-4ae7ef0d1251] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 628.473923] env[61923]: DEBUG oslo_concurrency.lockutils [None req-f3aa5656-98df-4c46-9436-b968a8770dbc tempest-ListServerFiltersTestJSON-483030130 tempest-ListServerFiltersTestJSON-483030130-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.938s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 628.474568] env[61923]: ERROR nova.compute.manager [None req-f3aa5656-98df-4c46-9436-b968a8770dbc tempest-ListServerFiltersTestJSON-483030130 tempest-ListServerFiltersTestJSON-483030130-project-member] [instance: 896dd689-4810-4f23-af0a-d2f557a0796b] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 53407134-6b24-4ff7-b5d0-ef4cf1885933, please check neutron logs for more information. [ 628.474568] env[61923]: ERROR nova.compute.manager [instance: 896dd689-4810-4f23-af0a-d2f557a0796b] Traceback (most recent call last): [ 628.474568] env[61923]: ERROR nova.compute.manager [instance: 896dd689-4810-4f23-af0a-d2f557a0796b] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 628.474568] env[61923]: ERROR nova.compute.manager [instance: 896dd689-4810-4f23-af0a-d2f557a0796b] self.driver.spawn(context, instance, image_meta, [ 628.474568] env[61923]: ERROR nova.compute.manager [instance: 896dd689-4810-4f23-af0a-d2f557a0796b] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 628.474568] env[61923]: ERROR nova.compute.manager [instance: 896dd689-4810-4f23-af0a-d2f557a0796b] self._vmops.spawn(context, instance, image_meta, injected_files, [ 628.474568] env[61923]: ERROR nova.compute.manager [instance: 896dd689-4810-4f23-af0a-d2f557a0796b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 628.474568] env[61923]: ERROR nova.compute.manager [instance: 896dd689-4810-4f23-af0a-d2f557a0796b] vm_ref = self.build_virtual_machine(instance, [ 628.474568] env[61923]: ERROR nova.compute.manager [instance: 896dd689-4810-4f23-af0a-d2f557a0796b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 628.474568] env[61923]: ERROR nova.compute.manager [instance: 896dd689-4810-4f23-af0a-d2f557a0796b] vif_infos = vmwarevif.get_vif_info(self._session, [ 628.474568] env[61923]: ERROR nova.compute.manager [instance: 896dd689-4810-4f23-af0a-d2f557a0796b] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 628.474913] env[61923]: ERROR nova.compute.manager [instance: 896dd689-4810-4f23-af0a-d2f557a0796b] for vif in network_info: [ 628.474913] env[61923]: ERROR nova.compute.manager [instance: 896dd689-4810-4f23-af0a-d2f557a0796b] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 628.474913] env[61923]: ERROR nova.compute.manager [instance: 896dd689-4810-4f23-af0a-d2f557a0796b] return self._sync_wrapper(fn, *args, **kwargs) [ 628.474913] env[61923]: ERROR nova.compute.manager [instance: 896dd689-4810-4f23-af0a-d2f557a0796b] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 628.474913] env[61923]: ERROR nova.compute.manager [instance: 896dd689-4810-4f23-af0a-d2f557a0796b] self.wait() [ 628.474913] env[61923]: ERROR nova.compute.manager [instance: 896dd689-4810-4f23-af0a-d2f557a0796b] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 628.474913] env[61923]: ERROR nova.compute.manager [instance: 896dd689-4810-4f23-af0a-d2f557a0796b] self[:] = self._gt.wait() [ 628.474913] env[61923]: ERROR nova.compute.manager [instance: 896dd689-4810-4f23-af0a-d2f557a0796b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 628.474913] env[61923]: ERROR nova.compute.manager [instance: 896dd689-4810-4f23-af0a-d2f557a0796b] return self._exit_event.wait() [ 628.474913] env[61923]: ERROR nova.compute.manager [instance: 896dd689-4810-4f23-af0a-d2f557a0796b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 628.474913] env[61923]: ERROR nova.compute.manager [instance: 896dd689-4810-4f23-af0a-d2f557a0796b] result = hub.switch() [ 628.474913] env[61923]: ERROR nova.compute.manager [instance: 896dd689-4810-4f23-af0a-d2f557a0796b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 628.474913] env[61923]: ERROR nova.compute.manager [instance: 896dd689-4810-4f23-af0a-d2f557a0796b] return self.greenlet.switch() [ 628.475326] env[61923]: ERROR nova.compute.manager [instance: 896dd689-4810-4f23-af0a-d2f557a0796b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 628.475326] env[61923]: ERROR nova.compute.manager [instance: 896dd689-4810-4f23-af0a-d2f557a0796b] result = function(*args, **kwargs) [ 628.475326] env[61923]: ERROR nova.compute.manager [instance: 896dd689-4810-4f23-af0a-d2f557a0796b] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 628.475326] env[61923]: ERROR nova.compute.manager [instance: 896dd689-4810-4f23-af0a-d2f557a0796b] return func(*args, **kwargs) [ 628.475326] env[61923]: ERROR nova.compute.manager [instance: 896dd689-4810-4f23-af0a-d2f557a0796b] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 628.475326] env[61923]: ERROR nova.compute.manager [instance: 896dd689-4810-4f23-af0a-d2f557a0796b] raise e [ 628.475326] env[61923]: ERROR nova.compute.manager [instance: 896dd689-4810-4f23-af0a-d2f557a0796b] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 628.475326] env[61923]: ERROR nova.compute.manager [instance: 896dd689-4810-4f23-af0a-d2f557a0796b] nwinfo = self.network_api.allocate_for_instance( [ 628.475326] env[61923]: ERROR nova.compute.manager [instance: 896dd689-4810-4f23-af0a-d2f557a0796b] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 628.475326] env[61923]: ERROR nova.compute.manager [instance: 896dd689-4810-4f23-af0a-d2f557a0796b] created_port_ids = self._update_ports_for_instance( [ 628.475326] env[61923]: ERROR nova.compute.manager [instance: 896dd689-4810-4f23-af0a-d2f557a0796b] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 628.475326] env[61923]: ERROR nova.compute.manager [instance: 896dd689-4810-4f23-af0a-d2f557a0796b] with excutils.save_and_reraise_exception(): [ 628.475326] env[61923]: ERROR nova.compute.manager [instance: 896dd689-4810-4f23-af0a-d2f557a0796b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 628.475747] env[61923]: ERROR nova.compute.manager [instance: 896dd689-4810-4f23-af0a-d2f557a0796b] self.force_reraise() [ 628.475747] env[61923]: ERROR nova.compute.manager [instance: 896dd689-4810-4f23-af0a-d2f557a0796b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 628.475747] env[61923]: ERROR nova.compute.manager [instance: 896dd689-4810-4f23-af0a-d2f557a0796b] raise self.value [ 628.475747] env[61923]: ERROR nova.compute.manager [instance: 896dd689-4810-4f23-af0a-d2f557a0796b] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 628.475747] env[61923]: ERROR nova.compute.manager [instance: 896dd689-4810-4f23-af0a-d2f557a0796b] updated_port = self._update_port( [ 628.475747] env[61923]: ERROR nova.compute.manager [instance: 896dd689-4810-4f23-af0a-d2f557a0796b] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 628.475747] env[61923]: ERROR nova.compute.manager [instance: 896dd689-4810-4f23-af0a-d2f557a0796b] _ensure_no_port_binding_failure(port) [ 628.475747] env[61923]: ERROR nova.compute.manager [instance: 896dd689-4810-4f23-af0a-d2f557a0796b] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 628.475747] env[61923]: ERROR nova.compute.manager [instance: 896dd689-4810-4f23-af0a-d2f557a0796b] raise exception.PortBindingFailed(port_id=port['id']) [ 628.475747] env[61923]: ERROR nova.compute.manager [instance: 896dd689-4810-4f23-af0a-d2f557a0796b] nova.exception.PortBindingFailed: Binding failed for port 53407134-6b24-4ff7-b5d0-ef4cf1885933, please check neutron logs for more information. [ 628.475747] env[61923]: ERROR nova.compute.manager [instance: 896dd689-4810-4f23-af0a-d2f557a0796b] [ 628.476034] env[61923]: DEBUG nova.compute.utils [None req-f3aa5656-98df-4c46-9436-b968a8770dbc tempest-ListServerFiltersTestJSON-483030130 tempest-ListServerFiltersTestJSON-483030130-project-member] [instance: 896dd689-4810-4f23-af0a-d2f557a0796b] Binding failed for port 53407134-6b24-4ff7-b5d0-ef4cf1885933, please check neutron logs for more information. {{(pid=61923) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 628.479951] env[61923]: DEBUG oslo_concurrency.lockutils [None req-95b9b70a-a835-4901-9a6d-c0fa1b2217f3 tempest-MigrationsAdminTest-1311988845 tempest-MigrationsAdminTest-1311988845-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.427s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 628.481767] env[61923]: INFO nova.compute.claims [None req-95b9b70a-a835-4901-9a6d-c0fa1b2217f3 tempest-MigrationsAdminTest-1311988845 tempest-MigrationsAdminTest-1311988845-project-member] [instance: a4b54a7a-3f3d-49ae-ba43-5a10c262d4e5] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 628.485571] env[61923]: DEBUG nova.compute.manager [None req-f3aa5656-98df-4c46-9436-b968a8770dbc tempest-ListServerFiltersTestJSON-483030130 tempest-ListServerFiltersTestJSON-483030130-project-member] [instance: 896dd689-4810-4f23-af0a-d2f557a0796b] Build of instance 896dd689-4810-4f23-af0a-d2f557a0796b was re-scheduled: Binding failed for port 53407134-6b24-4ff7-b5d0-ef4cf1885933, please check neutron logs for more information. {{(pid=61923) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 628.486009] env[61923]: DEBUG nova.compute.manager [None req-f3aa5656-98df-4c46-9436-b968a8770dbc tempest-ListServerFiltersTestJSON-483030130 tempest-ListServerFiltersTestJSON-483030130-project-member] [instance: 896dd689-4810-4f23-af0a-d2f557a0796b] Unplugging VIFs for instance {{(pid=61923) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 628.486295] env[61923]: DEBUG oslo_concurrency.lockutils [None req-f3aa5656-98df-4c46-9436-b968a8770dbc tempest-ListServerFiltersTestJSON-483030130 tempest-ListServerFiltersTestJSON-483030130-project-member] Acquiring lock "refresh_cache-896dd689-4810-4f23-af0a-d2f557a0796b" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 628.486384] env[61923]: DEBUG oslo_concurrency.lockutils [None req-f3aa5656-98df-4c46-9436-b968a8770dbc tempest-ListServerFiltersTestJSON-483030130 tempest-ListServerFiltersTestJSON-483030130-project-member] Acquired lock "refresh_cache-896dd689-4810-4f23-af0a-d2f557a0796b" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 628.486537] env[61923]: DEBUG nova.network.neutron [None req-f3aa5656-98df-4c46-9436-b968a8770dbc tempest-ListServerFiltersTestJSON-483030130 tempest-ListServerFiltersTestJSON-483030130-project-member] [instance: 896dd689-4810-4f23-af0a-d2f557a0796b] Building network info cache for instance {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 628.555362] env[61923]: DEBUG nova.compute.manager [None req-0c55849d-766e-4fac-b1c4-e4e97ba688a6 tempest-DeleteServersAdminTestJSON-590116201 tempest-DeleteServersAdminTestJSON-590116201-project-member] [instance: 1b509288-2df3-4927-afcf-cb76918d3da3] Start spawning the instance on the hypervisor. {{(pid=61923) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 628.582452] env[61923]: DEBUG nova.virt.hardware [None req-0c55849d-766e-4fac-b1c4-e4e97ba688a6 tempest-DeleteServersAdminTestJSON-590116201 tempest-DeleteServersAdminTestJSON-590116201-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-29T20:07:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-29T20:07:35Z,direct_url=,disk_format='vmdk',id=0f153f63-ae0a-45b1-b7f5-7f9b673c947f,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='4e7b5e3b3c3443c8bd5c70f8a15739f5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-29T20:07:36Z,virtual_size=,visibility=), allow threads: False {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 628.582735] env[61923]: DEBUG nova.virt.hardware [None req-0c55849d-766e-4fac-b1c4-e4e97ba688a6 tempest-DeleteServersAdminTestJSON-590116201 tempest-DeleteServersAdminTestJSON-590116201-project-member] Flavor limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 628.582890] env[61923]: DEBUG nova.virt.hardware [None req-0c55849d-766e-4fac-b1c4-e4e97ba688a6 tempest-DeleteServersAdminTestJSON-590116201 tempest-DeleteServersAdminTestJSON-590116201-project-member] Image limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 628.583076] env[61923]: DEBUG nova.virt.hardware [None req-0c55849d-766e-4fac-b1c4-e4e97ba688a6 tempest-DeleteServersAdminTestJSON-590116201 tempest-DeleteServersAdminTestJSON-590116201-project-member] Flavor pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 628.583228] env[61923]: DEBUG nova.virt.hardware [None req-0c55849d-766e-4fac-b1c4-e4e97ba688a6 tempest-DeleteServersAdminTestJSON-590116201 tempest-DeleteServersAdminTestJSON-590116201-project-member] Image pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 628.583370] env[61923]: DEBUG nova.virt.hardware [None req-0c55849d-766e-4fac-b1c4-e4e97ba688a6 tempest-DeleteServersAdminTestJSON-590116201 tempest-DeleteServersAdminTestJSON-590116201-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 628.583578] env[61923]: DEBUG nova.virt.hardware [None req-0c55849d-766e-4fac-b1c4-e4e97ba688a6 tempest-DeleteServersAdminTestJSON-590116201 tempest-DeleteServersAdminTestJSON-590116201-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 628.583844] env[61923]: DEBUG nova.virt.hardware [None req-0c55849d-766e-4fac-b1c4-e4e97ba688a6 tempest-DeleteServersAdminTestJSON-590116201 tempest-DeleteServersAdminTestJSON-590116201-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 628.584064] env[61923]: DEBUG nova.virt.hardware [None req-0c55849d-766e-4fac-b1c4-e4e97ba688a6 tempest-DeleteServersAdminTestJSON-590116201 tempest-DeleteServersAdminTestJSON-590116201-project-member] Got 1 possible topologies {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 628.584267] env[61923]: DEBUG nova.virt.hardware [None req-0c55849d-766e-4fac-b1c4-e4e97ba688a6 tempest-DeleteServersAdminTestJSON-590116201 tempest-DeleteServersAdminTestJSON-590116201-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 628.584459] env[61923]: DEBUG nova.virt.hardware [None req-0c55849d-766e-4fac-b1c4-e4e97ba688a6 tempest-DeleteServersAdminTestJSON-590116201 tempest-DeleteServersAdminTestJSON-590116201-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 628.585801] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1338a7a-40f1-49be-8322-70aaf5f26d20 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.593842] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7b9d5eb-8e72-4aee-a25d-bc06bd45d5e6 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.745993] env[61923]: ERROR nova.compute.manager [None req-0c55849d-766e-4fac-b1c4-e4e97ba688a6 tempest-DeleteServersAdminTestJSON-590116201 tempest-DeleteServersAdminTestJSON-590116201-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port dbf4bfa8-cc4a-4be3-a70c-7f709e040750, please check neutron logs for more information. [ 628.745993] env[61923]: ERROR nova.compute.manager Traceback (most recent call last): [ 628.745993] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 628.745993] env[61923]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 628.745993] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 628.745993] env[61923]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 628.745993] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 628.745993] env[61923]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 628.745993] env[61923]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 628.745993] env[61923]: ERROR nova.compute.manager self.force_reraise() [ 628.745993] env[61923]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 628.745993] env[61923]: ERROR nova.compute.manager raise self.value [ 628.745993] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 628.745993] env[61923]: ERROR nova.compute.manager updated_port = self._update_port( [ 628.745993] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 628.745993] env[61923]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 628.746578] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 628.746578] env[61923]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 628.746578] env[61923]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port dbf4bfa8-cc4a-4be3-a70c-7f709e040750, please check neutron logs for more information. [ 628.746578] env[61923]: ERROR nova.compute.manager [ 628.746578] env[61923]: Traceback (most recent call last): [ 628.746578] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 628.746578] env[61923]: listener.cb(fileno) [ 628.746578] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 628.746578] env[61923]: result = function(*args, **kwargs) [ 628.746578] env[61923]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 628.746578] env[61923]: return func(*args, **kwargs) [ 628.746578] env[61923]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 628.746578] env[61923]: raise e [ 628.746578] env[61923]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 628.746578] env[61923]: nwinfo = self.network_api.allocate_for_instance( [ 628.746578] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 628.746578] env[61923]: created_port_ids = self._update_ports_for_instance( [ 628.746578] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 628.746578] env[61923]: with excutils.save_and_reraise_exception(): [ 628.746578] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 628.746578] env[61923]: self.force_reraise() [ 628.746578] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 628.746578] env[61923]: raise self.value [ 628.746578] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 628.746578] env[61923]: updated_port = self._update_port( [ 628.746578] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 628.746578] env[61923]: _ensure_no_port_binding_failure(port) [ 628.746578] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 628.746578] env[61923]: raise exception.PortBindingFailed(port_id=port['id']) [ 628.747465] env[61923]: nova.exception.PortBindingFailed: Binding failed for port dbf4bfa8-cc4a-4be3-a70c-7f709e040750, please check neutron logs for more information. [ 628.747465] env[61923]: Removing descriptor: 17 [ 628.747465] env[61923]: ERROR nova.compute.manager [None req-0c55849d-766e-4fac-b1c4-e4e97ba688a6 tempest-DeleteServersAdminTestJSON-590116201 tempest-DeleteServersAdminTestJSON-590116201-project-member] [instance: 1b509288-2df3-4927-afcf-cb76918d3da3] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port dbf4bfa8-cc4a-4be3-a70c-7f709e040750, please check neutron logs for more information. [ 628.747465] env[61923]: ERROR nova.compute.manager [instance: 1b509288-2df3-4927-afcf-cb76918d3da3] Traceback (most recent call last): [ 628.747465] env[61923]: ERROR nova.compute.manager [instance: 1b509288-2df3-4927-afcf-cb76918d3da3] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 628.747465] env[61923]: ERROR nova.compute.manager [instance: 1b509288-2df3-4927-afcf-cb76918d3da3] yield resources [ 628.747465] env[61923]: ERROR nova.compute.manager [instance: 1b509288-2df3-4927-afcf-cb76918d3da3] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 628.747465] env[61923]: ERROR nova.compute.manager [instance: 1b509288-2df3-4927-afcf-cb76918d3da3] self.driver.spawn(context, instance, image_meta, [ 628.747465] env[61923]: ERROR nova.compute.manager [instance: 1b509288-2df3-4927-afcf-cb76918d3da3] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 628.747465] env[61923]: ERROR nova.compute.manager [instance: 1b509288-2df3-4927-afcf-cb76918d3da3] self._vmops.spawn(context, instance, image_meta, injected_files, [ 628.747465] env[61923]: ERROR nova.compute.manager [instance: 1b509288-2df3-4927-afcf-cb76918d3da3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 628.747465] env[61923]: ERROR nova.compute.manager [instance: 1b509288-2df3-4927-afcf-cb76918d3da3] vm_ref = self.build_virtual_machine(instance, [ 628.747816] env[61923]: ERROR nova.compute.manager [instance: 1b509288-2df3-4927-afcf-cb76918d3da3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 628.747816] env[61923]: ERROR nova.compute.manager [instance: 1b509288-2df3-4927-afcf-cb76918d3da3] vif_infos = vmwarevif.get_vif_info(self._session, [ 628.747816] env[61923]: ERROR nova.compute.manager [instance: 1b509288-2df3-4927-afcf-cb76918d3da3] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 628.747816] env[61923]: ERROR nova.compute.manager [instance: 1b509288-2df3-4927-afcf-cb76918d3da3] for vif in network_info: [ 628.747816] env[61923]: ERROR nova.compute.manager [instance: 1b509288-2df3-4927-afcf-cb76918d3da3] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 628.747816] env[61923]: ERROR nova.compute.manager [instance: 1b509288-2df3-4927-afcf-cb76918d3da3] return self._sync_wrapper(fn, *args, **kwargs) [ 628.747816] env[61923]: ERROR nova.compute.manager [instance: 1b509288-2df3-4927-afcf-cb76918d3da3] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 628.747816] env[61923]: ERROR nova.compute.manager [instance: 1b509288-2df3-4927-afcf-cb76918d3da3] self.wait() [ 628.747816] env[61923]: ERROR nova.compute.manager [instance: 1b509288-2df3-4927-afcf-cb76918d3da3] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 628.747816] env[61923]: ERROR nova.compute.manager [instance: 1b509288-2df3-4927-afcf-cb76918d3da3] self[:] = self._gt.wait() [ 628.747816] env[61923]: ERROR nova.compute.manager [instance: 1b509288-2df3-4927-afcf-cb76918d3da3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 628.747816] env[61923]: ERROR nova.compute.manager [instance: 1b509288-2df3-4927-afcf-cb76918d3da3] return self._exit_event.wait() [ 628.747816] env[61923]: ERROR nova.compute.manager [instance: 1b509288-2df3-4927-afcf-cb76918d3da3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 628.748195] env[61923]: ERROR nova.compute.manager [instance: 1b509288-2df3-4927-afcf-cb76918d3da3] result = hub.switch() [ 628.748195] env[61923]: ERROR nova.compute.manager [instance: 1b509288-2df3-4927-afcf-cb76918d3da3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 628.748195] env[61923]: ERROR nova.compute.manager [instance: 1b509288-2df3-4927-afcf-cb76918d3da3] return self.greenlet.switch() [ 628.748195] env[61923]: ERROR nova.compute.manager [instance: 1b509288-2df3-4927-afcf-cb76918d3da3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 628.748195] env[61923]: ERROR nova.compute.manager [instance: 1b509288-2df3-4927-afcf-cb76918d3da3] result = function(*args, **kwargs) [ 628.748195] env[61923]: ERROR nova.compute.manager [instance: 1b509288-2df3-4927-afcf-cb76918d3da3] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 628.748195] env[61923]: ERROR nova.compute.manager [instance: 1b509288-2df3-4927-afcf-cb76918d3da3] return func(*args, **kwargs) [ 628.748195] env[61923]: ERROR nova.compute.manager [instance: 1b509288-2df3-4927-afcf-cb76918d3da3] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 628.748195] env[61923]: ERROR nova.compute.manager [instance: 1b509288-2df3-4927-afcf-cb76918d3da3] raise e [ 628.748195] env[61923]: ERROR nova.compute.manager [instance: 1b509288-2df3-4927-afcf-cb76918d3da3] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 628.748195] env[61923]: ERROR nova.compute.manager [instance: 1b509288-2df3-4927-afcf-cb76918d3da3] nwinfo = self.network_api.allocate_for_instance( [ 628.748195] env[61923]: ERROR nova.compute.manager [instance: 1b509288-2df3-4927-afcf-cb76918d3da3] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 628.748195] env[61923]: ERROR nova.compute.manager [instance: 1b509288-2df3-4927-afcf-cb76918d3da3] created_port_ids = self._update_ports_for_instance( [ 628.748581] env[61923]: ERROR nova.compute.manager [instance: 1b509288-2df3-4927-afcf-cb76918d3da3] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 628.748581] env[61923]: ERROR nova.compute.manager [instance: 1b509288-2df3-4927-afcf-cb76918d3da3] with excutils.save_and_reraise_exception(): [ 628.748581] env[61923]: ERROR nova.compute.manager [instance: 1b509288-2df3-4927-afcf-cb76918d3da3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 628.748581] env[61923]: ERROR nova.compute.manager [instance: 1b509288-2df3-4927-afcf-cb76918d3da3] self.force_reraise() [ 628.748581] env[61923]: ERROR nova.compute.manager [instance: 1b509288-2df3-4927-afcf-cb76918d3da3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 628.748581] env[61923]: ERROR nova.compute.manager [instance: 1b509288-2df3-4927-afcf-cb76918d3da3] raise self.value [ 628.748581] env[61923]: ERROR nova.compute.manager [instance: 1b509288-2df3-4927-afcf-cb76918d3da3] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 628.748581] env[61923]: ERROR nova.compute.manager [instance: 1b509288-2df3-4927-afcf-cb76918d3da3] updated_port = self._update_port( [ 628.748581] env[61923]: ERROR nova.compute.manager [instance: 1b509288-2df3-4927-afcf-cb76918d3da3] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 628.748581] env[61923]: ERROR nova.compute.manager [instance: 1b509288-2df3-4927-afcf-cb76918d3da3] _ensure_no_port_binding_failure(port) [ 628.748581] env[61923]: ERROR nova.compute.manager [instance: 1b509288-2df3-4927-afcf-cb76918d3da3] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 628.748581] env[61923]: ERROR nova.compute.manager [instance: 1b509288-2df3-4927-afcf-cb76918d3da3] raise exception.PortBindingFailed(port_id=port['id']) [ 628.748947] env[61923]: ERROR nova.compute.manager [instance: 1b509288-2df3-4927-afcf-cb76918d3da3] nova.exception.PortBindingFailed: Binding failed for port dbf4bfa8-cc4a-4be3-a70c-7f709e040750, please check neutron logs for more information. [ 628.748947] env[61923]: ERROR nova.compute.manager [instance: 1b509288-2df3-4927-afcf-cb76918d3da3] [ 628.748947] env[61923]: INFO nova.compute.manager [None req-0c55849d-766e-4fac-b1c4-e4e97ba688a6 tempest-DeleteServersAdminTestJSON-590116201 tempest-DeleteServersAdminTestJSON-590116201-project-member] [instance: 1b509288-2df3-4927-afcf-cb76918d3da3] Terminating instance [ 628.749602] env[61923]: DEBUG oslo_concurrency.lockutils [None req-0c55849d-766e-4fac-b1c4-e4e97ba688a6 tempest-DeleteServersAdminTestJSON-590116201 tempest-DeleteServersAdminTestJSON-590116201-project-member] Acquiring lock "refresh_cache-1b509288-2df3-4927-afcf-cb76918d3da3" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 628.749762] env[61923]: DEBUG oslo_concurrency.lockutils [None req-0c55849d-766e-4fac-b1c4-e4e97ba688a6 tempest-DeleteServersAdminTestJSON-590116201 tempest-DeleteServersAdminTestJSON-590116201-project-member] Acquired lock "refresh_cache-1b509288-2df3-4927-afcf-cb76918d3da3" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 628.749926] env[61923]: DEBUG nova.network.neutron [None req-0c55849d-766e-4fac-b1c4-e4e97ba688a6 tempest-DeleteServersAdminTestJSON-590116201 tempest-DeleteServersAdminTestJSON-590116201-project-member] [instance: 1b509288-2df3-4927-afcf-cb76918d3da3] Building network info cache for instance {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 628.892703] env[61923]: DEBUG nova.network.neutron [-] [instance: 0c1b8117-199f-4101-ab81-4ae7ef0d1251] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 629.005419] env[61923]: DEBUG nova.network.neutron [None req-f3aa5656-98df-4c46-9436-b968a8770dbc tempest-ListServerFiltersTestJSON-483030130 tempest-ListServerFiltersTestJSON-483030130-project-member] [instance: 896dd689-4810-4f23-af0a-d2f557a0796b] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 629.113988] env[61923]: DEBUG nova.network.neutron [None req-f3aa5656-98df-4c46-9436-b968a8770dbc tempest-ListServerFiltersTestJSON-483030130 tempest-ListServerFiltersTestJSON-483030130-project-member] [instance: 896dd689-4810-4f23-af0a-d2f557a0796b] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 629.265663] env[61923]: DEBUG nova.network.neutron [None req-0c55849d-766e-4fac-b1c4-e4e97ba688a6 tempest-DeleteServersAdminTestJSON-590116201 tempest-DeleteServersAdminTestJSON-590116201-project-member] [instance: 1b509288-2df3-4927-afcf-cb76918d3da3] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 629.307562] env[61923]: DEBUG nova.network.neutron [None req-0c55849d-766e-4fac-b1c4-e4e97ba688a6 tempest-DeleteServersAdminTestJSON-590116201 tempest-DeleteServersAdminTestJSON-590116201-project-member] [instance: 1b509288-2df3-4927-afcf-cb76918d3da3] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 629.396364] env[61923]: INFO nova.compute.manager [-] [instance: 0c1b8117-199f-4101-ab81-4ae7ef0d1251] Took 1.03 seconds to deallocate network for instance. [ 629.397699] env[61923]: DEBUG nova.compute.claims [None req-aa9c464b-ac6b-44fe-9384-28cd9bdfecbc tempest-ImagesOneServerTestJSON-63657725 tempest-ImagesOneServerTestJSON-63657725-project-member] [instance: 0c1b8117-199f-4101-ab81-4ae7ef0d1251] Aborting claim: {{(pid=61923) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 629.397874] env[61923]: DEBUG oslo_concurrency.lockutils [None req-aa9c464b-ac6b-44fe-9384-28cd9bdfecbc tempest-ImagesOneServerTestJSON-63657725 tempest-ImagesOneServerTestJSON-63657725-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 629.616556] env[61923]: DEBUG oslo_concurrency.lockutils [None req-f3aa5656-98df-4c46-9436-b968a8770dbc tempest-ListServerFiltersTestJSON-483030130 tempest-ListServerFiltersTestJSON-483030130-project-member] Releasing lock "refresh_cache-896dd689-4810-4f23-af0a-d2f557a0796b" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 629.616850] env[61923]: DEBUG nova.compute.manager [None req-f3aa5656-98df-4c46-9436-b968a8770dbc tempest-ListServerFiltersTestJSON-483030130 tempest-ListServerFiltersTestJSON-483030130-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61923) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 629.617148] env[61923]: DEBUG nova.compute.manager [None req-f3aa5656-98df-4c46-9436-b968a8770dbc tempest-ListServerFiltersTestJSON-483030130 tempest-ListServerFiltersTestJSON-483030130-project-member] [instance: 896dd689-4810-4f23-af0a-d2f557a0796b] Deallocating network for instance {{(pid=61923) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 629.617368] env[61923]: DEBUG nova.network.neutron [None req-f3aa5656-98df-4c46-9436-b968a8770dbc tempest-ListServerFiltersTestJSON-483030130 tempest-ListServerFiltersTestJSON-483030130-project-member] [instance: 896dd689-4810-4f23-af0a-d2f557a0796b] deallocate_for_instance() {{(pid=61923) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 629.640761] env[61923]: DEBUG nova.network.neutron [None req-f3aa5656-98df-4c46-9436-b968a8770dbc tempest-ListServerFiltersTestJSON-483030130 tempest-ListServerFiltersTestJSON-483030130-project-member] [instance: 896dd689-4810-4f23-af0a-d2f557a0796b] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 629.782627] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ade80547-7443-4de9-a55f-18b0d9109b79 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.789654] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63ec2a62-9afd-4a0e-b567-b000417100b3 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.818806] env[61923]: DEBUG oslo_concurrency.lockutils [None req-0c55849d-766e-4fac-b1c4-e4e97ba688a6 tempest-DeleteServersAdminTestJSON-590116201 tempest-DeleteServersAdminTestJSON-590116201-project-member] Releasing lock "refresh_cache-1b509288-2df3-4927-afcf-cb76918d3da3" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 629.819191] env[61923]: DEBUG nova.compute.manager [None req-0c55849d-766e-4fac-b1c4-e4e97ba688a6 tempest-DeleteServersAdminTestJSON-590116201 tempest-DeleteServersAdminTestJSON-590116201-project-member] [instance: 1b509288-2df3-4927-afcf-cb76918d3da3] Start destroying the instance on the hypervisor. {{(pid=61923) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 629.819396] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-0c55849d-766e-4fac-b1c4-e4e97ba688a6 tempest-DeleteServersAdminTestJSON-590116201 tempest-DeleteServersAdminTestJSON-590116201-project-member] [instance: 1b509288-2df3-4927-afcf-cb76918d3da3] Destroying instance {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 629.819798] env[61923]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-f03d0b74-4491-4b90-b223-ae268bd4c086 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.822036] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed5bf985-dc30-4a72-9871-73f7368282a7 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.830128] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cae2c54e-db55-413d-ac80-bbff53053031 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.836585] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a6c6fb6-ac40-4a4b-b9e7-da608d523c43 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 629.855011] env[61923]: DEBUG nova.compute.provider_tree [None req-95b9b70a-a835-4901-9a6d-c0fa1b2217f3 tempest-MigrationsAdminTest-1311988845 tempest-MigrationsAdminTest-1311988845-project-member] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 629.860469] env[61923]: WARNING nova.virt.vmwareapi.vmops [None req-0c55849d-766e-4fac-b1c4-e4e97ba688a6 tempest-DeleteServersAdminTestJSON-590116201 tempest-DeleteServersAdminTestJSON-590116201-project-member] [instance: 1b509288-2df3-4927-afcf-cb76918d3da3] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 1b509288-2df3-4927-afcf-cb76918d3da3 could not be found. [ 629.860681] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-0c55849d-766e-4fac-b1c4-e4e97ba688a6 tempest-DeleteServersAdminTestJSON-590116201 tempest-DeleteServersAdminTestJSON-590116201-project-member] [instance: 1b509288-2df3-4927-afcf-cb76918d3da3] Instance destroyed {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 629.860860] env[61923]: INFO nova.compute.manager [None req-0c55849d-766e-4fac-b1c4-e4e97ba688a6 tempest-DeleteServersAdminTestJSON-590116201 tempest-DeleteServersAdminTestJSON-590116201-project-member] [instance: 1b509288-2df3-4927-afcf-cb76918d3da3] Took 0.04 seconds to destroy the instance on the hypervisor. [ 629.861112] env[61923]: DEBUG oslo.service.loopingcall [None req-0c55849d-766e-4fac-b1c4-e4e97ba688a6 tempest-DeleteServersAdminTestJSON-590116201 tempest-DeleteServersAdminTestJSON-590116201-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61923) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 629.861543] env[61923]: DEBUG nova.compute.manager [-] [instance: 1b509288-2df3-4927-afcf-cb76918d3da3] Deallocating network for instance {{(pid=61923) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 629.861643] env[61923]: DEBUG nova.network.neutron [-] [instance: 1b509288-2df3-4927-afcf-cb76918d3da3] deallocate_for_instance() {{(pid=61923) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 629.885520] env[61923]: DEBUG nova.network.neutron [-] [instance: 1b509288-2df3-4927-afcf-cb76918d3da3] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 630.065114] env[61923]: DEBUG nova.compute.manager [req-6a3d6c65-bd7f-449d-80a1-b1847968ac46 req-1a7b5c39-8cf3-4c63-acd0-44b2799abac5 service nova] [instance: 1b509288-2df3-4927-afcf-cb76918d3da3] Received event network-changed-dbf4bfa8-cc4a-4be3-a70c-7f709e040750 {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 630.065373] env[61923]: DEBUG nova.compute.manager [req-6a3d6c65-bd7f-449d-80a1-b1847968ac46 req-1a7b5c39-8cf3-4c63-acd0-44b2799abac5 service nova] [instance: 1b509288-2df3-4927-afcf-cb76918d3da3] Refreshing instance network info cache due to event network-changed-dbf4bfa8-cc4a-4be3-a70c-7f709e040750. {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 630.065555] env[61923]: DEBUG oslo_concurrency.lockutils [req-6a3d6c65-bd7f-449d-80a1-b1847968ac46 req-1a7b5c39-8cf3-4c63-acd0-44b2799abac5 service nova] Acquiring lock "refresh_cache-1b509288-2df3-4927-afcf-cb76918d3da3" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 630.065762] env[61923]: DEBUG oslo_concurrency.lockutils [req-6a3d6c65-bd7f-449d-80a1-b1847968ac46 req-1a7b5c39-8cf3-4c63-acd0-44b2799abac5 service nova] Acquired lock "refresh_cache-1b509288-2df3-4927-afcf-cb76918d3da3" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 630.065853] env[61923]: DEBUG nova.network.neutron [req-6a3d6c65-bd7f-449d-80a1-b1847968ac46 req-1a7b5c39-8cf3-4c63-acd0-44b2799abac5 service nova] [instance: 1b509288-2df3-4927-afcf-cb76918d3da3] Refreshing network info cache for port dbf4bfa8-cc4a-4be3-a70c-7f709e040750 {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 630.144600] env[61923]: DEBUG nova.network.neutron [None req-f3aa5656-98df-4c46-9436-b968a8770dbc tempest-ListServerFiltersTestJSON-483030130 tempest-ListServerFiltersTestJSON-483030130-project-member] [instance: 896dd689-4810-4f23-af0a-d2f557a0796b] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 630.364111] env[61923]: DEBUG nova.scheduler.client.report [None req-95b9b70a-a835-4901-9a6d-c0fa1b2217f3 tempest-MigrationsAdminTest-1311988845 tempest-MigrationsAdminTest-1311988845-project-member] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 630.387316] env[61923]: DEBUG nova.network.neutron [-] [instance: 1b509288-2df3-4927-afcf-cb76918d3da3] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 630.595170] env[61923]: DEBUG nova.network.neutron [req-6a3d6c65-bd7f-449d-80a1-b1847968ac46 req-1a7b5c39-8cf3-4c63-acd0-44b2799abac5 service nova] [instance: 1b509288-2df3-4927-afcf-cb76918d3da3] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 630.646717] env[61923]: INFO nova.compute.manager [None req-f3aa5656-98df-4c46-9436-b968a8770dbc tempest-ListServerFiltersTestJSON-483030130 tempest-ListServerFiltersTestJSON-483030130-project-member] [instance: 896dd689-4810-4f23-af0a-d2f557a0796b] Took 1.03 seconds to deallocate network for instance. [ 630.663794] env[61923]: DEBUG nova.network.neutron [req-6a3d6c65-bd7f-449d-80a1-b1847968ac46 req-1a7b5c39-8cf3-4c63-acd0-44b2799abac5 service nova] [instance: 1b509288-2df3-4927-afcf-cb76918d3da3] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 630.868413] env[61923]: DEBUG oslo_concurrency.lockutils [None req-95b9b70a-a835-4901-9a6d-c0fa1b2217f3 tempest-MigrationsAdminTest-1311988845 tempest-MigrationsAdminTest-1311988845-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.388s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 630.868907] env[61923]: DEBUG nova.compute.manager [None req-95b9b70a-a835-4901-9a6d-c0fa1b2217f3 tempest-MigrationsAdminTest-1311988845 tempest-MigrationsAdminTest-1311988845-project-member] [instance: a4b54a7a-3f3d-49ae-ba43-5a10c262d4e5] Start building networks asynchronously for instance. {{(pid=61923) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 630.871496] env[61923]: DEBUG oslo_concurrency.lockutils [None req-cb394400-3d4b-4149-a290-413641bb6f18 tempest-InstanceActionsV221TestJSON-1730893270 tempest-InstanceActionsV221TestJSON-1730893270-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.789s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 630.872880] env[61923]: INFO nova.compute.claims [None req-cb394400-3d4b-4149-a290-413641bb6f18 tempest-InstanceActionsV221TestJSON-1730893270 tempest-InstanceActionsV221TestJSON-1730893270-project-member] [instance: dc8cc22d-aba9-4dba-b456-1b323013d234] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 630.890096] env[61923]: INFO nova.compute.manager [-] [instance: 1b509288-2df3-4927-afcf-cb76918d3da3] Took 1.03 seconds to deallocate network for instance. [ 630.892370] env[61923]: DEBUG nova.compute.claims [None req-0c55849d-766e-4fac-b1c4-e4e97ba688a6 tempest-DeleteServersAdminTestJSON-590116201 tempest-DeleteServersAdminTestJSON-590116201-project-member] [instance: 1b509288-2df3-4927-afcf-cb76918d3da3] Aborting claim: {{(pid=61923) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 630.892547] env[61923]: DEBUG oslo_concurrency.lockutils [None req-0c55849d-766e-4fac-b1c4-e4e97ba688a6 tempest-DeleteServersAdminTestJSON-590116201 tempest-DeleteServersAdminTestJSON-590116201-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 631.165833] env[61923]: DEBUG oslo_concurrency.lockutils [req-6a3d6c65-bd7f-449d-80a1-b1847968ac46 req-1a7b5c39-8cf3-4c63-acd0-44b2799abac5 service nova] Releasing lock "refresh_cache-1b509288-2df3-4927-afcf-cb76918d3da3" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 631.166119] env[61923]: DEBUG nova.compute.manager [req-6a3d6c65-bd7f-449d-80a1-b1847968ac46 req-1a7b5c39-8cf3-4c63-acd0-44b2799abac5 service nova] [instance: 1b509288-2df3-4927-afcf-cb76918d3da3] Received event network-vif-deleted-dbf4bfa8-cc4a-4be3-a70c-7f709e040750 {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 631.377725] env[61923]: DEBUG nova.compute.utils [None req-95b9b70a-a835-4901-9a6d-c0fa1b2217f3 tempest-MigrationsAdminTest-1311988845 tempest-MigrationsAdminTest-1311988845-project-member] Using /dev/sd instead of None {{(pid=61923) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 631.381050] env[61923]: DEBUG nova.compute.manager [None req-95b9b70a-a835-4901-9a6d-c0fa1b2217f3 tempest-MigrationsAdminTest-1311988845 tempest-MigrationsAdminTest-1311988845-project-member] [instance: a4b54a7a-3f3d-49ae-ba43-5a10c262d4e5] Allocating IP information in the background. {{(pid=61923) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 631.381050] env[61923]: DEBUG nova.network.neutron [None req-95b9b70a-a835-4901-9a6d-c0fa1b2217f3 tempest-MigrationsAdminTest-1311988845 tempest-MigrationsAdminTest-1311988845-project-member] [instance: a4b54a7a-3f3d-49ae-ba43-5a10c262d4e5] allocate_for_instance() {{(pid=61923) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 631.428622] env[61923]: DEBUG nova.policy [None req-95b9b70a-a835-4901-9a6d-c0fa1b2217f3 tempest-MigrationsAdminTest-1311988845 tempest-MigrationsAdminTest-1311988845-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5eaeac65cba14de5b8cd759cf5fba8f4', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'feb670247d4a4baabfb19df3f9399223', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61923) authorize /opt/stack/nova/nova/policy.py:201}} [ 631.684878] env[61923]: INFO nova.scheduler.client.report [None req-f3aa5656-98df-4c46-9436-b968a8770dbc tempest-ListServerFiltersTestJSON-483030130 tempest-ListServerFiltersTestJSON-483030130-project-member] Deleted allocations for instance 896dd689-4810-4f23-af0a-d2f557a0796b [ 631.799188] env[61923]: DEBUG nova.network.neutron [None req-95b9b70a-a835-4901-9a6d-c0fa1b2217f3 tempest-MigrationsAdminTest-1311988845 tempest-MigrationsAdminTest-1311988845-project-member] [instance: a4b54a7a-3f3d-49ae-ba43-5a10c262d4e5] Successfully created port: 2110c7b3-1dff-4f09-b631-618ae3b8467e {{(pid=61923) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 631.881503] env[61923]: DEBUG nova.compute.manager [None req-95b9b70a-a835-4901-9a6d-c0fa1b2217f3 tempest-MigrationsAdminTest-1311988845 tempest-MigrationsAdminTest-1311988845-project-member] [instance: a4b54a7a-3f3d-49ae-ba43-5a10c262d4e5] Start building block device mappings for instance. {{(pid=61923) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 632.195958] env[61923]: DEBUG oslo_concurrency.lockutils [None req-f3aa5656-98df-4c46-9436-b968a8770dbc tempest-ListServerFiltersTestJSON-483030130 tempest-ListServerFiltersTestJSON-483030130-project-member] Lock "896dd689-4810-4f23-af0a-d2f557a0796b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 90.475s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 632.208863] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35377d49-f1b1-473f-8691-48261b20ddb7 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.218204] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5d472e0-11f2-4208-8a98-e60344b406c4 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.249552] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40f90932-9445-4f1d-a96c-ed7b71422e33 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.256679] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97699589-883c-4479-bb97-592429d4c6f8 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.273861] env[61923]: DEBUG nova.compute.provider_tree [None req-cb394400-3d4b-4149-a290-413641bb6f18 tempest-InstanceActionsV221TestJSON-1730893270 tempest-InstanceActionsV221TestJSON-1730893270-project-member] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 632.700596] env[61923]: DEBUG nova.compute.manager [None req-56950b6c-0b87-48bf-a575-f699fcd1a5b0 tempest-FloatingIPsAssociationTestJSON-1125124803 tempest-FloatingIPsAssociationTestJSON-1125124803-project-member] [instance: aeca6a94-478d-435a-9d93-59936f26654e] Starting instance... {{(pid=61923) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 632.782055] env[61923]: DEBUG nova.scheduler.client.report [None req-cb394400-3d4b-4149-a290-413641bb6f18 tempest-InstanceActionsV221TestJSON-1730893270 tempest-InstanceActionsV221TestJSON-1730893270-project-member] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 632.895189] env[61923]: DEBUG nova.compute.manager [None req-95b9b70a-a835-4901-9a6d-c0fa1b2217f3 tempest-MigrationsAdminTest-1311988845 tempest-MigrationsAdminTest-1311988845-project-member] [instance: a4b54a7a-3f3d-49ae-ba43-5a10c262d4e5] Start spawning the instance on the hypervisor. {{(pid=61923) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 632.923963] env[61923]: DEBUG nova.virt.hardware [None req-95b9b70a-a835-4901-9a6d-c0fa1b2217f3 tempest-MigrationsAdminTest-1311988845 tempest-MigrationsAdminTest-1311988845-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-29T20:09:58Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='110705e7-7a2e-417c-94b4-11e653f53955',id=26,is_public=True,memory_mb=192,name='tempest-test_resize_flavor_-1307836108',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-29T20:07:35Z,direct_url=,disk_format='vmdk',id=0f153f63-ae0a-45b1-b7f5-7f9b673c947f,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='4e7b5e3b3c3443c8bd5c70f8a15739f5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-29T20:07:36Z,virtual_size=,visibility=), allow threads: False {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 632.923963] env[61923]: DEBUG nova.virt.hardware [None req-95b9b70a-a835-4901-9a6d-c0fa1b2217f3 tempest-MigrationsAdminTest-1311988845 tempest-MigrationsAdminTest-1311988845-project-member] Flavor limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 632.923963] env[61923]: DEBUG nova.virt.hardware [None req-95b9b70a-a835-4901-9a6d-c0fa1b2217f3 tempest-MigrationsAdminTest-1311988845 tempest-MigrationsAdminTest-1311988845-project-member] Image limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 632.924160] env[61923]: DEBUG nova.virt.hardware [None req-95b9b70a-a835-4901-9a6d-c0fa1b2217f3 tempest-MigrationsAdminTest-1311988845 tempest-MigrationsAdminTest-1311988845-project-member] Flavor pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 632.924379] env[61923]: DEBUG nova.virt.hardware [None req-95b9b70a-a835-4901-9a6d-c0fa1b2217f3 tempest-MigrationsAdminTest-1311988845 tempest-MigrationsAdminTest-1311988845-project-member] Image pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 632.924989] env[61923]: DEBUG nova.virt.hardware [None req-95b9b70a-a835-4901-9a6d-c0fa1b2217f3 tempest-MigrationsAdminTest-1311988845 tempest-MigrationsAdminTest-1311988845-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 632.925368] env[61923]: DEBUG nova.virt.hardware [None req-95b9b70a-a835-4901-9a6d-c0fa1b2217f3 tempest-MigrationsAdminTest-1311988845 tempest-MigrationsAdminTest-1311988845-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 632.928188] env[61923]: DEBUG nova.virt.hardware [None req-95b9b70a-a835-4901-9a6d-c0fa1b2217f3 tempest-MigrationsAdminTest-1311988845 tempest-MigrationsAdminTest-1311988845-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 632.928188] env[61923]: DEBUG nova.virt.hardware [None req-95b9b70a-a835-4901-9a6d-c0fa1b2217f3 tempest-MigrationsAdminTest-1311988845 tempest-MigrationsAdminTest-1311988845-project-member] Got 1 possible topologies {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 632.928188] env[61923]: DEBUG nova.virt.hardware [None req-95b9b70a-a835-4901-9a6d-c0fa1b2217f3 tempest-MigrationsAdminTest-1311988845 tempest-MigrationsAdminTest-1311988845-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 632.928188] env[61923]: DEBUG nova.virt.hardware [None req-95b9b70a-a835-4901-9a6d-c0fa1b2217f3 tempest-MigrationsAdminTest-1311988845 tempest-MigrationsAdminTest-1311988845-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 632.928188] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bba32bfb-83e4-4efa-93a6-8b780727e8db {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 632.935322] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-032be5be-a4a1-4c7e-bbaf-90194c082622 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.038736] env[61923]: DEBUG nova.compute.manager [req-6a89204a-e979-4c81-baaf-95679a332296 req-9fd3799a-c31e-4798-814f-aaaac58178e7 service nova] [instance: a4b54a7a-3f3d-49ae-ba43-5a10c262d4e5] Received event network-changed-2110c7b3-1dff-4f09-b631-618ae3b8467e {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 633.039125] env[61923]: DEBUG nova.compute.manager [req-6a89204a-e979-4c81-baaf-95679a332296 req-9fd3799a-c31e-4798-814f-aaaac58178e7 service nova] [instance: a4b54a7a-3f3d-49ae-ba43-5a10c262d4e5] Refreshing instance network info cache due to event network-changed-2110c7b3-1dff-4f09-b631-618ae3b8467e. {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 633.039530] env[61923]: DEBUG oslo_concurrency.lockutils [req-6a89204a-e979-4c81-baaf-95679a332296 req-9fd3799a-c31e-4798-814f-aaaac58178e7 service nova] Acquiring lock "refresh_cache-a4b54a7a-3f3d-49ae-ba43-5a10c262d4e5" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 633.039793] env[61923]: DEBUG oslo_concurrency.lockutils [req-6a89204a-e979-4c81-baaf-95679a332296 req-9fd3799a-c31e-4798-814f-aaaac58178e7 service nova] Acquired lock "refresh_cache-a4b54a7a-3f3d-49ae-ba43-5a10c262d4e5" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 633.040069] env[61923]: DEBUG nova.network.neutron [req-6a89204a-e979-4c81-baaf-95679a332296 req-9fd3799a-c31e-4798-814f-aaaac58178e7 service nova] [instance: a4b54a7a-3f3d-49ae-ba43-5a10c262d4e5] Refreshing network info cache for port 2110c7b3-1dff-4f09-b631-618ae3b8467e {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 633.222210] env[61923]: DEBUG oslo_concurrency.lockutils [None req-56950b6c-0b87-48bf-a575-f699fcd1a5b0 tempest-FloatingIPsAssociationTestJSON-1125124803 tempest-FloatingIPsAssociationTestJSON-1125124803-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 633.242156] env[61923]: ERROR nova.compute.manager [None req-95b9b70a-a835-4901-9a6d-c0fa1b2217f3 tempest-MigrationsAdminTest-1311988845 tempest-MigrationsAdminTest-1311988845-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 2110c7b3-1dff-4f09-b631-618ae3b8467e, please check neutron logs for more information. [ 633.242156] env[61923]: ERROR nova.compute.manager Traceback (most recent call last): [ 633.242156] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 633.242156] env[61923]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 633.242156] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 633.242156] env[61923]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 633.242156] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 633.242156] env[61923]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 633.242156] env[61923]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 633.242156] env[61923]: ERROR nova.compute.manager self.force_reraise() [ 633.242156] env[61923]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 633.242156] env[61923]: ERROR nova.compute.manager raise self.value [ 633.242156] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 633.242156] env[61923]: ERROR nova.compute.manager updated_port = self._update_port( [ 633.242156] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 633.242156] env[61923]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 633.242626] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 633.242626] env[61923]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 633.242626] env[61923]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 2110c7b3-1dff-4f09-b631-618ae3b8467e, please check neutron logs for more information. [ 633.242626] env[61923]: ERROR nova.compute.manager [ 633.242626] env[61923]: Traceback (most recent call last): [ 633.242626] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 633.242626] env[61923]: listener.cb(fileno) [ 633.242626] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 633.242626] env[61923]: result = function(*args, **kwargs) [ 633.242626] env[61923]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 633.242626] env[61923]: return func(*args, **kwargs) [ 633.242626] env[61923]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 633.242626] env[61923]: raise e [ 633.242626] env[61923]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 633.242626] env[61923]: nwinfo = self.network_api.allocate_for_instance( [ 633.242626] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 633.242626] env[61923]: created_port_ids = self._update_ports_for_instance( [ 633.242626] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 633.242626] env[61923]: with excutils.save_and_reraise_exception(): [ 633.242626] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 633.242626] env[61923]: self.force_reraise() [ 633.242626] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 633.242626] env[61923]: raise self.value [ 633.242626] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 633.242626] env[61923]: updated_port = self._update_port( [ 633.242626] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 633.242626] env[61923]: _ensure_no_port_binding_failure(port) [ 633.242626] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 633.242626] env[61923]: raise exception.PortBindingFailed(port_id=port['id']) [ 633.243462] env[61923]: nova.exception.PortBindingFailed: Binding failed for port 2110c7b3-1dff-4f09-b631-618ae3b8467e, please check neutron logs for more information. [ 633.243462] env[61923]: Removing descriptor: 17 [ 633.243462] env[61923]: ERROR nova.compute.manager [None req-95b9b70a-a835-4901-9a6d-c0fa1b2217f3 tempest-MigrationsAdminTest-1311988845 tempest-MigrationsAdminTest-1311988845-project-member] [instance: a4b54a7a-3f3d-49ae-ba43-5a10c262d4e5] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 2110c7b3-1dff-4f09-b631-618ae3b8467e, please check neutron logs for more information. [ 633.243462] env[61923]: ERROR nova.compute.manager [instance: a4b54a7a-3f3d-49ae-ba43-5a10c262d4e5] Traceback (most recent call last): [ 633.243462] env[61923]: ERROR nova.compute.manager [instance: a4b54a7a-3f3d-49ae-ba43-5a10c262d4e5] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 633.243462] env[61923]: ERROR nova.compute.manager [instance: a4b54a7a-3f3d-49ae-ba43-5a10c262d4e5] yield resources [ 633.243462] env[61923]: ERROR nova.compute.manager [instance: a4b54a7a-3f3d-49ae-ba43-5a10c262d4e5] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 633.243462] env[61923]: ERROR nova.compute.manager [instance: a4b54a7a-3f3d-49ae-ba43-5a10c262d4e5] self.driver.spawn(context, instance, image_meta, [ 633.243462] env[61923]: ERROR nova.compute.manager [instance: a4b54a7a-3f3d-49ae-ba43-5a10c262d4e5] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 633.243462] env[61923]: ERROR nova.compute.manager [instance: a4b54a7a-3f3d-49ae-ba43-5a10c262d4e5] self._vmops.spawn(context, instance, image_meta, injected_files, [ 633.243462] env[61923]: ERROR nova.compute.manager [instance: a4b54a7a-3f3d-49ae-ba43-5a10c262d4e5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 633.243462] env[61923]: ERROR nova.compute.manager [instance: a4b54a7a-3f3d-49ae-ba43-5a10c262d4e5] vm_ref = self.build_virtual_machine(instance, [ 633.243883] env[61923]: ERROR nova.compute.manager [instance: a4b54a7a-3f3d-49ae-ba43-5a10c262d4e5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 633.243883] env[61923]: ERROR nova.compute.manager [instance: a4b54a7a-3f3d-49ae-ba43-5a10c262d4e5] vif_infos = vmwarevif.get_vif_info(self._session, [ 633.243883] env[61923]: ERROR nova.compute.manager [instance: a4b54a7a-3f3d-49ae-ba43-5a10c262d4e5] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 633.243883] env[61923]: ERROR nova.compute.manager [instance: a4b54a7a-3f3d-49ae-ba43-5a10c262d4e5] for vif in network_info: [ 633.243883] env[61923]: ERROR nova.compute.manager [instance: a4b54a7a-3f3d-49ae-ba43-5a10c262d4e5] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 633.243883] env[61923]: ERROR nova.compute.manager [instance: a4b54a7a-3f3d-49ae-ba43-5a10c262d4e5] return self._sync_wrapper(fn, *args, **kwargs) [ 633.243883] env[61923]: ERROR nova.compute.manager [instance: a4b54a7a-3f3d-49ae-ba43-5a10c262d4e5] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 633.243883] env[61923]: ERROR nova.compute.manager [instance: a4b54a7a-3f3d-49ae-ba43-5a10c262d4e5] self.wait() [ 633.243883] env[61923]: ERROR nova.compute.manager [instance: a4b54a7a-3f3d-49ae-ba43-5a10c262d4e5] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 633.243883] env[61923]: ERROR nova.compute.manager [instance: a4b54a7a-3f3d-49ae-ba43-5a10c262d4e5] self[:] = self._gt.wait() [ 633.243883] env[61923]: ERROR nova.compute.manager [instance: a4b54a7a-3f3d-49ae-ba43-5a10c262d4e5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 633.243883] env[61923]: ERROR nova.compute.manager [instance: a4b54a7a-3f3d-49ae-ba43-5a10c262d4e5] return self._exit_event.wait() [ 633.243883] env[61923]: ERROR nova.compute.manager [instance: a4b54a7a-3f3d-49ae-ba43-5a10c262d4e5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 633.244257] env[61923]: ERROR nova.compute.manager [instance: a4b54a7a-3f3d-49ae-ba43-5a10c262d4e5] result = hub.switch() [ 633.244257] env[61923]: ERROR nova.compute.manager [instance: a4b54a7a-3f3d-49ae-ba43-5a10c262d4e5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 633.244257] env[61923]: ERROR nova.compute.manager [instance: a4b54a7a-3f3d-49ae-ba43-5a10c262d4e5] return self.greenlet.switch() [ 633.244257] env[61923]: ERROR nova.compute.manager [instance: a4b54a7a-3f3d-49ae-ba43-5a10c262d4e5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 633.244257] env[61923]: ERROR nova.compute.manager [instance: a4b54a7a-3f3d-49ae-ba43-5a10c262d4e5] result = function(*args, **kwargs) [ 633.244257] env[61923]: ERROR nova.compute.manager [instance: a4b54a7a-3f3d-49ae-ba43-5a10c262d4e5] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 633.244257] env[61923]: ERROR nova.compute.manager [instance: a4b54a7a-3f3d-49ae-ba43-5a10c262d4e5] return func(*args, **kwargs) [ 633.244257] env[61923]: ERROR nova.compute.manager [instance: a4b54a7a-3f3d-49ae-ba43-5a10c262d4e5] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 633.244257] env[61923]: ERROR nova.compute.manager [instance: a4b54a7a-3f3d-49ae-ba43-5a10c262d4e5] raise e [ 633.244257] env[61923]: ERROR nova.compute.manager [instance: a4b54a7a-3f3d-49ae-ba43-5a10c262d4e5] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 633.244257] env[61923]: ERROR nova.compute.manager [instance: a4b54a7a-3f3d-49ae-ba43-5a10c262d4e5] nwinfo = self.network_api.allocate_for_instance( [ 633.244257] env[61923]: ERROR nova.compute.manager [instance: a4b54a7a-3f3d-49ae-ba43-5a10c262d4e5] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 633.244257] env[61923]: ERROR nova.compute.manager [instance: a4b54a7a-3f3d-49ae-ba43-5a10c262d4e5] created_port_ids = self._update_ports_for_instance( [ 633.244621] env[61923]: ERROR nova.compute.manager [instance: a4b54a7a-3f3d-49ae-ba43-5a10c262d4e5] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 633.244621] env[61923]: ERROR nova.compute.manager [instance: a4b54a7a-3f3d-49ae-ba43-5a10c262d4e5] with excutils.save_and_reraise_exception(): [ 633.244621] env[61923]: ERROR nova.compute.manager [instance: a4b54a7a-3f3d-49ae-ba43-5a10c262d4e5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 633.244621] env[61923]: ERROR nova.compute.manager [instance: a4b54a7a-3f3d-49ae-ba43-5a10c262d4e5] self.force_reraise() [ 633.244621] env[61923]: ERROR nova.compute.manager [instance: a4b54a7a-3f3d-49ae-ba43-5a10c262d4e5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 633.244621] env[61923]: ERROR nova.compute.manager [instance: a4b54a7a-3f3d-49ae-ba43-5a10c262d4e5] raise self.value [ 633.244621] env[61923]: ERROR nova.compute.manager [instance: a4b54a7a-3f3d-49ae-ba43-5a10c262d4e5] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 633.244621] env[61923]: ERROR nova.compute.manager [instance: a4b54a7a-3f3d-49ae-ba43-5a10c262d4e5] updated_port = self._update_port( [ 633.244621] env[61923]: ERROR nova.compute.manager [instance: a4b54a7a-3f3d-49ae-ba43-5a10c262d4e5] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 633.244621] env[61923]: ERROR nova.compute.manager [instance: a4b54a7a-3f3d-49ae-ba43-5a10c262d4e5] _ensure_no_port_binding_failure(port) [ 633.244621] env[61923]: ERROR nova.compute.manager [instance: a4b54a7a-3f3d-49ae-ba43-5a10c262d4e5] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 633.244621] env[61923]: ERROR nova.compute.manager [instance: a4b54a7a-3f3d-49ae-ba43-5a10c262d4e5] raise exception.PortBindingFailed(port_id=port['id']) [ 633.244971] env[61923]: ERROR nova.compute.manager [instance: a4b54a7a-3f3d-49ae-ba43-5a10c262d4e5] nova.exception.PortBindingFailed: Binding failed for port 2110c7b3-1dff-4f09-b631-618ae3b8467e, please check neutron logs for more information. [ 633.244971] env[61923]: ERROR nova.compute.manager [instance: a4b54a7a-3f3d-49ae-ba43-5a10c262d4e5] [ 633.244971] env[61923]: INFO nova.compute.manager [None req-95b9b70a-a835-4901-9a6d-c0fa1b2217f3 tempest-MigrationsAdminTest-1311988845 tempest-MigrationsAdminTest-1311988845-project-member] [instance: a4b54a7a-3f3d-49ae-ba43-5a10c262d4e5] Terminating instance [ 633.246054] env[61923]: DEBUG oslo_concurrency.lockutils [None req-95b9b70a-a835-4901-9a6d-c0fa1b2217f3 tempest-MigrationsAdminTest-1311988845 tempest-MigrationsAdminTest-1311988845-project-member] Acquiring lock "refresh_cache-a4b54a7a-3f3d-49ae-ba43-5a10c262d4e5" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 633.287150] env[61923]: DEBUG oslo_concurrency.lockutils [None req-cb394400-3d4b-4149-a290-413641bb6f18 tempest-InstanceActionsV221TestJSON-1730893270 tempest-InstanceActionsV221TestJSON-1730893270-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.415s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 633.287689] env[61923]: DEBUG nova.compute.manager [None req-cb394400-3d4b-4149-a290-413641bb6f18 tempest-InstanceActionsV221TestJSON-1730893270 tempest-InstanceActionsV221TestJSON-1730893270-project-member] [instance: dc8cc22d-aba9-4dba-b456-1b323013d234] Start building networks asynchronously for instance. {{(pid=61923) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 633.291307] env[61923]: DEBUG oslo_concurrency.lockutils [None req-753531fa-39a1-4c82-9915-d195bf0fe1b2 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.837s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 633.294342] env[61923]: INFO nova.compute.claims [None req-753531fa-39a1-4c82-9915-d195bf0fe1b2 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] [instance: 03100310-86ce-4739-8636-cd2aa1b542ff] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 633.557805] env[61923]: DEBUG nova.network.neutron [req-6a89204a-e979-4c81-baaf-95679a332296 req-9fd3799a-c31e-4798-814f-aaaac58178e7 service nova] [instance: a4b54a7a-3f3d-49ae-ba43-5a10c262d4e5] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 633.615492] env[61923]: DEBUG nova.network.neutron [req-6a89204a-e979-4c81-baaf-95679a332296 req-9fd3799a-c31e-4798-814f-aaaac58178e7 service nova] [instance: a4b54a7a-3f3d-49ae-ba43-5a10c262d4e5] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 633.798651] env[61923]: DEBUG nova.compute.utils [None req-cb394400-3d4b-4149-a290-413641bb6f18 tempest-InstanceActionsV221TestJSON-1730893270 tempest-InstanceActionsV221TestJSON-1730893270-project-member] Using /dev/sd instead of None {{(pid=61923) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 633.802372] env[61923]: DEBUG nova.compute.manager [None req-cb394400-3d4b-4149-a290-413641bb6f18 tempest-InstanceActionsV221TestJSON-1730893270 tempest-InstanceActionsV221TestJSON-1730893270-project-member] [instance: dc8cc22d-aba9-4dba-b456-1b323013d234] Allocating IP information in the background. {{(pid=61923) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 633.802540] env[61923]: DEBUG nova.network.neutron [None req-cb394400-3d4b-4149-a290-413641bb6f18 tempest-InstanceActionsV221TestJSON-1730893270 tempest-InstanceActionsV221TestJSON-1730893270-project-member] [instance: dc8cc22d-aba9-4dba-b456-1b323013d234] allocate_for_instance() {{(pid=61923) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 633.865462] env[61923]: DEBUG nova.policy [None req-cb394400-3d4b-4149-a290-413641bb6f18 tempest-InstanceActionsV221TestJSON-1730893270 tempest-InstanceActionsV221TestJSON-1730893270-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0b89a8db1d9b42d0aa498f10d3093e1f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '74e4387c2bcb488885143a50a4c79cb9', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61923) authorize /opt/stack/nova/nova/policy.py:201}} [ 634.117958] env[61923]: DEBUG oslo_concurrency.lockutils [req-6a89204a-e979-4c81-baaf-95679a332296 req-9fd3799a-c31e-4798-814f-aaaac58178e7 service nova] Releasing lock "refresh_cache-a4b54a7a-3f3d-49ae-ba43-5a10c262d4e5" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 634.118572] env[61923]: DEBUG oslo_concurrency.lockutils [None req-95b9b70a-a835-4901-9a6d-c0fa1b2217f3 tempest-MigrationsAdminTest-1311988845 tempest-MigrationsAdminTest-1311988845-project-member] Acquired lock "refresh_cache-a4b54a7a-3f3d-49ae-ba43-5a10c262d4e5" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 634.118866] env[61923]: DEBUG nova.network.neutron [None req-95b9b70a-a835-4901-9a6d-c0fa1b2217f3 tempest-MigrationsAdminTest-1311988845 tempest-MigrationsAdminTest-1311988845-project-member] [instance: a4b54a7a-3f3d-49ae-ba43-5a10c262d4e5] Building network info cache for instance {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 634.208491] env[61923]: DEBUG nova.network.neutron [None req-cb394400-3d4b-4149-a290-413641bb6f18 tempest-InstanceActionsV221TestJSON-1730893270 tempest-InstanceActionsV221TestJSON-1730893270-project-member] [instance: dc8cc22d-aba9-4dba-b456-1b323013d234] Successfully created port: 07f299cc-2ca7-4beb-9f20-7509c49ed4f0 {{(pid=61923) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 634.303883] env[61923]: DEBUG nova.compute.manager [None req-cb394400-3d4b-4149-a290-413641bb6f18 tempest-InstanceActionsV221TestJSON-1730893270 tempest-InstanceActionsV221TestJSON-1730893270-project-member] [instance: dc8cc22d-aba9-4dba-b456-1b323013d234] Start building block device mappings for instance. {{(pid=61923) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 634.642054] env[61923]: DEBUG nova.network.neutron [None req-95b9b70a-a835-4901-9a6d-c0fa1b2217f3 tempest-MigrationsAdminTest-1311988845 tempest-MigrationsAdminTest-1311988845-project-member] [instance: a4b54a7a-3f3d-49ae-ba43-5a10c262d4e5] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 634.701519] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0afc0e6-406a-407b-af49-f4f2f0772d1f {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.709524] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-246f1ef2-6b8e-479b-b32e-08f7cc196f05 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.740086] env[61923]: DEBUG nova.network.neutron [None req-95b9b70a-a835-4901-9a6d-c0fa1b2217f3 tempest-MigrationsAdminTest-1311988845 tempest-MigrationsAdminTest-1311988845-project-member] [instance: a4b54a7a-3f3d-49ae-ba43-5a10c262d4e5] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 634.743397] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c150cbbe-563c-4e97-9605-11dbfec81b6f {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.751085] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84a1f9d5-2aba-48a9-aec0-0beb26c05ba1 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.766346] env[61923]: DEBUG nova.compute.provider_tree [None req-753531fa-39a1-4c82-9915-d195bf0fe1b2 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 635.062847] env[61923]: DEBUG nova.compute.manager [req-945dc763-fb83-4085-93c1-687ef94d8966 req-3f1f6b2d-778c-4aac-84b6-737d707870d8 service nova] [instance: a4b54a7a-3f3d-49ae-ba43-5a10c262d4e5] Received event network-vif-deleted-2110c7b3-1dff-4f09-b631-618ae3b8467e {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 635.244323] env[61923]: DEBUG oslo_concurrency.lockutils [None req-95b9b70a-a835-4901-9a6d-c0fa1b2217f3 tempest-MigrationsAdminTest-1311988845 tempest-MigrationsAdminTest-1311988845-project-member] Releasing lock "refresh_cache-a4b54a7a-3f3d-49ae-ba43-5a10c262d4e5" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 635.244323] env[61923]: DEBUG nova.compute.manager [None req-95b9b70a-a835-4901-9a6d-c0fa1b2217f3 tempest-MigrationsAdminTest-1311988845 tempest-MigrationsAdminTest-1311988845-project-member] [instance: a4b54a7a-3f3d-49ae-ba43-5a10c262d4e5] Start destroying the instance on the hypervisor. {{(pid=61923) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 635.244323] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-95b9b70a-a835-4901-9a6d-c0fa1b2217f3 tempest-MigrationsAdminTest-1311988845 tempest-MigrationsAdminTest-1311988845-project-member] [instance: a4b54a7a-3f3d-49ae-ba43-5a10c262d4e5] Destroying instance {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 635.244323] env[61923]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d0c7abac-d93a-4be3-a3a9-8baaaeb2171d {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.252657] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc947423-0bcb-4756-b451-587bb42833d1 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.269640] env[61923]: DEBUG nova.scheduler.client.report [None req-753531fa-39a1-4c82-9915-d195bf0fe1b2 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 635.276669] env[61923]: WARNING nova.virt.vmwareapi.vmops [None req-95b9b70a-a835-4901-9a6d-c0fa1b2217f3 tempest-MigrationsAdminTest-1311988845 tempest-MigrationsAdminTest-1311988845-project-member] [instance: a4b54a7a-3f3d-49ae-ba43-5a10c262d4e5] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance a4b54a7a-3f3d-49ae-ba43-5a10c262d4e5 could not be found. [ 635.276753] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-95b9b70a-a835-4901-9a6d-c0fa1b2217f3 tempest-MigrationsAdminTest-1311988845 tempest-MigrationsAdminTest-1311988845-project-member] [instance: a4b54a7a-3f3d-49ae-ba43-5a10c262d4e5] Instance destroyed {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 635.276948] env[61923]: INFO nova.compute.manager [None req-95b9b70a-a835-4901-9a6d-c0fa1b2217f3 tempest-MigrationsAdminTest-1311988845 tempest-MigrationsAdminTest-1311988845-project-member] [instance: a4b54a7a-3f3d-49ae-ba43-5a10c262d4e5] Took 0.03 seconds to destroy the instance on the hypervisor. [ 635.277250] env[61923]: DEBUG oslo.service.loopingcall [None req-95b9b70a-a835-4901-9a6d-c0fa1b2217f3 tempest-MigrationsAdminTest-1311988845 tempest-MigrationsAdminTest-1311988845-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61923) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 635.277694] env[61923]: DEBUG nova.compute.manager [-] [instance: a4b54a7a-3f3d-49ae-ba43-5a10c262d4e5] Deallocating network for instance {{(pid=61923) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 635.277803] env[61923]: DEBUG nova.network.neutron [-] [instance: a4b54a7a-3f3d-49ae-ba43-5a10c262d4e5] deallocate_for_instance() {{(pid=61923) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 635.299888] env[61923]: DEBUG nova.network.neutron [-] [instance: a4b54a7a-3f3d-49ae-ba43-5a10c262d4e5] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 635.313410] env[61923]: DEBUG nova.compute.manager [None req-cb394400-3d4b-4149-a290-413641bb6f18 tempest-InstanceActionsV221TestJSON-1730893270 tempest-InstanceActionsV221TestJSON-1730893270-project-member] [instance: dc8cc22d-aba9-4dba-b456-1b323013d234] Start spawning the instance on the hypervisor. {{(pid=61923) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 635.339078] env[61923]: DEBUG nova.virt.hardware [None req-cb394400-3d4b-4149-a290-413641bb6f18 tempest-InstanceActionsV221TestJSON-1730893270 tempest-InstanceActionsV221TestJSON-1730893270-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-29T20:07:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-29T20:07:35Z,direct_url=,disk_format='vmdk',id=0f153f63-ae0a-45b1-b7f5-7f9b673c947f,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='4e7b5e3b3c3443c8bd5c70f8a15739f5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-29T20:07:36Z,virtual_size=,visibility=), allow threads: False {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 635.339350] env[61923]: DEBUG nova.virt.hardware [None req-cb394400-3d4b-4149-a290-413641bb6f18 tempest-InstanceActionsV221TestJSON-1730893270 tempest-InstanceActionsV221TestJSON-1730893270-project-member] Flavor limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 635.339525] env[61923]: DEBUG nova.virt.hardware [None req-cb394400-3d4b-4149-a290-413641bb6f18 tempest-InstanceActionsV221TestJSON-1730893270 tempest-InstanceActionsV221TestJSON-1730893270-project-member] Image limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 635.339705] env[61923]: DEBUG nova.virt.hardware [None req-cb394400-3d4b-4149-a290-413641bb6f18 tempest-InstanceActionsV221TestJSON-1730893270 tempest-InstanceActionsV221TestJSON-1730893270-project-member] Flavor pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 635.340946] env[61923]: DEBUG nova.virt.hardware [None req-cb394400-3d4b-4149-a290-413641bb6f18 tempest-InstanceActionsV221TestJSON-1730893270 tempest-InstanceActionsV221TestJSON-1730893270-project-member] Image pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 635.340946] env[61923]: DEBUG nova.virt.hardware [None req-cb394400-3d4b-4149-a290-413641bb6f18 tempest-InstanceActionsV221TestJSON-1730893270 tempest-InstanceActionsV221TestJSON-1730893270-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 635.340946] env[61923]: DEBUG nova.virt.hardware [None req-cb394400-3d4b-4149-a290-413641bb6f18 tempest-InstanceActionsV221TestJSON-1730893270 tempest-InstanceActionsV221TestJSON-1730893270-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 635.340946] env[61923]: DEBUG nova.virt.hardware [None req-cb394400-3d4b-4149-a290-413641bb6f18 tempest-InstanceActionsV221TestJSON-1730893270 tempest-InstanceActionsV221TestJSON-1730893270-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 635.340946] env[61923]: DEBUG nova.virt.hardware [None req-cb394400-3d4b-4149-a290-413641bb6f18 tempest-InstanceActionsV221TestJSON-1730893270 tempest-InstanceActionsV221TestJSON-1730893270-project-member] Got 1 possible topologies {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 635.341152] env[61923]: DEBUG nova.virt.hardware [None req-cb394400-3d4b-4149-a290-413641bb6f18 tempest-InstanceActionsV221TestJSON-1730893270 tempest-InstanceActionsV221TestJSON-1730893270-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 635.341152] env[61923]: DEBUG nova.virt.hardware [None req-cb394400-3d4b-4149-a290-413641bb6f18 tempest-InstanceActionsV221TestJSON-1730893270 tempest-InstanceActionsV221TestJSON-1730893270-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 635.341998] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0573d0a9-4ad8-42b4-a370-94d23f400aef {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.349919] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23c4a32a-25cc-4461-bd71-9d8def344b7f {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.432706] env[61923]: ERROR nova.compute.manager [None req-cb394400-3d4b-4149-a290-413641bb6f18 tempest-InstanceActionsV221TestJSON-1730893270 tempest-InstanceActionsV221TestJSON-1730893270-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 07f299cc-2ca7-4beb-9f20-7509c49ed4f0, please check neutron logs for more information. [ 635.432706] env[61923]: ERROR nova.compute.manager Traceback (most recent call last): [ 635.432706] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 635.432706] env[61923]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 635.432706] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 635.432706] env[61923]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 635.432706] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 635.432706] env[61923]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 635.432706] env[61923]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 635.432706] env[61923]: ERROR nova.compute.manager self.force_reraise() [ 635.432706] env[61923]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 635.432706] env[61923]: ERROR nova.compute.manager raise self.value [ 635.432706] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 635.432706] env[61923]: ERROR nova.compute.manager updated_port = self._update_port( [ 635.432706] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 635.432706] env[61923]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 635.433395] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 635.433395] env[61923]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 635.433395] env[61923]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 07f299cc-2ca7-4beb-9f20-7509c49ed4f0, please check neutron logs for more information. [ 635.433395] env[61923]: ERROR nova.compute.manager [ 635.433395] env[61923]: Traceback (most recent call last): [ 635.433395] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 635.433395] env[61923]: listener.cb(fileno) [ 635.433395] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 635.433395] env[61923]: result = function(*args, **kwargs) [ 635.433395] env[61923]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 635.433395] env[61923]: return func(*args, **kwargs) [ 635.433395] env[61923]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 635.433395] env[61923]: raise e [ 635.433395] env[61923]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 635.433395] env[61923]: nwinfo = self.network_api.allocate_for_instance( [ 635.433395] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 635.433395] env[61923]: created_port_ids = self._update_ports_for_instance( [ 635.433395] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 635.433395] env[61923]: with excutils.save_and_reraise_exception(): [ 635.433395] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 635.433395] env[61923]: self.force_reraise() [ 635.433395] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 635.433395] env[61923]: raise self.value [ 635.433395] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 635.433395] env[61923]: updated_port = self._update_port( [ 635.433395] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 635.433395] env[61923]: _ensure_no_port_binding_failure(port) [ 635.433395] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 635.433395] env[61923]: raise exception.PortBindingFailed(port_id=port['id']) [ 635.434530] env[61923]: nova.exception.PortBindingFailed: Binding failed for port 07f299cc-2ca7-4beb-9f20-7509c49ed4f0, please check neutron logs for more information. [ 635.434530] env[61923]: Removing descriptor: 17 [ 635.434530] env[61923]: ERROR nova.compute.manager [None req-cb394400-3d4b-4149-a290-413641bb6f18 tempest-InstanceActionsV221TestJSON-1730893270 tempest-InstanceActionsV221TestJSON-1730893270-project-member] [instance: dc8cc22d-aba9-4dba-b456-1b323013d234] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 07f299cc-2ca7-4beb-9f20-7509c49ed4f0, please check neutron logs for more information. [ 635.434530] env[61923]: ERROR nova.compute.manager [instance: dc8cc22d-aba9-4dba-b456-1b323013d234] Traceback (most recent call last): [ 635.434530] env[61923]: ERROR nova.compute.manager [instance: dc8cc22d-aba9-4dba-b456-1b323013d234] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 635.434530] env[61923]: ERROR nova.compute.manager [instance: dc8cc22d-aba9-4dba-b456-1b323013d234] yield resources [ 635.434530] env[61923]: ERROR nova.compute.manager [instance: dc8cc22d-aba9-4dba-b456-1b323013d234] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 635.434530] env[61923]: ERROR nova.compute.manager [instance: dc8cc22d-aba9-4dba-b456-1b323013d234] self.driver.spawn(context, instance, image_meta, [ 635.434530] env[61923]: ERROR nova.compute.manager [instance: dc8cc22d-aba9-4dba-b456-1b323013d234] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 635.434530] env[61923]: ERROR nova.compute.manager [instance: dc8cc22d-aba9-4dba-b456-1b323013d234] self._vmops.spawn(context, instance, image_meta, injected_files, [ 635.434530] env[61923]: ERROR nova.compute.manager [instance: dc8cc22d-aba9-4dba-b456-1b323013d234] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 635.434530] env[61923]: ERROR nova.compute.manager [instance: dc8cc22d-aba9-4dba-b456-1b323013d234] vm_ref = self.build_virtual_machine(instance, [ 635.435027] env[61923]: ERROR nova.compute.manager [instance: dc8cc22d-aba9-4dba-b456-1b323013d234] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 635.435027] env[61923]: ERROR nova.compute.manager [instance: dc8cc22d-aba9-4dba-b456-1b323013d234] vif_infos = vmwarevif.get_vif_info(self._session, [ 635.435027] env[61923]: ERROR nova.compute.manager [instance: dc8cc22d-aba9-4dba-b456-1b323013d234] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 635.435027] env[61923]: ERROR nova.compute.manager [instance: dc8cc22d-aba9-4dba-b456-1b323013d234] for vif in network_info: [ 635.435027] env[61923]: ERROR nova.compute.manager [instance: dc8cc22d-aba9-4dba-b456-1b323013d234] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 635.435027] env[61923]: ERROR nova.compute.manager [instance: dc8cc22d-aba9-4dba-b456-1b323013d234] return self._sync_wrapper(fn, *args, **kwargs) [ 635.435027] env[61923]: ERROR nova.compute.manager [instance: dc8cc22d-aba9-4dba-b456-1b323013d234] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 635.435027] env[61923]: ERROR nova.compute.manager [instance: dc8cc22d-aba9-4dba-b456-1b323013d234] self.wait() [ 635.435027] env[61923]: ERROR nova.compute.manager [instance: dc8cc22d-aba9-4dba-b456-1b323013d234] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 635.435027] env[61923]: ERROR nova.compute.manager [instance: dc8cc22d-aba9-4dba-b456-1b323013d234] self[:] = self._gt.wait() [ 635.435027] env[61923]: ERROR nova.compute.manager [instance: dc8cc22d-aba9-4dba-b456-1b323013d234] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 635.435027] env[61923]: ERROR nova.compute.manager [instance: dc8cc22d-aba9-4dba-b456-1b323013d234] return self._exit_event.wait() [ 635.435027] env[61923]: ERROR nova.compute.manager [instance: dc8cc22d-aba9-4dba-b456-1b323013d234] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 635.436072] env[61923]: ERROR nova.compute.manager [instance: dc8cc22d-aba9-4dba-b456-1b323013d234] result = hub.switch() [ 635.436072] env[61923]: ERROR nova.compute.manager [instance: dc8cc22d-aba9-4dba-b456-1b323013d234] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 635.436072] env[61923]: ERROR nova.compute.manager [instance: dc8cc22d-aba9-4dba-b456-1b323013d234] return self.greenlet.switch() [ 635.436072] env[61923]: ERROR nova.compute.manager [instance: dc8cc22d-aba9-4dba-b456-1b323013d234] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 635.436072] env[61923]: ERROR nova.compute.manager [instance: dc8cc22d-aba9-4dba-b456-1b323013d234] result = function(*args, **kwargs) [ 635.436072] env[61923]: ERROR nova.compute.manager [instance: dc8cc22d-aba9-4dba-b456-1b323013d234] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 635.436072] env[61923]: ERROR nova.compute.manager [instance: dc8cc22d-aba9-4dba-b456-1b323013d234] return func(*args, **kwargs) [ 635.436072] env[61923]: ERROR nova.compute.manager [instance: dc8cc22d-aba9-4dba-b456-1b323013d234] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 635.436072] env[61923]: ERROR nova.compute.manager [instance: dc8cc22d-aba9-4dba-b456-1b323013d234] raise e [ 635.436072] env[61923]: ERROR nova.compute.manager [instance: dc8cc22d-aba9-4dba-b456-1b323013d234] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 635.436072] env[61923]: ERROR nova.compute.manager [instance: dc8cc22d-aba9-4dba-b456-1b323013d234] nwinfo = self.network_api.allocate_for_instance( [ 635.436072] env[61923]: ERROR nova.compute.manager [instance: dc8cc22d-aba9-4dba-b456-1b323013d234] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 635.436072] env[61923]: ERROR nova.compute.manager [instance: dc8cc22d-aba9-4dba-b456-1b323013d234] created_port_ids = self._update_ports_for_instance( [ 635.436573] env[61923]: ERROR nova.compute.manager [instance: dc8cc22d-aba9-4dba-b456-1b323013d234] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 635.436573] env[61923]: ERROR nova.compute.manager [instance: dc8cc22d-aba9-4dba-b456-1b323013d234] with excutils.save_and_reraise_exception(): [ 635.436573] env[61923]: ERROR nova.compute.manager [instance: dc8cc22d-aba9-4dba-b456-1b323013d234] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 635.436573] env[61923]: ERROR nova.compute.manager [instance: dc8cc22d-aba9-4dba-b456-1b323013d234] self.force_reraise() [ 635.436573] env[61923]: ERROR nova.compute.manager [instance: dc8cc22d-aba9-4dba-b456-1b323013d234] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 635.436573] env[61923]: ERROR nova.compute.manager [instance: dc8cc22d-aba9-4dba-b456-1b323013d234] raise self.value [ 635.436573] env[61923]: ERROR nova.compute.manager [instance: dc8cc22d-aba9-4dba-b456-1b323013d234] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 635.436573] env[61923]: ERROR nova.compute.manager [instance: dc8cc22d-aba9-4dba-b456-1b323013d234] updated_port = self._update_port( [ 635.436573] env[61923]: ERROR nova.compute.manager [instance: dc8cc22d-aba9-4dba-b456-1b323013d234] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 635.436573] env[61923]: ERROR nova.compute.manager [instance: dc8cc22d-aba9-4dba-b456-1b323013d234] _ensure_no_port_binding_failure(port) [ 635.436573] env[61923]: ERROR nova.compute.manager [instance: dc8cc22d-aba9-4dba-b456-1b323013d234] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 635.436573] env[61923]: ERROR nova.compute.manager [instance: dc8cc22d-aba9-4dba-b456-1b323013d234] raise exception.PortBindingFailed(port_id=port['id']) [ 635.436857] env[61923]: ERROR nova.compute.manager [instance: dc8cc22d-aba9-4dba-b456-1b323013d234] nova.exception.PortBindingFailed: Binding failed for port 07f299cc-2ca7-4beb-9f20-7509c49ed4f0, please check neutron logs for more information. [ 635.436857] env[61923]: ERROR nova.compute.manager [instance: dc8cc22d-aba9-4dba-b456-1b323013d234] [ 635.436857] env[61923]: INFO nova.compute.manager [None req-cb394400-3d4b-4149-a290-413641bb6f18 tempest-InstanceActionsV221TestJSON-1730893270 tempest-InstanceActionsV221TestJSON-1730893270-project-member] [instance: dc8cc22d-aba9-4dba-b456-1b323013d234] Terminating instance [ 635.436857] env[61923]: DEBUG oslo_concurrency.lockutils [None req-cb394400-3d4b-4149-a290-413641bb6f18 tempest-InstanceActionsV221TestJSON-1730893270 tempest-InstanceActionsV221TestJSON-1730893270-project-member] Acquiring lock "refresh_cache-dc8cc22d-aba9-4dba-b456-1b323013d234" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 635.436857] env[61923]: DEBUG oslo_concurrency.lockutils [None req-cb394400-3d4b-4149-a290-413641bb6f18 tempest-InstanceActionsV221TestJSON-1730893270 tempest-InstanceActionsV221TestJSON-1730893270-project-member] Acquired lock "refresh_cache-dc8cc22d-aba9-4dba-b456-1b323013d234" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 635.436857] env[61923]: DEBUG nova.network.neutron [None req-cb394400-3d4b-4149-a290-413641bb6f18 tempest-InstanceActionsV221TestJSON-1730893270 tempest-InstanceActionsV221TestJSON-1730893270-project-member] [instance: dc8cc22d-aba9-4dba-b456-1b323013d234] Building network info cache for instance {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 635.779011] env[61923]: DEBUG oslo_concurrency.lockutils [None req-753531fa-39a1-4c82-9915-d195bf0fe1b2 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.488s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 635.779623] env[61923]: DEBUG nova.compute.manager [None req-753531fa-39a1-4c82-9915-d195bf0fe1b2 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] [instance: 03100310-86ce-4739-8636-cd2aa1b542ff] Start building networks asynchronously for instance. {{(pid=61923) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 635.782281] env[61923]: DEBUG oslo_concurrency.lockutils [None req-b2fe22fb-da8f-41ba-8845-95fc6177390e tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.775s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 635.802463] env[61923]: DEBUG nova.network.neutron [-] [instance: a4b54a7a-3f3d-49ae-ba43-5a10c262d4e5] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 635.963303] env[61923]: DEBUG nova.network.neutron [None req-cb394400-3d4b-4149-a290-413641bb6f18 tempest-InstanceActionsV221TestJSON-1730893270 tempest-InstanceActionsV221TestJSON-1730893270-project-member] [instance: dc8cc22d-aba9-4dba-b456-1b323013d234] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 636.060093] env[61923]: DEBUG nova.network.neutron [None req-cb394400-3d4b-4149-a290-413641bb6f18 tempest-InstanceActionsV221TestJSON-1730893270 tempest-InstanceActionsV221TestJSON-1730893270-project-member] [instance: dc8cc22d-aba9-4dba-b456-1b323013d234] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 636.287081] env[61923]: DEBUG nova.compute.utils [None req-753531fa-39a1-4c82-9915-d195bf0fe1b2 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] Using /dev/sd instead of None {{(pid=61923) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 636.291151] env[61923]: DEBUG nova.compute.manager [None req-753531fa-39a1-4c82-9915-d195bf0fe1b2 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] [instance: 03100310-86ce-4739-8636-cd2aa1b542ff] Allocating IP information in the background. {{(pid=61923) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 636.291360] env[61923]: DEBUG nova.network.neutron [None req-753531fa-39a1-4c82-9915-d195bf0fe1b2 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] [instance: 03100310-86ce-4739-8636-cd2aa1b542ff] allocate_for_instance() {{(pid=61923) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 636.305102] env[61923]: INFO nova.compute.manager [-] [instance: a4b54a7a-3f3d-49ae-ba43-5a10c262d4e5] Took 1.03 seconds to deallocate network for instance. [ 636.307554] env[61923]: DEBUG nova.compute.claims [None req-95b9b70a-a835-4901-9a6d-c0fa1b2217f3 tempest-MigrationsAdminTest-1311988845 tempest-MigrationsAdminTest-1311988845-project-member] [instance: a4b54a7a-3f3d-49ae-ba43-5a10c262d4e5] Aborting claim: {{(pid=61923) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 636.307554] env[61923]: DEBUG oslo_concurrency.lockutils [None req-95b9b70a-a835-4901-9a6d-c0fa1b2217f3 tempest-MigrationsAdminTest-1311988845 tempest-MigrationsAdminTest-1311988845-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 636.339554] env[61923]: DEBUG nova.policy [None req-753531fa-39a1-4c82-9915-d195bf0fe1b2 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '629d5a54afa44599a4c0756d997cabc1', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '87f548d252244b759e3f157af9ab875d', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61923) authorize /opt/stack/nova/nova/policy.py:201}} [ 636.564791] env[61923]: DEBUG oslo_concurrency.lockutils [None req-cb394400-3d4b-4149-a290-413641bb6f18 tempest-InstanceActionsV221TestJSON-1730893270 tempest-InstanceActionsV221TestJSON-1730893270-project-member] Releasing lock "refresh_cache-dc8cc22d-aba9-4dba-b456-1b323013d234" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 636.565240] env[61923]: DEBUG nova.compute.manager [None req-cb394400-3d4b-4149-a290-413641bb6f18 tempest-InstanceActionsV221TestJSON-1730893270 tempest-InstanceActionsV221TestJSON-1730893270-project-member] [instance: dc8cc22d-aba9-4dba-b456-1b323013d234] Start destroying the instance on the hypervisor. {{(pid=61923) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 636.565442] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-cb394400-3d4b-4149-a290-413641bb6f18 tempest-InstanceActionsV221TestJSON-1730893270 tempest-InstanceActionsV221TestJSON-1730893270-project-member] [instance: dc8cc22d-aba9-4dba-b456-1b323013d234] Destroying instance {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 636.566258] env[61923]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-a6cb6b7f-f369-46a9-978e-76e184cb219c {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.574443] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f094661e-f8ac-4e40-882f-53ca6a38ea94 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.599452] env[61923]: WARNING nova.virt.vmwareapi.vmops [None req-cb394400-3d4b-4149-a290-413641bb6f18 tempest-InstanceActionsV221TestJSON-1730893270 tempest-InstanceActionsV221TestJSON-1730893270-project-member] [instance: dc8cc22d-aba9-4dba-b456-1b323013d234] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance dc8cc22d-aba9-4dba-b456-1b323013d234 could not be found. [ 636.599723] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-cb394400-3d4b-4149-a290-413641bb6f18 tempest-InstanceActionsV221TestJSON-1730893270 tempest-InstanceActionsV221TestJSON-1730893270-project-member] [instance: dc8cc22d-aba9-4dba-b456-1b323013d234] Instance destroyed {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 636.600107] env[61923]: INFO nova.compute.manager [None req-cb394400-3d4b-4149-a290-413641bb6f18 tempest-InstanceActionsV221TestJSON-1730893270 tempest-InstanceActionsV221TestJSON-1730893270-project-member] [instance: dc8cc22d-aba9-4dba-b456-1b323013d234] Took 0.03 seconds to destroy the instance on the hypervisor. [ 636.600175] env[61923]: DEBUG oslo.service.loopingcall [None req-cb394400-3d4b-4149-a290-413641bb6f18 tempest-InstanceActionsV221TestJSON-1730893270 tempest-InstanceActionsV221TestJSON-1730893270-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61923) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 636.602625] env[61923]: DEBUG nova.compute.manager [-] [instance: dc8cc22d-aba9-4dba-b456-1b323013d234] Deallocating network for instance {{(pid=61923) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 636.602724] env[61923]: DEBUG nova.network.neutron [-] [instance: dc8cc22d-aba9-4dba-b456-1b323013d234] deallocate_for_instance() {{(pid=61923) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 636.614510] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c85f72c-bd6a-40a4-8bbc-9f2ca3e4fee3 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.621418] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33d886b4-68c8-43b0-914d-b0e19958ae7f {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.656846] env[61923]: DEBUG nova.network.neutron [-] [instance: dc8cc22d-aba9-4dba-b456-1b323013d234] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 636.658783] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-782bdcf1-9d0c-4fcf-9866-d4927eaec2cb {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.667300] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3830ff7-b627-4760-a0cc-68bcb7416f04 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.682484] env[61923]: DEBUG nova.compute.provider_tree [None req-b2fe22fb-da8f-41ba-8845-95fc6177390e tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 636.756943] env[61923]: DEBUG nova.network.neutron [None req-753531fa-39a1-4c82-9915-d195bf0fe1b2 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] [instance: 03100310-86ce-4739-8636-cd2aa1b542ff] Successfully created port: 5a9feebd-c914-422f-af20-45e4709470e8 {{(pid=61923) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 636.797028] env[61923]: DEBUG nova.compute.manager [None req-753531fa-39a1-4c82-9915-d195bf0fe1b2 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] [instance: 03100310-86ce-4739-8636-cd2aa1b542ff] Start building block device mappings for instance. {{(pid=61923) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 637.090475] env[61923]: DEBUG nova.compute.manager [req-4b13a9a5-f2bb-496b-9399-f80db8362d64 req-b17e2a7d-ac0b-498b-ac78-c1fdfe4f52ec service nova] [instance: dc8cc22d-aba9-4dba-b456-1b323013d234] Received event network-changed-07f299cc-2ca7-4beb-9f20-7509c49ed4f0 {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 637.090553] env[61923]: DEBUG nova.compute.manager [req-4b13a9a5-f2bb-496b-9399-f80db8362d64 req-b17e2a7d-ac0b-498b-ac78-c1fdfe4f52ec service nova] [instance: dc8cc22d-aba9-4dba-b456-1b323013d234] Refreshing instance network info cache due to event network-changed-07f299cc-2ca7-4beb-9f20-7509c49ed4f0. {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 637.090737] env[61923]: DEBUG oslo_concurrency.lockutils [req-4b13a9a5-f2bb-496b-9399-f80db8362d64 req-b17e2a7d-ac0b-498b-ac78-c1fdfe4f52ec service nova] Acquiring lock "refresh_cache-dc8cc22d-aba9-4dba-b456-1b323013d234" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 637.090882] env[61923]: DEBUG oslo_concurrency.lockutils [req-4b13a9a5-f2bb-496b-9399-f80db8362d64 req-b17e2a7d-ac0b-498b-ac78-c1fdfe4f52ec service nova] Acquired lock "refresh_cache-dc8cc22d-aba9-4dba-b456-1b323013d234" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 637.091433] env[61923]: DEBUG nova.network.neutron [req-4b13a9a5-f2bb-496b-9399-f80db8362d64 req-b17e2a7d-ac0b-498b-ac78-c1fdfe4f52ec service nova] [instance: dc8cc22d-aba9-4dba-b456-1b323013d234] Refreshing network info cache for port 07f299cc-2ca7-4beb-9f20-7509c49ed4f0 {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 637.164108] env[61923]: DEBUG nova.network.neutron [-] [instance: dc8cc22d-aba9-4dba-b456-1b323013d234] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 637.185093] env[61923]: DEBUG nova.scheduler.client.report [None req-b2fe22fb-da8f-41ba-8845-95fc6177390e tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 637.667130] env[61923]: INFO nova.compute.manager [-] [instance: dc8cc22d-aba9-4dba-b456-1b323013d234] Took 1.06 seconds to deallocate network for instance. [ 637.670454] env[61923]: DEBUG nova.compute.claims [None req-cb394400-3d4b-4149-a290-413641bb6f18 tempest-InstanceActionsV221TestJSON-1730893270 tempest-InstanceActionsV221TestJSON-1730893270-project-member] [instance: dc8cc22d-aba9-4dba-b456-1b323013d234] Aborting claim: {{(pid=61923) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 637.670705] env[61923]: DEBUG oslo_concurrency.lockutils [None req-cb394400-3d4b-4149-a290-413641bb6f18 tempest-InstanceActionsV221TestJSON-1730893270 tempest-InstanceActionsV221TestJSON-1730893270-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 637.690673] env[61923]: DEBUG oslo_concurrency.lockutils [None req-b2fe22fb-da8f-41ba-8845-95fc6177390e tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.908s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 637.691921] env[61923]: ERROR nova.compute.manager [None req-b2fe22fb-da8f-41ba-8845-95fc6177390e tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] [instance: 119bb4a7-2d71-44d7-ae81-9238512937c8] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 37c91d0a-5c78-46e0-800e-9c9760eb9fbc, please check neutron logs for more information. [ 637.691921] env[61923]: ERROR nova.compute.manager [instance: 119bb4a7-2d71-44d7-ae81-9238512937c8] Traceback (most recent call last): [ 637.691921] env[61923]: ERROR nova.compute.manager [instance: 119bb4a7-2d71-44d7-ae81-9238512937c8] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 637.691921] env[61923]: ERROR nova.compute.manager [instance: 119bb4a7-2d71-44d7-ae81-9238512937c8] self.driver.spawn(context, instance, image_meta, [ 637.691921] env[61923]: ERROR nova.compute.manager [instance: 119bb4a7-2d71-44d7-ae81-9238512937c8] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 637.691921] env[61923]: ERROR nova.compute.manager [instance: 119bb4a7-2d71-44d7-ae81-9238512937c8] self._vmops.spawn(context, instance, image_meta, injected_files, [ 637.691921] env[61923]: ERROR nova.compute.manager [instance: 119bb4a7-2d71-44d7-ae81-9238512937c8] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 637.691921] env[61923]: ERROR nova.compute.manager [instance: 119bb4a7-2d71-44d7-ae81-9238512937c8] vm_ref = self.build_virtual_machine(instance, [ 637.691921] env[61923]: ERROR nova.compute.manager [instance: 119bb4a7-2d71-44d7-ae81-9238512937c8] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 637.691921] env[61923]: ERROR nova.compute.manager [instance: 119bb4a7-2d71-44d7-ae81-9238512937c8] vif_infos = vmwarevif.get_vif_info(self._session, [ 637.691921] env[61923]: ERROR nova.compute.manager [instance: 119bb4a7-2d71-44d7-ae81-9238512937c8] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 637.692399] env[61923]: ERROR nova.compute.manager [instance: 119bb4a7-2d71-44d7-ae81-9238512937c8] for vif in network_info: [ 637.692399] env[61923]: ERROR nova.compute.manager [instance: 119bb4a7-2d71-44d7-ae81-9238512937c8] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 637.692399] env[61923]: ERROR nova.compute.manager [instance: 119bb4a7-2d71-44d7-ae81-9238512937c8] return self._sync_wrapper(fn, *args, **kwargs) [ 637.692399] env[61923]: ERROR nova.compute.manager [instance: 119bb4a7-2d71-44d7-ae81-9238512937c8] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 637.692399] env[61923]: ERROR nova.compute.manager [instance: 119bb4a7-2d71-44d7-ae81-9238512937c8] self.wait() [ 637.692399] env[61923]: ERROR nova.compute.manager [instance: 119bb4a7-2d71-44d7-ae81-9238512937c8] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 637.692399] env[61923]: ERROR nova.compute.manager [instance: 119bb4a7-2d71-44d7-ae81-9238512937c8] self[:] = self._gt.wait() [ 637.692399] env[61923]: ERROR nova.compute.manager [instance: 119bb4a7-2d71-44d7-ae81-9238512937c8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 637.692399] env[61923]: ERROR nova.compute.manager [instance: 119bb4a7-2d71-44d7-ae81-9238512937c8] return self._exit_event.wait() [ 637.692399] env[61923]: ERROR nova.compute.manager [instance: 119bb4a7-2d71-44d7-ae81-9238512937c8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 637.692399] env[61923]: ERROR nova.compute.manager [instance: 119bb4a7-2d71-44d7-ae81-9238512937c8] result = hub.switch() [ 637.692399] env[61923]: ERROR nova.compute.manager [instance: 119bb4a7-2d71-44d7-ae81-9238512937c8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 637.692399] env[61923]: ERROR nova.compute.manager [instance: 119bb4a7-2d71-44d7-ae81-9238512937c8] return self.greenlet.switch() [ 637.692811] env[61923]: ERROR nova.compute.manager [instance: 119bb4a7-2d71-44d7-ae81-9238512937c8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 637.692811] env[61923]: ERROR nova.compute.manager [instance: 119bb4a7-2d71-44d7-ae81-9238512937c8] result = function(*args, **kwargs) [ 637.692811] env[61923]: ERROR nova.compute.manager [instance: 119bb4a7-2d71-44d7-ae81-9238512937c8] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 637.692811] env[61923]: ERROR nova.compute.manager [instance: 119bb4a7-2d71-44d7-ae81-9238512937c8] return func(*args, **kwargs) [ 637.692811] env[61923]: ERROR nova.compute.manager [instance: 119bb4a7-2d71-44d7-ae81-9238512937c8] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 637.692811] env[61923]: ERROR nova.compute.manager [instance: 119bb4a7-2d71-44d7-ae81-9238512937c8] raise e [ 637.692811] env[61923]: ERROR nova.compute.manager [instance: 119bb4a7-2d71-44d7-ae81-9238512937c8] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 637.692811] env[61923]: ERROR nova.compute.manager [instance: 119bb4a7-2d71-44d7-ae81-9238512937c8] nwinfo = self.network_api.allocate_for_instance( [ 637.692811] env[61923]: ERROR nova.compute.manager [instance: 119bb4a7-2d71-44d7-ae81-9238512937c8] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 637.692811] env[61923]: ERROR nova.compute.manager [instance: 119bb4a7-2d71-44d7-ae81-9238512937c8] created_port_ids = self._update_ports_for_instance( [ 637.692811] env[61923]: ERROR nova.compute.manager [instance: 119bb4a7-2d71-44d7-ae81-9238512937c8] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 637.692811] env[61923]: ERROR nova.compute.manager [instance: 119bb4a7-2d71-44d7-ae81-9238512937c8] with excutils.save_and_reraise_exception(): [ 637.692811] env[61923]: ERROR nova.compute.manager [instance: 119bb4a7-2d71-44d7-ae81-9238512937c8] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 637.694571] env[61923]: ERROR nova.compute.manager [instance: 119bb4a7-2d71-44d7-ae81-9238512937c8] self.force_reraise() [ 637.694571] env[61923]: ERROR nova.compute.manager [instance: 119bb4a7-2d71-44d7-ae81-9238512937c8] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 637.694571] env[61923]: ERROR nova.compute.manager [instance: 119bb4a7-2d71-44d7-ae81-9238512937c8] raise self.value [ 637.694571] env[61923]: ERROR nova.compute.manager [instance: 119bb4a7-2d71-44d7-ae81-9238512937c8] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 637.694571] env[61923]: ERROR nova.compute.manager [instance: 119bb4a7-2d71-44d7-ae81-9238512937c8] updated_port = self._update_port( [ 637.694571] env[61923]: ERROR nova.compute.manager [instance: 119bb4a7-2d71-44d7-ae81-9238512937c8] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 637.694571] env[61923]: ERROR nova.compute.manager [instance: 119bb4a7-2d71-44d7-ae81-9238512937c8] _ensure_no_port_binding_failure(port) [ 637.694571] env[61923]: ERROR nova.compute.manager [instance: 119bb4a7-2d71-44d7-ae81-9238512937c8] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 637.694571] env[61923]: ERROR nova.compute.manager [instance: 119bb4a7-2d71-44d7-ae81-9238512937c8] raise exception.PortBindingFailed(port_id=port['id']) [ 637.694571] env[61923]: ERROR nova.compute.manager [instance: 119bb4a7-2d71-44d7-ae81-9238512937c8] nova.exception.PortBindingFailed: Binding failed for port 37c91d0a-5c78-46e0-800e-9c9760eb9fbc, please check neutron logs for more information. [ 637.694571] env[61923]: ERROR nova.compute.manager [instance: 119bb4a7-2d71-44d7-ae81-9238512937c8] [ 637.694928] env[61923]: DEBUG nova.compute.utils [None req-b2fe22fb-da8f-41ba-8845-95fc6177390e tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] [instance: 119bb4a7-2d71-44d7-ae81-9238512937c8] Binding failed for port 37c91d0a-5c78-46e0-800e-9c9760eb9fbc, please check neutron logs for more information. {{(pid=61923) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 637.694928] env[61923]: DEBUG oslo_concurrency.lockutils [None req-76010439-9fb6-4296-9b5a-107842db76db tempest-ServerExternalEventsTest-1479796458 tempest-ServerExternalEventsTest-1479796458-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.419s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 637.694928] env[61923]: INFO nova.compute.claims [None req-76010439-9fb6-4296-9b5a-107842db76db tempest-ServerExternalEventsTest-1479796458 tempest-ServerExternalEventsTest-1479796458-project-member] [instance: c0f87aa5-447b-4920-8251-354d89a14fbb] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 637.700017] env[61923]: DEBUG nova.compute.manager [None req-b2fe22fb-da8f-41ba-8845-95fc6177390e tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] [instance: 119bb4a7-2d71-44d7-ae81-9238512937c8] Build of instance 119bb4a7-2d71-44d7-ae81-9238512937c8 was re-scheduled: Binding failed for port 37c91d0a-5c78-46e0-800e-9c9760eb9fbc, please check neutron logs for more information. {{(pid=61923) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 637.700017] env[61923]: DEBUG nova.compute.manager [None req-b2fe22fb-da8f-41ba-8845-95fc6177390e tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] [instance: 119bb4a7-2d71-44d7-ae81-9238512937c8] Unplugging VIFs for instance {{(pid=61923) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 637.700017] env[61923]: DEBUG oslo_concurrency.lockutils [None req-b2fe22fb-da8f-41ba-8845-95fc6177390e tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Acquiring lock "refresh_cache-119bb4a7-2d71-44d7-ae81-9238512937c8" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 637.700017] env[61923]: DEBUG oslo_concurrency.lockutils [None req-b2fe22fb-da8f-41ba-8845-95fc6177390e tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Acquired lock "refresh_cache-119bb4a7-2d71-44d7-ae81-9238512937c8" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 637.700017] env[61923]: DEBUG nova.network.neutron [None req-b2fe22fb-da8f-41ba-8845-95fc6177390e tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] [instance: 119bb4a7-2d71-44d7-ae81-9238512937c8] Building network info cache for instance {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 637.709693] env[61923]: DEBUG nova.network.neutron [req-4b13a9a5-f2bb-496b-9399-f80db8362d64 req-b17e2a7d-ac0b-498b-ac78-c1fdfe4f52ec service nova] [instance: dc8cc22d-aba9-4dba-b456-1b323013d234] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 637.804933] env[61923]: DEBUG nova.compute.manager [None req-753531fa-39a1-4c82-9915-d195bf0fe1b2 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] [instance: 03100310-86ce-4739-8636-cd2aa1b542ff] Start spawning the instance on the hypervisor. {{(pid=61923) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 637.843046] env[61923]: DEBUG nova.virt.hardware [None req-753531fa-39a1-4c82-9915-d195bf0fe1b2 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-29T20:07:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-29T20:07:35Z,direct_url=,disk_format='vmdk',id=0f153f63-ae0a-45b1-b7f5-7f9b673c947f,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='4e7b5e3b3c3443c8bd5c70f8a15739f5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-29T20:07:36Z,virtual_size=,visibility=), allow threads: False {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 637.843046] env[61923]: DEBUG nova.virt.hardware [None req-753531fa-39a1-4c82-9915-d195bf0fe1b2 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] Flavor limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 637.843046] env[61923]: DEBUG nova.virt.hardware [None req-753531fa-39a1-4c82-9915-d195bf0fe1b2 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] Image limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 637.843371] env[61923]: DEBUG nova.virt.hardware [None req-753531fa-39a1-4c82-9915-d195bf0fe1b2 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] Flavor pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 637.843371] env[61923]: DEBUG nova.virt.hardware [None req-753531fa-39a1-4c82-9915-d195bf0fe1b2 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] Image pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 637.843371] env[61923]: DEBUG nova.virt.hardware [None req-753531fa-39a1-4c82-9915-d195bf0fe1b2 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 637.843371] env[61923]: DEBUG nova.virt.hardware [None req-753531fa-39a1-4c82-9915-d195bf0fe1b2 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 637.843371] env[61923]: DEBUG nova.virt.hardware [None req-753531fa-39a1-4c82-9915-d195bf0fe1b2 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 637.843529] env[61923]: DEBUG nova.virt.hardware [None req-753531fa-39a1-4c82-9915-d195bf0fe1b2 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] Got 1 possible topologies {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 637.843697] env[61923]: DEBUG nova.virt.hardware [None req-753531fa-39a1-4c82-9915-d195bf0fe1b2 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 637.844068] env[61923]: DEBUG nova.virt.hardware [None req-753531fa-39a1-4c82-9915-d195bf0fe1b2 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 637.845305] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-749ce4aa-a2d0-4108-b195-ea286dcbbf97 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.856160] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a67a57d9-56c9-4874-918c-e94cc62cf22c {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.903579] env[61923]: DEBUG nova.network.neutron [req-4b13a9a5-f2bb-496b-9399-f80db8362d64 req-b17e2a7d-ac0b-498b-ac78-c1fdfe4f52ec service nova] [instance: dc8cc22d-aba9-4dba-b456-1b323013d234] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 638.130110] env[61923]: ERROR nova.compute.manager [None req-753531fa-39a1-4c82-9915-d195bf0fe1b2 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 5a9feebd-c914-422f-af20-45e4709470e8, please check neutron logs for more information. [ 638.130110] env[61923]: ERROR nova.compute.manager Traceback (most recent call last): [ 638.130110] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 638.130110] env[61923]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 638.130110] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 638.130110] env[61923]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 638.130110] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 638.130110] env[61923]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 638.130110] env[61923]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 638.130110] env[61923]: ERROR nova.compute.manager self.force_reraise() [ 638.130110] env[61923]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 638.130110] env[61923]: ERROR nova.compute.manager raise self.value [ 638.130110] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 638.130110] env[61923]: ERROR nova.compute.manager updated_port = self._update_port( [ 638.130110] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 638.130110] env[61923]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 638.130586] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 638.130586] env[61923]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 638.130586] env[61923]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 5a9feebd-c914-422f-af20-45e4709470e8, please check neutron logs for more information. [ 638.130586] env[61923]: ERROR nova.compute.manager [ 638.130586] env[61923]: Traceback (most recent call last): [ 638.130586] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 638.130586] env[61923]: listener.cb(fileno) [ 638.130586] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 638.130586] env[61923]: result = function(*args, **kwargs) [ 638.130586] env[61923]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 638.130586] env[61923]: return func(*args, **kwargs) [ 638.130586] env[61923]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 638.130586] env[61923]: raise e [ 638.130586] env[61923]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 638.130586] env[61923]: nwinfo = self.network_api.allocate_for_instance( [ 638.130586] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 638.130586] env[61923]: created_port_ids = self._update_ports_for_instance( [ 638.130586] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 638.130586] env[61923]: with excutils.save_and_reraise_exception(): [ 638.130586] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 638.130586] env[61923]: self.force_reraise() [ 638.130586] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 638.130586] env[61923]: raise self.value [ 638.130586] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 638.130586] env[61923]: updated_port = self._update_port( [ 638.130586] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 638.130586] env[61923]: _ensure_no_port_binding_failure(port) [ 638.130586] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 638.130586] env[61923]: raise exception.PortBindingFailed(port_id=port['id']) [ 638.131373] env[61923]: nova.exception.PortBindingFailed: Binding failed for port 5a9feebd-c914-422f-af20-45e4709470e8, please check neutron logs for more information. [ 638.131373] env[61923]: Removing descriptor: 17 [ 638.131373] env[61923]: ERROR nova.compute.manager [None req-753531fa-39a1-4c82-9915-d195bf0fe1b2 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] [instance: 03100310-86ce-4739-8636-cd2aa1b542ff] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 5a9feebd-c914-422f-af20-45e4709470e8, please check neutron logs for more information. [ 638.131373] env[61923]: ERROR nova.compute.manager [instance: 03100310-86ce-4739-8636-cd2aa1b542ff] Traceback (most recent call last): [ 638.131373] env[61923]: ERROR nova.compute.manager [instance: 03100310-86ce-4739-8636-cd2aa1b542ff] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 638.131373] env[61923]: ERROR nova.compute.manager [instance: 03100310-86ce-4739-8636-cd2aa1b542ff] yield resources [ 638.131373] env[61923]: ERROR nova.compute.manager [instance: 03100310-86ce-4739-8636-cd2aa1b542ff] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 638.131373] env[61923]: ERROR nova.compute.manager [instance: 03100310-86ce-4739-8636-cd2aa1b542ff] self.driver.spawn(context, instance, image_meta, [ 638.131373] env[61923]: ERROR nova.compute.manager [instance: 03100310-86ce-4739-8636-cd2aa1b542ff] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 638.131373] env[61923]: ERROR nova.compute.manager [instance: 03100310-86ce-4739-8636-cd2aa1b542ff] self._vmops.spawn(context, instance, image_meta, injected_files, [ 638.131373] env[61923]: ERROR nova.compute.manager [instance: 03100310-86ce-4739-8636-cd2aa1b542ff] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 638.131373] env[61923]: ERROR nova.compute.manager [instance: 03100310-86ce-4739-8636-cd2aa1b542ff] vm_ref = self.build_virtual_machine(instance, [ 638.131706] env[61923]: ERROR nova.compute.manager [instance: 03100310-86ce-4739-8636-cd2aa1b542ff] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 638.131706] env[61923]: ERROR nova.compute.manager [instance: 03100310-86ce-4739-8636-cd2aa1b542ff] vif_infos = vmwarevif.get_vif_info(self._session, [ 638.131706] env[61923]: ERROR nova.compute.manager [instance: 03100310-86ce-4739-8636-cd2aa1b542ff] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 638.131706] env[61923]: ERROR nova.compute.manager [instance: 03100310-86ce-4739-8636-cd2aa1b542ff] for vif in network_info: [ 638.131706] env[61923]: ERROR nova.compute.manager [instance: 03100310-86ce-4739-8636-cd2aa1b542ff] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 638.131706] env[61923]: ERROR nova.compute.manager [instance: 03100310-86ce-4739-8636-cd2aa1b542ff] return self._sync_wrapper(fn, *args, **kwargs) [ 638.131706] env[61923]: ERROR nova.compute.manager [instance: 03100310-86ce-4739-8636-cd2aa1b542ff] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 638.131706] env[61923]: ERROR nova.compute.manager [instance: 03100310-86ce-4739-8636-cd2aa1b542ff] self.wait() [ 638.131706] env[61923]: ERROR nova.compute.manager [instance: 03100310-86ce-4739-8636-cd2aa1b542ff] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 638.131706] env[61923]: ERROR nova.compute.manager [instance: 03100310-86ce-4739-8636-cd2aa1b542ff] self[:] = self._gt.wait() [ 638.131706] env[61923]: ERROR nova.compute.manager [instance: 03100310-86ce-4739-8636-cd2aa1b542ff] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 638.131706] env[61923]: ERROR nova.compute.manager [instance: 03100310-86ce-4739-8636-cd2aa1b542ff] return self._exit_event.wait() [ 638.131706] env[61923]: ERROR nova.compute.manager [instance: 03100310-86ce-4739-8636-cd2aa1b542ff] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 638.132085] env[61923]: ERROR nova.compute.manager [instance: 03100310-86ce-4739-8636-cd2aa1b542ff] result = hub.switch() [ 638.132085] env[61923]: ERROR nova.compute.manager [instance: 03100310-86ce-4739-8636-cd2aa1b542ff] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 638.132085] env[61923]: ERROR nova.compute.manager [instance: 03100310-86ce-4739-8636-cd2aa1b542ff] return self.greenlet.switch() [ 638.132085] env[61923]: ERROR nova.compute.manager [instance: 03100310-86ce-4739-8636-cd2aa1b542ff] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 638.132085] env[61923]: ERROR nova.compute.manager [instance: 03100310-86ce-4739-8636-cd2aa1b542ff] result = function(*args, **kwargs) [ 638.132085] env[61923]: ERROR nova.compute.manager [instance: 03100310-86ce-4739-8636-cd2aa1b542ff] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 638.132085] env[61923]: ERROR nova.compute.manager [instance: 03100310-86ce-4739-8636-cd2aa1b542ff] return func(*args, **kwargs) [ 638.132085] env[61923]: ERROR nova.compute.manager [instance: 03100310-86ce-4739-8636-cd2aa1b542ff] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 638.132085] env[61923]: ERROR nova.compute.manager [instance: 03100310-86ce-4739-8636-cd2aa1b542ff] raise e [ 638.132085] env[61923]: ERROR nova.compute.manager [instance: 03100310-86ce-4739-8636-cd2aa1b542ff] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 638.132085] env[61923]: ERROR nova.compute.manager [instance: 03100310-86ce-4739-8636-cd2aa1b542ff] nwinfo = self.network_api.allocate_for_instance( [ 638.132085] env[61923]: ERROR nova.compute.manager [instance: 03100310-86ce-4739-8636-cd2aa1b542ff] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 638.132085] env[61923]: ERROR nova.compute.manager [instance: 03100310-86ce-4739-8636-cd2aa1b542ff] created_port_ids = self._update_ports_for_instance( [ 638.132446] env[61923]: ERROR nova.compute.manager [instance: 03100310-86ce-4739-8636-cd2aa1b542ff] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 638.132446] env[61923]: ERROR nova.compute.manager [instance: 03100310-86ce-4739-8636-cd2aa1b542ff] with excutils.save_and_reraise_exception(): [ 638.132446] env[61923]: ERROR nova.compute.manager [instance: 03100310-86ce-4739-8636-cd2aa1b542ff] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 638.132446] env[61923]: ERROR nova.compute.manager [instance: 03100310-86ce-4739-8636-cd2aa1b542ff] self.force_reraise() [ 638.132446] env[61923]: ERROR nova.compute.manager [instance: 03100310-86ce-4739-8636-cd2aa1b542ff] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 638.132446] env[61923]: ERROR nova.compute.manager [instance: 03100310-86ce-4739-8636-cd2aa1b542ff] raise self.value [ 638.132446] env[61923]: ERROR nova.compute.manager [instance: 03100310-86ce-4739-8636-cd2aa1b542ff] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 638.132446] env[61923]: ERROR nova.compute.manager [instance: 03100310-86ce-4739-8636-cd2aa1b542ff] updated_port = self._update_port( [ 638.132446] env[61923]: ERROR nova.compute.manager [instance: 03100310-86ce-4739-8636-cd2aa1b542ff] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 638.132446] env[61923]: ERROR nova.compute.manager [instance: 03100310-86ce-4739-8636-cd2aa1b542ff] _ensure_no_port_binding_failure(port) [ 638.132446] env[61923]: ERROR nova.compute.manager [instance: 03100310-86ce-4739-8636-cd2aa1b542ff] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 638.132446] env[61923]: ERROR nova.compute.manager [instance: 03100310-86ce-4739-8636-cd2aa1b542ff] raise exception.PortBindingFailed(port_id=port['id']) [ 638.132775] env[61923]: ERROR nova.compute.manager [instance: 03100310-86ce-4739-8636-cd2aa1b542ff] nova.exception.PortBindingFailed: Binding failed for port 5a9feebd-c914-422f-af20-45e4709470e8, please check neutron logs for more information. [ 638.132775] env[61923]: ERROR nova.compute.manager [instance: 03100310-86ce-4739-8636-cd2aa1b542ff] [ 638.132775] env[61923]: INFO nova.compute.manager [None req-753531fa-39a1-4c82-9915-d195bf0fe1b2 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] [instance: 03100310-86ce-4739-8636-cd2aa1b542ff] Terminating instance [ 638.135179] env[61923]: DEBUG oslo_concurrency.lockutils [None req-753531fa-39a1-4c82-9915-d195bf0fe1b2 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] Acquiring lock "refresh_cache-03100310-86ce-4739-8636-cd2aa1b542ff" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 638.135179] env[61923]: DEBUG oslo_concurrency.lockutils [None req-753531fa-39a1-4c82-9915-d195bf0fe1b2 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] Acquired lock "refresh_cache-03100310-86ce-4739-8636-cd2aa1b542ff" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 638.135179] env[61923]: DEBUG nova.network.neutron [None req-753531fa-39a1-4c82-9915-d195bf0fe1b2 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] [instance: 03100310-86ce-4739-8636-cd2aa1b542ff] Building network info cache for instance {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 638.216689] env[61923]: DEBUG nova.network.neutron [None req-b2fe22fb-da8f-41ba-8845-95fc6177390e tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] [instance: 119bb4a7-2d71-44d7-ae81-9238512937c8] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 638.294501] env[61923]: DEBUG nova.network.neutron [None req-b2fe22fb-da8f-41ba-8845-95fc6177390e tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] [instance: 119bb4a7-2d71-44d7-ae81-9238512937c8] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 638.406553] env[61923]: DEBUG oslo_concurrency.lockutils [req-4b13a9a5-f2bb-496b-9399-f80db8362d64 req-b17e2a7d-ac0b-498b-ac78-c1fdfe4f52ec service nova] Releasing lock "refresh_cache-dc8cc22d-aba9-4dba-b456-1b323013d234" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 638.406818] env[61923]: DEBUG nova.compute.manager [req-4b13a9a5-f2bb-496b-9399-f80db8362d64 req-b17e2a7d-ac0b-498b-ac78-c1fdfe4f52ec service nova] [instance: dc8cc22d-aba9-4dba-b456-1b323013d234] Received event network-vif-deleted-07f299cc-2ca7-4beb-9f20-7509c49ed4f0 {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 638.652566] env[61923]: DEBUG nova.network.neutron [None req-753531fa-39a1-4c82-9915-d195bf0fe1b2 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] [instance: 03100310-86ce-4739-8636-cd2aa1b542ff] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 638.725534] env[61923]: DEBUG nova.network.neutron [None req-753531fa-39a1-4c82-9915-d195bf0fe1b2 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] [instance: 03100310-86ce-4739-8636-cd2aa1b542ff] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 638.796666] env[61923]: DEBUG oslo_concurrency.lockutils [None req-b2fe22fb-da8f-41ba-8845-95fc6177390e tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Releasing lock "refresh_cache-119bb4a7-2d71-44d7-ae81-9238512937c8" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 638.796905] env[61923]: DEBUG nova.compute.manager [None req-b2fe22fb-da8f-41ba-8845-95fc6177390e tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61923) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 638.797098] env[61923]: DEBUG nova.compute.manager [None req-b2fe22fb-da8f-41ba-8845-95fc6177390e tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] [instance: 119bb4a7-2d71-44d7-ae81-9238512937c8] Deallocating network for instance {{(pid=61923) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 638.797387] env[61923]: DEBUG nova.network.neutron [None req-b2fe22fb-da8f-41ba-8845-95fc6177390e tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] [instance: 119bb4a7-2d71-44d7-ae81-9238512937c8] deallocate_for_instance() {{(pid=61923) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 638.812160] env[61923]: DEBUG nova.network.neutron [None req-b2fe22fb-da8f-41ba-8845-95fc6177390e tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] [instance: 119bb4a7-2d71-44d7-ae81-9238512937c8] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 639.006514] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3f4a241-3165-4112-ac8b-423a4c42490e {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.013500] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9fd402bd-c048-4a22-80e5-437519567930 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.042411] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4b5147e-1b40-4627-8f5b-3aff462355ea {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.049527] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-451f87aa-d290-4240-bbff-b4a6287c38a6 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.062306] env[61923]: DEBUG nova.compute.provider_tree [None req-76010439-9fb6-4296-9b5a-107842db76db tempest-ServerExternalEventsTest-1479796458 tempest-ServerExternalEventsTest-1479796458-project-member] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 639.133678] env[61923]: DEBUG nova.compute.manager [req-bf606cee-4fb6-4927-859a-74b721741096 req-3bc7fc9a-00b1-4476-b122-18481390b128 service nova] [instance: 03100310-86ce-4739-8636-cd2aa1b542ff] Received event network-changed-5a9feebd-c914-422f-af20-45e4709470e8 {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 639.133678] env[61923]: DEBUG nova.compute.manager [req-bf606cee-4fb6-4927-859a-74b721741096 req-3bc7fc9a-00b1-4476-b122-18481390b128 service nova] [instance: 03100310-86ce-4739-8636-cd2aa1b542ff] Refreshing instance network info cache due to event network-changed-5a9feebd-c914-422f-af20-45e4709470e8. {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 639.133678] env[61923]: DEBUG oslo_concurrency.lockutils [req-bf606cee-4fb6-4927-859a-74b721741096 req-3bc7fc9a-00b1-4476-b122-18481390b128 service nova] Acquiring lock "refresh_cache-03100310-86ce-4739-8636-cd2aa1b542ff" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 639.230051] env[61923]: DEBUG oslo_concurrency.lockutils [None req-753531fa-39a1-4c82-9915-d195bf0fe1b2 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] Releasing lock "refresh_cache-03100310-86ce-4739-8636-cd2aa1b542ff" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 639.230501] env[61923]: DEBUG nova.compute.manager [None req-753531fa-39a1-4c82-9915-d195bf0fe1b2 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] [instance: 03100310-86ce-4739-8636-cd2aa1b542ff] Start destroying the instance on the hypervisor. {{(pid=61923) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 639.230699] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-753531fa-39a1-4c82-9915-d195bf0fe1b2 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] [instance: 03100310-86ce-4739-8636-cd2aa1b542ff] Destroying instance {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 639.231021] env[61923]: DEBUG oslo_concurrency.lockutils [req-bf606cee-4fb6-4927-859a-74b721741096 req-3bc7fc9a-00b1-4476-b122-18481390b128 service nova] Acquired lock "refresh_cache-03100310-86ce-4739-8636-cd2aa1b542ff" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 639.231197] env[61923]: DEBUG nova.network.neutron [req-bf606cee-4fb6-4927-859a-74b721741096 req-3bc7fc9a-00b1-4476-b122-18481390b128 service nova] [instance: 03100310-86ce-4739-8636-cd2aa1b542ff] Refreshing network info cache for port 5a9feebd-c914-422f-af20-45e4709470e8 {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 639.232259] env[61923]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-a0de33cd-d7af-419b-b23e-c49b36d5dd3f {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.243043] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6f9397c-dc1b-4c14-9bd1-0ff550cfa612 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.264784] env[61923]: WARNING nova.virt.vmwareapi.vmops [None req-753531fa-39a1-4c82-9915-d195bf0fe1b2 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] [instance: 03100310-86ce-4739-8636-cd2aa1b542ff] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 03100310-86ce-4739-8636-cd2aa1b542ff could not be found. [ 639.265014] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-753531fa-39a1-4c82-9915-d195bf0fe1b2 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] [instance: 03100310-86ce-4739-8636-cd2aa1b542ff] Instance destroyed {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 639.265209] env[61923]: INFO nova.compute.manager [None req-753531fa-39a1-4c82-9915-d195bf0fe1b2 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] [instance: 03100310-86ce-4739-8636-cd2aa1b542ff] Took 0.03 seconds to destroy the instance on the hypervisor. [ 639.265471] env[61923]: DEBUG oslo.service.loopingcall [None req-753531fa-39a1-4c82-9915-d195bf0fe1b2 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61923) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 639.265694] env[61923]: DEBUG nova.compute.manager [-] [instance: 03100310-86ce-4739-8636-cd2aa1b542ff] Deallocating network for instance {{(pid=61923) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 639.265785] env[61923]: DEBUG nova.network.neutron [-] [instance: 03100310-86ce-4739-8636-cd2aa1b542ff] deallocate_for_instance() {{(pid=61923) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 639.279171] env[61923]: DEBUG nova.network.neutron [-] [instance: 03100310-86ce-4739-8636-cd2aa1b542ff] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 639.317452] env[61923]: DEBUG nova.network.neutron [None req-b2fe22fb-da8f-41ba-8845-95fc6177390e tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] [instance: 119bb4a7-2d71-44d7-ae81-9238512937c8] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 639.566485] env[61923]: DEBUG nova.scheduler.client.report [None req-76010439-9fb6-4296-9b5a-107842db76db tempest-ServerExternalEventsTest-1479796458 tempest-ServerExternalEventsTest-1479796458-project-member] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 639.752309] env[61923]: DEBUG nova.network.neutron [req-bf606cee-4fb6-4927-859a-74b721741096 req-3bc7fc9a-00b1-4476-b122-18481390b128 service nova] [instance: 03100310-86ce-4739-8636-cd2aa1b542ff] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 639.783435] env[61923]: DEBUG nova.network.neutron [-] [instance: 03100310-86ce-4739-8636-cd2aa1b542ff] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 639.819434] env[61923]: INFO nova.compute.manager [None req-b2fe22fb-da8f-41ba-8845-95fc6177390e tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] [instance: 119bb4a7-2d71-44d7-ae81-9238512937c8] Took 1.02 seconds to deallocate network for instance. [ 639.823330] env[61923]: DEBUG nova.network.neutron [req-bf606cee-4fb6-4927-859a-74b721741096 req-3bc7fc9a-00b1-4476-b122-18481390b128 service nova] [instance: 03100310-86ce-4739-8636-cd2aa1b542ff] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 640.074085] env[61923]: DEBUG oslo_concurrency.lockutils [None req-76010439-9fb6-4296-9b5a-107842db76db tempest-ServerExternalEventsTest-1479796458 tempest-ServerExternalEventsTest-1479796458-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.381s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 640.074647] env[61923]: DEBUG nova.compute.manager [None req-76010439-9fb6-4296-9b5a-107842db76db tempest-ServerExternalEventsTest-1479796458 tempest-ServerExternalEventsTest-1479796458-project-member] [instance: c0f87aa5-447b-4920-8251-354d89a14fbb] Start building networks asynchronously for instance. {{(pid=61923) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 640.077685] env[61923]: DEBUG oslo_concurrency.lockutils [None req-85ac5bbf-a6d0-481b-b0bf-1428265b44a9 tempest-AttachInterfacesV270Test-411696101 tempest-AttachInterfacesV270Test-411696101-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.772s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 640.078750] env[61923]: INFO nova.compute.claims [None req-85ac5bbf-a6d0-481b-b0bf-1428265b44a9 tempest-AttachInterfacesV270Test-411696101 tempest-AttachInterfacesV270Test-411696101-project-member] [instance: 525f37d5-2262-40c8-a339-fc262c53ce6e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 640.286222] env[61923]: INFO nova.compute.manager [-] [instance: 03100310-86ce-4739-8636-cd2aa1b542ff] Took 1.02 seconds to deallocate network for instance. [ 640.288733] env[61923]: DEBUG nova.compute.claims [None req-753531fa-39a1-4c82-9915-d195bf0fe1b2 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] [instance: 03100310-86ce-4739-8636-cd2aa1b542ff] Aborting claim: {{(pid=61923) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 640.288875] env[61923]: DEBUG oslo_concurrency.lockutils [None req-753531fa-39a1-4c82-9915-d195bf0fe1b2 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 640.328053] env[61923]: DEBUG oslo_concurrency.lockutils [req-bf606cee-4fb6-4927-859a-74b721741096 req-3bc7fc9a-00b1-4476-b122-18481390b128 service nova] Releasing lock "refresh_cache-03100310-86ce-4739-8636-cd2aa1b542ff" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 640.328320] env[61923]: DEBUG nova.compute.manager [req-bf606cee-4fb6-4927-859a-74b721741096 req-3bc7fc9a-00b1-4476-b122-18481390b128 service nova] [instance: 03100310-86ce-4739-8636-cd2aa1b542ff] Received event network-vif-deleted-5a9feebd-c914-422f-af20-45e4709470e8 {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 640.583128] env[61923]: DEBUG nova.compute.utils [None req-76010439-9fb6-4296-9b5a-107842db76db tempest-ServerExternalEventsTest-1479796458 tempest-ServerExternalEventsTest-1479796458-project-member] Using /dev/sd instead of None {{(pid=61923) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 640.584586] env[61923]: DEBUG nova.compute.manager [None req-76010439-9fb6-4296-9b5a-107842db76db tempest-ServerExternalEventsTest-1479796458 tempest-ServerExternalEventsTest-1479796458-project-member] [instance: c0f87aa5-447b-4920-8251-354d89a14fbb] Allocating IP information in the background. {{(pid=61923) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 640.585019] env[61923]: DEBUG nova.network.neutron [None req-76010439-9fb6-4296-9b5a-107842db76db tempest-ServerExternalEventsTest-1479796458 tempest-ServerExternalEventsTest-1479796458-project-member] [instance: c0f87aa5-447b-4920-8251-354d89a14fbb] allocate_for_instance() {{(pid=61923) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 640.645545] env[61923]: DEBUG nova.policy [None req-76010439-9fb6-4296-9b5a-107842db76db tempest-ServerExternalEventsTest-1479796458 tempest-ServerExternalEventsTest-1479796458-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'b4e5982c4f2246ce873cb185114d7011', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f111424450e646f6b4cac54d338923c5', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61923) authorize /opt/stack/nova/nova/policy.py:201}} [ 640.847990] env[61923]: INFO nova.scheduler.client.report [None req-b2fe22fb-da8f-41ba-8845-95fc6177390e tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Deleted allocations for instance 119bb4a7-2d71-44d7-ae81-9238512937c8 [ 641.047626] env[61923]: DEBUG nova.network.neutron [None req-76010439-9fb6-4296-9b5a-107842db76db tempest-ServerExternalEventsTest-1479796458 tempest-ServerExternalEventsTest-1479796458-project-member] [instance: c0f87aa5-447b-4920-8251-354d89a14fbb] Successfully created port: c67bdeb7-ce64-47ec-9e91-62b8f48de195 {{(pid=61923) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 641.087582] env[61923]: DEBUG nova.compute.manager [None req-76010439-9fb6-4296-9b5a-107842db76db tempest-ServerExternalEventsTest-1479796458 tempest-ServerExternalEventsTest-1479796458-project-member] [instance: c0f87aa5-447b-4920-8251-354d89a14fbb] Start building block device mappings for instance. {{(pid=61923) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 641.358602] env[61923]: DEBUG oslo_concurrency.lockutils [None req-b2fe22fb-da8f-41ba-8845-95fc6177390e tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Lock "119bb4a7-2d71-44d7-ae81-9238512937c8" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 98.549s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 641.424025] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60ead6b0-f260-43df-b7d7-e9ea71622be2 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.431559] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ad1d7d8-903c-40fc-b323-c5c8ccfc8dba {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.461824] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d63c03f-bc2a-4c6d-ae93-1f9167e842a8 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.469093] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b8b6ac0-7cb6-4dd7-a4e6-bca79a8e59ba {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.481852] env[61923]: DEBUG nova.compute.provider_tree [None req-85ac5bbf-a6d0-481b-b0bf-1428265b44a9 tempest-AttachInterfacesV270Test-411696101 tempest-AttachInterfacesV270Test-411696101-project-member] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 641.864476] env[61923]: DEBUG nova.compute.manager [None req-859ffdff-aaee-46b4-bdd6-0a15cce9792f tempest-ServerRescueNegativeTestJSON-1435265126 tempest-ServerRescueNegativeTestJSON-1435265126-project-member] [instance: 41b6744e-d548-442c-b6cc-e12bd403f229] Starting instance... {{(pid=61923) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 641.988243] env[61923]: DEBUG nova.scheduler.client.report [None req-85ac5bbf-a6d0-481b-b0bf-1428265b44a9 tempest-AttachInterfacesV270Test-411696101 tempest-AttachInterfacesV270Test-411696101-project-member] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 642.099253] env[61923]: DEBUG nova.compute.manager [None req-76010439-9fb6-4296-9b5a-107842db76db tempest-ServerExternalEventsTest-1479796458 tempest-ServerExternalEventsTest-1479796458-project-member] [instance: c0f87aa5-447b-4920-8251-354d89a14fbb] Start spawning the instance on the hypervisor. {{(pid=61923) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 642.125029] env[61923]: DEBUG nova.virt.hardware [None req-76010439-9fb6-4296-9b5a-107842db76db tempest-ServerExternalEventsTest-1479796458 tempest-ServerExternalEventsTest-1479796458-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-29T20:07:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-29T20:07:35Z,direct_url=,disk_format='vmdk',id=0f153f63-ae0a-45b1-b7f5-7f9b673c947f,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='4e7b5e3b3c3443c8bd5c70f8a15739f5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-29T20:07:36Z,virtual_size=,visibility=), allow threads: False {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 642.125029] env[61923]: DEBUG nova.virt.hardware [None req-76010439-9fb6-4296-9b5a-107842db76db tempest-ServerExternalEventsTest-1479796458 tempest-ServerExternalEventsTest-1479796458-project-member] Flavor limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 642.125029] env[61923]: DEBUG nova.virt.hardware [None req-76010439-9fb6-4296-9b5a-107842db76db tempest-ServerExternalEventsTest-1479796458 tempest-ServerExternalEventsTest-1479796458-project-member] Image limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 642.125224] env[61923]: DEBUG nova.virt.hardware [None req-76010439-9fb6-4296-9b5a-107842db76db tempest-ServerExternalEventsTest-1479796458 tempest-ServerExternalEventsTest-1479796458-project-member] Flavor pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 642.125224] env[61923]: DEBUG nova.virt.hardware [None req-76010439-9fb6-4296-9b5a-107842db76db tempest-ServerExternalEventsTest-1479796458 tempest-ServerExternalEventsTest-1479796458-project-member] Image pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 642.125224] env[61923]: DEBUG nova.virt.hardware [None req-76010439-9fb6-4296-9b5a-107842db76db tempest-ServerExternalEventsTest-1479796458 tempest-ServerExternalEventsTest-1479796458-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 642.125224] env[61923]: DEBUG nova.virt.hardware [None req-76010439-9fb6-4296-9b5a-107842db76db tempest-ServerExternalEventsTest-1479796458 tempest-ServerExternalEventsTest-1479796458-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 642.125224] env[61923]: DEBUG nova.virt.hardware [None req-76010439-9fb6-4296-9b5a-107842db76db tempest-ServerExternalEventsTest-1479796458 tempest-ServerExternalEventsTest-1479796458-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 642.125357] env[61923]: DEBUG nova.virt.hardware [None req-76010439-9fb6-4296-9b5a-107842db76db tempest-ServerExternalEventsTest-1479796458 tempest-ServerExternalEventsTest-1479796458-project-member] Got 1 possible topologies {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 642.125357] env[61923]: DEBUG nova.virt.hardware [None req-76010439-9fb6-4296-9b5a-107842db76db tempest-ServerExternalEventsTest-1479796458 tempest-ServerExternalEventsTest-1479796458-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 642.125682] env[61923]: DEBUG nova.virt.hardware [None req-76010439-9fb6-4296-9b5a-107842db76db tempest-ServerExternalEventsTest-1479796458 tempest-ServerExternalEventsTest-1479796458-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 642.126675] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68cfe3b8-bc50-4fd8-94b7-5c2906895926 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 642.138124] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50ef775c-f248-4754-9188-d4cba36e6acb {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 642.217409] env[61923]: DEBUG nova.compute.manager [req-caef5341-7d56-48df-bb7c-c8417945fce8 req-0461ac69-6ce5-4804-b839-33643e55035f service nova] [instance: c0f87aa5-447b-4920-8251-354d89a14fbb] Received event network-changed-c67bdeb7-ce64-47ec-9e91-62b8f48de195 {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 642.217616] env[61923]: DEBUG nova.compute.manager [req-caef5341-7d56-48df-bb7c-c8417945fce8 req-0461ac69-6ce5-4804-b839-33643e55035f service nova] [instance: c0f87aa5-447b-4920-8251-354d89a14fbb] Refreshing instance network info cache due to event network-changed-c67bdeb7-ce64-47ec-9e91-62b8f48de195. {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 642.217828] env[61923]: DEBUG oslo_concurrency.lockutils [req-caef5341-7d56-48df-bb7c-c8417945fce8 req-0461ac69-6ce5-4804-b839-33643e55035f service nova] Acquiring lock "refresh_cache-c0f87aa5-447b-4920-8251-354d89a14fbb" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 642.217969] env[61923]: DEBUG oslo_concurrency.lockutils [req-caef5341-7d56-48df-bb7c-c8417945fce8 req-0461ac69-6ce5-4804-b839-33643e55035f service nova] Acquired lock "refresh_cache-c0f87aa5-447b-4920-8251-354d89a14fbb" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 642.218143] env[61923]: DEBUG nova.network.neutron [req-caef5341-7d56-48df-bb7c-c8417945fce8 req-0461ac69-6ce5-4804-b839-33643e55035f service nova] [instance: c0f87aa5-447b-4920-8251-354d89a14fbb] Refreshing network info cache for port c67bdeb7-ce64-47ec-9e91-62b8f48de195 {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 642.382254] env[61923]: DEBUG oslo_concurrency.lockutils [None req-859ffdff-aaee-46b4-bdd6-0a15cce9792f tempest-ServerRescueNegativeTestJSON-1435265126 tempest-ServerRescueNegativeTestJSON-1435265126-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 642.452620] env[61923]: ERROR nova.compute.manager [None req-76010439-9fb6-4296-9b5a-107842db76db tempest-ServerExternalEventsTest-1479796458 tempest-ServerExternalEventsTest-1479796458-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port c67bdeb7-ce64-47ec-9e91-62b8f48de195, please check neutron logs for more information. [ 642.452620] env[61923]: ERROR nova.compute.manager Traceback (most recent call last): [ 642.452620] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 642.452620] env[61923]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 642.452620] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 642.452620] env[61923]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 642.452620] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 642.452620] env[61923]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 642.452620] env[61923]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 642.452620] env[61923]: ERROR nova.compute.manager self.force_reraise() [ 642.452620] env[61923]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 642.452620] env[61923]: ERROR nova.compute.manager raise self.value [ 642.452620] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 642.452620] env[61923]: ERROR nova.compute.manager updated_port = self._update_port( [ 642.452620] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 642.452620] env[61923]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 642.453127] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 642.453127] env[61923]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 642.453127] env[61923]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port c67bdeb7-ce64-47ec-9e91-62b8f48de195, please check neutron logs for more information. [ 642.453127] env[61923]: ERROR nova.compute.manager [ 642.453127] env[61923]: Traceback (most recent call last): [ 642.453127] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 642.453127] env[61923]: listener.cb(fileno) [ 642.453127] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 642.453127] env[61923]: result = function(*args, **kwargs) [ 642.453127] env[61923]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 642.453127] env[61923]: return func(*args, **kwargs) [ 642.453127] env[61923]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 642.453127] env[61923]: raise e [ 642.453127] env[61923]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 642.453127] env[61923]: nwinfo = self.network_api.allocate_for_instance( [ 642.453127] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 642.453127] env[61923]: created_port_ids = self._update_ports_for_instance( [ 642.453127] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 642.453127] env[61923]: with excutils.save_and_reraise_exception(): [ 642.453127] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 642.453127] env[61923]: self.force_reraise() [ 642.453127] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 642.453127] env[61923]: raise self.value [ 642.453127] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 642.453127] env[61923]: updated_port = self._update_port( [ 642.453127] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 642.453127] env[61923]: _ensure_no_port_binding_failure(port) [ 642.453127] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 642.453127] env[61923]: raise exception.PortBindingFailed(port_id=port['id']) [ 642.453917] env[61923]: nova.exception.PortBindingFailed: Binding failed for port c67bdeb7-ce64-47ec-9e91-62b8f48de195, please check neutron logs for more information. [ 642.453917] env[61923]: Removing descriptor: 17 [ 642.454252] env[61923]: ERROR nova.compute.manager [None req-76010439-9fb6-4296-9b5a-107842db76db tempest-ServerExternalEventsTest-1479796458 tempest-ServerExternalEventsTest-1479796458-project-member] [instance: c0f87aa5-447b-4920-8251-354d89a14fbb] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port c67bdeb7-ce64-47ec-9e91-62b8f48de195, please check neutron logs for more information. [ 642.454252] env[61923]: ERROR nova.compute.manager [instance: c0f87aa5-447b-4920-8251-354d89a14fbb] Traceback (most recent call last): [ 642.454252] env[61923]: ERROR nova.compute.manager [instance: c0f87aa5-447b-4920-8251-354d89a14fbb] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 642.454252] env[61923]: ERROR nova.compute.manager [instance: c0f87aa5-447b-4920-8251-354d89a14fbb] yield resources [ 642.454252] env[61923]: ERROR nova.compute.manager [instance: c0f87aa5-447b-4920-8251-354d89a14fbb] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 642.454252] env[61923]: ERROR nova.compute.manager [instance: c0f87aa5-447b-4920-8251-354d89a14fbb] self.driver.spawn(context, instance, image_meta, [ 642.454252] env[61923]: ERROR nova.compute.manager [instance: c0f87aa5-447b-4920-8251-354d89a14fbb] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 642.454252] env[61923]: ERROR nova.compute.manager [instance: c0f87aa5-447b-4920-8251-354d89a14fbb] self._vmops.spawn(context, instance, image_meta, injected_files, [ 642.454252] env[61923]: ERROR nova.compute.manager [instance: c0f87aa5-447b-4920-8251-354d89a14fbb] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 642.454252] env[61923]: ERROR nova.compute.manager [instance: c0f87aa5-447b-4920-8251-354d89a14fbb] vm_ref = self.build_virtual_machine(instance, [ 642.454252] env[61923]: ERROR nova.compute.manager [instance: c0f87aa5-447b-4920-8251-354d89a14fbb] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 642.454588] env[61923]: ERROR nova.compute.manager [instance: c0f87aa5-447b-4920-8251-354d89a14fbb] vif_infos = vmwarevif.get_vif_info(self._session, [ 642.454588] env[61923]: ERROR nova.compute.manager [instance: c0f87aa5-447b-4920-8251-354d89a14fbb] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 642.454588] env[61923]: ERROR nova.compute.manager [instance: c0f87aa5-447b-4920-8251-354d89a14fbb] for vif in network_info: [ 642.454588] env[61923]: ERROR nova.compute.manager [instance: c0f87aa5-447b-4920-8251-354d89a14fbb] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 642.454588] env[61923]: ERROR nova.compute.manager [instance: c0f87aa5-447b-4920-8251-354d89a14fbb] return self._sync_wrapper(fn, *args, **kwargs) [ 642.454588] env[61923]: ERROR nova.compute.manager [instance: c0f87aa5-447b-4920-8251-354d89a14fbb] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 642.454588] env[61923]: ERROR nova.compute.manager [instance: c0f87aa5-447b-4920-8251-354d89a14fbb] self.wait() [ 642.454588] env[61923]: ERROR nova.compute.manager [instance: c0f87aa5-447b-4920-8251-354d89a14fbb] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 642.454588] env[61923]: ERROR nova.compute.manager [instance: c0f87aa5-447b-4920-8251-354d89a14fbb] self[:] = self._gt.wait() [ 642.454588] env[61923]: ERROR nova.compute.manager [instance: c0f87aa5-447b-4920-8251-354d89a14fbb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 642.454588] env[61923]: ERROR nova.compute.manager [instance: c0f87aa5-447b-4920-8251-354d89a14fbb] return self._exit_event.wait() [ 642.454588] env[61923]: ERROR nova.compute.manager [instance: c0f87aa5-447b-4920-8251-354d89a14fbb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 642.454588] env[61923]: ERROR nova.compute.manager [instance: c0f87aa5-447b-4920-8251-354d89a14fbb] result = hub.switch() [ 642.455024] env[61923]: ERROR nova.compute.manager [instance: c0f87aa5-447b-4920-8251-354d89a14fbb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 642.455024] env[61923]: ERROR nova.compute.manager [instance: c0f87aa5-447b-4920-8251-354d89a14fbb] return self.greenlet.switch() [ 642.455024] env[61923]: ERROR nova.compute.manager [instance: c0f87aa5-447b-4920-8251-354d89a14fbb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 642.455024] env[61923]: ERROR nova.compute.manager [instance: c0f87aa5-447b-4920-8251-354d89a14fbb] result = function(*args, **kwargs) [ 642.455024] env[61923]: ERROR nova.compute.manager [instance: c0f87aa5-447b-4920-8251-354d89a14fbb] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 642.455024] env[61923]: ERROR nova.compute.manager [instance: c0f87aa5-447b-4920-8251-354d89a14fbb] return func(*args, **kwargs) [ 642.455024] env[61923]: ERROR nova.compute.manager [instance: c0f87aa5-447b-4920-8251-354d89a14fbb] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 642.455024] env[61923]: ERROR nova.compute.manager [instance: c0f87aa5-447b-4920-8251-354d89a14fbb] raise e [ 642.455024] env[61923]: ERROR nova.compute.manager [instance: c0f87aa5-447b-4920-8251-354d89a14fbb] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 642.455024] env[61923]: ERROR nova.compute.manager [instance: c0f87aa5-447b-4920-8251-354d89a14fbb] nwinfo = self.network_api.allocate_for_instance( [ 642.455024] env[61923]: ERROR nova.compute.manager [instance: c0f87aa5-447b-4920-8251-354d89a14fbb] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 642.455024] env[61923]: ERROR nova.compute.manager [instance: c0f87aa5-447b-4920-8251-354d89a14fbb] created_port_ids = self._update_ports_for_instance( [ 642.455024] env[61923]: ERROR nova.compute.manager [instance: c0f87aa5-447b-4920-8251-354d89a14fbb] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 642.455439] env[61923]: ERROR nova.compute.manager [instance: c0f87aa5-447b-4920-8251-354d89a14fbb] with excutils.save_and_reraise_exception(): [ 642.455439] env[61923]: ERROR nova.compute.manager [instance: c0f87aa5-447b-4920-8251-354d89a14fbb] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 642.455439] env[61923]: ERROR nova.compute.manager [instance: c0f87aa5-447b-4920-8251-354d89a14fbb] self.force_reraise() [ 642.455439] env[61923]: ERROR nova.compute.manager [instance: c0f87aa5-447b-4920-8251-354d89a14fbb] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 642.455439] env[61923]: ERROR nova.compute.manager [instance: c0f87aa5-447b-4920-8251-354d89a14fbb] raise self.value [ 642.455439] env[61923]: ERROR nova.compute.manager [instance: c0f87aa5-447b-4920-8251-354d89a14fbb] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 642.455439] env[61923]: ERROR nova.compute.manager [instance: c0f87aa5-447b-4920-8251-354d89a14fbb] updated_port = self._update_port( [ 642.455439] env[61923]: ERROR nova.compute.manager [instance: c0f87aa5-447b-4920-8251-354d89a14fbb] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 642.455439] env[61923]: ERROR nova.compute.manager [instance: c0f87aa5-447b-4920-8251-354d89a14fbb] _ensure_no_port_binding_failure(port) [ 642.455439] env[61923]: ERROR nova.compute.manager [instance: c0f87aa5-447b-4920-8251-354d89a14fbb] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 642.455439] env[61923]: ERROR nova.compute.manager [instance: c0f87aa5-447b-4920-8251-354d89a14fbb] raise exception.PortBindingFailed(port_id=port['id']) [ 642.455439] env[61923]: ERROR nova.compute.manager [instance: c0f87aa5-447b-4920-8251-354d89a14fbb] nova.exception.PortBindingFailed: Binding failed for port c67bdeb7-ce64-47ec-9e91-62b8f48de195, please check neutron logs for more information. [ 642.455439] env[61923]: ERROR nova.compute.manager [instance: c0f87aa5-447b-4920-8251-354d89a14fbb] [ 642.455804] env[61923]: INFO nova.compute.manager [None req-76010439-9fb6-4296-9b5a-107842db76db tempest-ServerExternalEventsTest-1479796458 tempest-ServerExternalEventsTest-1479796458-project-member] [instance: c0f87aa5-447b-4920-8251-354d89a14fbb] Terminating instance [ 642.457040] env[61923]: DEBUG oslo_concurrency.lockutils [None req-76010439-9fb6-4296-9b5a-107842db76db tempest-ServerExternalEventsTest-1479796458 tempest-ServerExternalEventsTest-1479796458-project-member] Acquiring lock "refresh_cache-c0f87aa5-447b-4920-8251-354d89a14fbb" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 642.493717] env[61923]: DEBUG oslo_concurrency.lockutils [None req-85ac5bbf-a6d0-481b-b0bf-1428265b44a9 tempest-AttachInterfacesV270Test-411696101 tempest-AttachInterfacesV270Test-411696101-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.416s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 642.494236] env[61923]: DEBUG nova.compute.manager [None req-85ac5bbf-a6d0-481b-b0bf-1428265b44a9 tempest-AttachInterfacesV270Test-411696101 tempest-AttachInterfacesV270Test-411696101-project-member] [instance: 525f37d5-2262-40c8-a339-fc262c53ce6e] Start building networks asynchronously for instance. {{(pid=61923) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 642.497380] env[61923]: DEBUG oslo_concurrency.lockutils [None req-4bd5c12b-d52a-4c54-93f7-87894f09013a tempest-ServersTestJSON-483293076 tempest-ServersTestJSON-483293076-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.144s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 642.498915] env[61923]: INFO nova.compute.claims [None req-4bd5c12b-d52a-4c54-93f7-87894f09013a tempest-ServersTestJSON-483293076 tempest-ServersTestJSON-483293076-project-member] [instance: 55b60fc2-6484-4fad-aa40-3e73e020841a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 642.590197] env[61923]: DEBUG oslo_concurrency.lockutils [None req-0708225f-80bb-41ec-b3ed-7dd45d589b95 tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Acquiring lock "62d8a170-e7ad-4d14-90d8-6f6ce32b7cf8" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 642.590560] env[61923]: DEBUG oslo_concurrency.lockutils [None req-0708225f-80bb-41ec-b3ed-7dd45d589b95 tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Lock "62d8a170-e7ad-4d14-90d8-6f6ce32b7cf8" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 642.734206] env[61923]: DEBUG nova.network.neutron [req-caef5341-7d56-48df-bb7c-c8417945fce8 req-0461ac69-6ce5-4804-b839-33643e55035f service nova] [instance: c0f87aa5-447b-4920-8251-354d89a14fbb] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 642.797406] env[61923]: DEBUG nova.network.neutron [req-caef5341-7d56-48df-bb7c-c8417945fce8 req-0461ac69-6ce5-4804-b839-33643e55035f service nova] [instance: c0f87aa5-447b-4920-8251-354d89a14fbb] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 643.003757] env[61923]: DEBUG nova.compute.utils [None req-85ac5bbf-a6d0-481b-b0bf-1428265b44a9 tempest-AttachInterfacesV270Test-411696101 tempest-AttachInterfacesV270Test-411696101-project-member] Using /dev/sd instead of None {{(pid=61923) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 643.008685] env[61923]: DEBUG nova.compute.manager [None req-85ac5bbf-a6d0-481b-b0bf-1428265b44a9 tempest-AttachInterfacesV270Test-411696101 tempest-AttachInterfacesV270Test-411696101-project-member] [instance: 525f37d5-2262-40c8-a339-fc262c53ce6e] Allocating IP information in the background. {{(pid=61923) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 643.008685] env[61923]: DEBUG nova.network.neutron [None req-85ac5bbf-a6d0-481b-b0bf-1428265b44a9 tempest-AttachInterfacesV270Test-411696101 tempest-AttachInterfacesV270Test-411696101-project-member] [instance: 525f37d5-2262-40c8-a339-fc262c53ce6e] allocate_for_instance() {{(pid=61923) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 643.051043] env[61923]: DEBUG nova.policy [None req-85ac5bbf-a6d0-481b-b0bf-1428265b44a9 tempest-AttachInterfacesV270Test-411696101 tempest-AttachInterfacesV270Test-411696101-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '88438311ba61481fb190e4316c105d9f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '506f8d36207e4483b78fd1afb9f5a4f8', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61923) authorize /opt/stack/nova/nova/policy.py:201}} [ 643.300093] env[61923]: DEBUG oslo_concurrency.lockutils [req-caef5341-7d56-48df-bb7c-c8417945fce8 req-0461ac69-6ce5-4804-b839-33643e55035f service nova] Releasing lock "refresh_cache-c0f87aa5-447b-4920-8251-354d89a14fbb" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 643.300575] env[61923]: DEBUG oslo_concurrency.lockutils [None req-76010439-9fb6-4296-9b5a-107842db76db tempest-ServerExternalEventsTest-1479796458 tempest-ServerExternalEventsTest-1479796458-project-member] Acquired lock "refresh_cache-c0f87aa5-447b-4920-8251-354d89a14fbb" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 643.300768] env[61923]: DEBUG nova.network.neutron [None req-76010439-9fb6-4296-9b5a-107842db76db tempest-ServerExternalEventsTest-1479796458 tempest-ServerExternalEventsTest-1479796458-project-member] [instance: c0f87aa5-447b-4920-8251-354d89a14fbb] Building network info cache for instance {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 643.508435] env[61923]: DEBUG nova.compute.manager [None req-85ac5bbf-a6d0-481b-b0bf-1428265b44a9 tempest-AttachInterfacesV270Test-411696101 tempest-AttachInterfacesV270Test-411696101-project-member] [instance: 525f37d5-2262-40c8-a339-fc262c53ce6e] Start building block device mappings for instance. {{(pid=61923) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 643.625011] env[61923]: DEBUG nova.network.neutron [None req-85ac5bbf-a6d0-481b-b0bf-1428265b44a9 tempest-AttachInterfacesV270Test-411696101 tempest-AttachInterfacesV270Test-411696101-project-member] [instance: 525f37d5-2262-40c8-a339-fc262c53ce6e] Successfully created port: 0aa60830-cb1d-45ad-8800-f2352dd43766 {{(pid=61923) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 643.833520] env[61923]: DEBUG nova.network.neutron [None req-76010439-9fb6-4296-9b5a-107842db76db tempest-ServerExternalEventsTest-1479796458 tempest-ServerExternalEventsTest-1479796458-project-member] [instance: c0f87aa5-447b-4920-8251-354d89a14fbb] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 643.878510] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ccdddaf8-9684-4b10-a618-62f3b29246e1 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.886093] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e032a77f-20be-402f-a69a-36f0eb96f520 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.917222] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c4c59f1-4365-4154-8b14-13a9a78fdaad {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.924824] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39b8ef92-a279-475c-86cb-7276e86485d8 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 643.937532] env[61923]: DEBUG nova.compute.provider_tree [None req-4bd5c12b-d52a-4c54-93f7-87894f09013a tempest-ServersTestJSON-483293076 tempest-ServersTestJSON-483293076-project-member] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 643.939365] env[61923]: DEBUG nova.network.neutron [None req-76010439-9fb6-4296-9b5a-107842db76db tempest-ServerExternalEventsTest-1479796458 tempest-ServerExternalEventsTest-1479796458-project-member] [instance: c0f87aa5-447b-4920-8251-354d89a14fbb] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 644.261165] env[61923]: DEBUG nova.compute.manager [req-bc3f2b0f-7d44-4d37-8bc1-4327201a8ae0 req-ef60c0e2-7ddd-40a3-a6ee-89e37e505ac8 service nova] [instance: c0f87aa5-447b-4920-8251-354d89a14fbb] Received event network-vif-deleted-c67bdeb7-ce64-47ec-9e91-62b8f48de195 {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 644.441512] env[61923]: DEBUG nova.scheduler.client.report [None req-4bd5c12b-d52a-4c54-93f7-87894f09013a tempest-ServersTestJSON-483293076 tempest-ServersTestJSON-483293076-project-member] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 644.444833] env[61923]: DEBUG oslo_concurrency.lockutils [None req-76010439-9fb6-4296-9b5a-107842db76db tempest-ServerExternalEventsTest-1479796458 tempest-ServerExternalEventsTest-1479796458-project-member] Releasing lock "refresh_cache-c0f87aa5-447b-4920-8251-354d89a14fbb" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 644.445247] env[61923]: DEBUG nova.compute.manager [None req-76010439-9fb6-4296-9b5a-107842db76db tempest-ServerExternalEventsTest-1479796458 tempest-ServerExternalEventsTest-1479796458-project-member] [instance: c0f87aa5-447b-4920-8251-354d89a14fbb] Start destroying the instance on the hypervisor. {{(pid=61923) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 644.445468] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-76010439-9fb6-4296-9b5a-107842db76db tempest-ServerExternalEventsTest-1479796458 tempest-ServerExternalEventsTest-1479796458-project-member] [instance: c0f87aa5-447b-4920-8251-354d89a14fbb] Destroying instance {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 644.445929] env[61923]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-b917d6e3-5f53-43b1-8ed0-2dcaf9b8f348 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.455905] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2650bf8-6e74-4d34-aa87-f9604049ff15 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.479320] env[61923]: WARNING nova.virt.vmwareapi.vmops [None req-76010439-9fb6-4296-9b5a-107842db76db tempest-ServerExternalEventsTest-1479796458 tempest-ServerExternalEventsTest-1479796458-project-member] [instance: c0f87aa5-447b-4920-8251-354d89a14fbb] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance c0f87aa5-447b-4920-8251-354d89a14fbb could not be found. [ 644.479477] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-76010439-9fb6-4296-9b5a-107842db76db tempest-ServerExternalEventsTest-1479796458 tempest-ServerExternalEventsTest-1479796458-project-member] [instance: c0f87aa5-447b-4920-8251-354d89a14fbb] Instance destroyed {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 644.479679] env[61923]: INFO nova.compute.manager [None req-76010439-9fb6-4296-9b5a-107842db76db tempest-ServerExternalEventsTest-1479796458 tempest-ServerExternalEventsTest-1479796458-project-member] [instance: c0f87aa5-447b-4920-8251-354d89a14fbb] Took 0.03 seconds to destroy the instance on the hypervisor. [ 644.479924] env[61923]: DEBUG oslo.service.loopingcall [None req-76010439-9fb6-4296-9b5a-107842db76db tempest-ServerExternalEventsTest-1479796458 tempest-ServerExternalEventsTest-1479796458-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61923) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 644.481933] env[61923]: DEBUG nova.compute.manager [-] [instance: c0f87aa5-447b-4920-8251-354d89a14fbb] Deallocating network for instance {{(pid=61923) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 644.482077] env[61923]: DEBUG nova.network.neutron [-] [instance: c0f87aa5-447b-4920-8251-354d89a14fbb] deallocate_for_instance() {{(pid=61923) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 644.519636] env[61923]: DEBUG nova.compute.manager [None req-85ac5bbf-a6d0-481b-b0bf-1428265b44a9 tempest-AttachInterfacesV270Test-411696101 tempest-AttachInterfacesV270Test-411696101-project-member] [instance: 525f37d5-2262-40c8-a339-fc262c53ce6e] Start spawning the instance on the hypervisor. {{(pid=61923) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 644.523010] env[61923]: DEBUG nova.network.neutron [-] [instance: c0f87aa5-447b-4920-8251-354d89a14fbb] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 644.545486] env[61923]: DEBUG nova.virt.hardware [None req-85ac5bbf-a6d0-481b-b0bf-1428265b44a9 tempest-AttachInterfacesV270Test-411696101 tempest-AttachInterfacesV270Test-411696101-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-29T20:07:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-29T20:07:35Z,direct_url=,disk_format='vmdk',id=0f153f63-ae0a-45b1-b7f5-7f9b673c947f,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='4e7b5e3b3c3443c8bd5c70f8a15739f5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-29T20:07:36Z,virtual_size=,visibility=), allow threads: False {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 644.545743] env[61923]: DEBUG nova.virt.hardware [None req-85ac5bbf-a6d0-481b-b0bf-1428265b44a9 tempest-AttachInterfacesV270Test-411696101 tempest-AttachInterfacesV270Test-411696101-project-member] Flavor limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 644.545933] env[61923]: DEBUG nova.virt.hardware [None req-85ac5bbf-a6d0-481b-b0bf-1428265b44a9 tempest-AttachInterfacesV270Test-411696101 tempest-AttachInterfacesV270Test-411696101-project-member] Image limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 644.546152] env[61923]: DEBUG nova.virt.hardware [None req-85ac5bbf-a6d0-481b-b0bf-1428265b44a9 tempest-AttachInterfacesV270Test-411696101 tempest-AttachInterfacesV270Test-411696101-project-member] Flavor pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 644.546301] env[61923]: DEBUG nova.virt.hardware [None req-85ac5bbf-a6d0-481b-b0bf-1428265b44a9 tempest-AttachInterfacesV270Test-411696101 tempest-AttachInterfacesV270Test-411696101-project-member] Image pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 644.546447] env[61923]: DEBUG nova.virt.hardware [None req-85ac5bbf-a6d0-481b-b0bf-1428265b44a9 tempest-AttachInterfacesV270Test-411696101 tempest-AttachInterfacesV270Test-411696101-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 644.546652] env[61923]: DEBUG nova.virt.hardware [None req-85ac5bbf-a6d0-481b-b0bf-1428265b44a9 tempest-AttachInterfacesV270Test-411696101 tempest-AttachInterfacesV270Test-411696101-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 644.546806] env[61923]: DEBUG nova.virt.hardware [None req-85ac5bbf-a6d0-481b-b0bf-1428265b44a9 tempest-AttachInterfacesV270Test-411696101 tempest-AttachInterfacesV270Test-411696101-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 644.546969] env[61923]: DEBUG nova.virt.hardware [None req-85ac5bbf-a6d0-481b-b0bf-1428265b44a9 tempest-AttachInterfacesV270Test-411696101 tempest-AttachInterfacesV270Test-411696101-project-member] Got 1 possible topologies {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 644.547158] env[61923]: DEBUG nova.virt.hardware [None req-85ac5bbf-a6d0-481b-b0bf-1428265b44a9 tempest-AttachInterfacesV270Test-411696101 tempest-AttachInterfacesV270Test-411696101-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 644.547354] env[61923]: DEBUG nova.virt.hardware [None req-85ac5bbf-a6d0-481b-b0bf-1428265b44a9 tempest-AttachInterfacesV270Test-411696101 tempest-AttachInterfacesV270Test-411696101-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 644.548448] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc51e5fe-1f45-4456-aac5-f4636b28f383 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.556130] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70252abe-d1ac-4ca5-815e-08e738090a68 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.863990] env[61923]: ERROR nova.compute.manager [None req-85ac5bbf-a6d0-481b-b0bf-1428265b44a9 tempest-AttachInterfacesV270Test-411696101 tempest-AttachInterfacesV270Test-411696101-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 0aa60830-cb1d-45ad-8800-f2352dd43766, please check neutron logs for more information. [ 644.863990] env[61923]: ERROR nova.compute.manager Traceback (most recent call last): [ 644.863990] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 644.863990] env[61923]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 644.863990] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 644.863990] env[61923]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 644.863990] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 644.863990] env[61923]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 644.863990] env[61923]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 644.863990] env[61923]: ERROR nova.compute.manager self.force_reraise() [ 644.863990] env[61923]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 644.863990] env[61923]: ERROR nova.compute.manager raise self.value [ 644.863990] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 644.863990] env[61923]: ERROR nova.compute.manager updated_port = self._update_port( [ 644.863990] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 644.863990] env[61923]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 644.864811] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 644.864811] env[61923]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 644.864811] env[61923]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 0aa60830-cb1d-45ad-8800-f2352dd43766, please check neutron logs for more information. [ 644.864811] env[61923]: ERROR nova.compute.manager [ 644.864811] env[61923]: Traceback (most recent call last): [ 644.864811] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 644.864811] env[61923]: listener.cb(fileno) [ 644.864811] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 644.864811] env[61923]: result = function(*args, **kwargs) [ 644.864811] env[61923]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 644.864811] env[61923]: return func(*args, **kwargs) [ 644.864811] env[61923]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 644.864811] env[61923]: raise e [ 644.864811] env[61923]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 644.864811] env[61923]: nwinfo = self.network_api.allocate_for_instance( [ 644.864811] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 644.864811] env[61923]: created_port_ids = self._update_ports_for_instance( [ 644.864811] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 644.864811] env[61923]: with excutils.save_and_reraise_exception(): [ 644.864811] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 644.864811] env[61923]: self.force_reraise() [ 644.864811] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 644.864811] env[61923]: raise self.value [ 644.864811] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 644.864811] env[61923]: updated_port = self._update_port( [ 644.864811] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 644.864811] env[61923]: _ensure_no_port_binding_failure(port) [ 644.864811] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 644.864811] env[61923]: raise exception.PortBindingFailed(port_id=port['id']) [ 644.867022] env[61923]: nova.exception.PortBindingFailed: Binding failed for port 0aa60830-cb1d-45ad-8800-f2352dd43766, please check neutron logs for more information. [ 644.867022] env[61923]: Removing descriptor: 17 [ 644.867022] env[61923]: ERROR nova.compute.manager [None req-85ac5bbf-a6d0-481b-b0bf-1428265b44a9 tempest-AttachInterfacesV270Test-411696101 tempest-AttachInterfacesV270Test-411696101-project-member] [instance: 525f37d5-2262-40c8-a339-fc262c53ce6e] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 0aa60830-cb1d-45ad-8800-f2352dd43766, please check neutron logs for more information. [ 644.867022] env[61923]: ERROR nova.compute.manager [instance: 525f37d5-2262-40c8-a339-fc262c53ce6e] Traceback (most recent call last): [ 644.867022] env[61923]: ERROR nova.compute.manager [instance: 525f37d5-2262-40c8-a339-fc262c53ce6e] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 644.867022] env[61923]: ERROR nova.compute.manager [instance: 525f37d5-2262-40c8-a339-fc262c53ce6e] yield resources [ 644.867022] env[61923]: ERROR nova.compute.manager [instance: 525f37d5-2262-40c8-a339-fc262c53ce6e] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 644.867022] env[61923]: ERROR nova.compute.manager [instance: 525f37d5-2262-40c8-a339-fc262c53ce6e] self.driver.spawn(context, instance, image_meta, [ 644.867022] env[61923]: ERROR nova.compute.manager [instance: 525f37d5-2262-40c8-a339-fc262c53ce6e] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 644.867022] env[61923]: ERROR nova.compute.manager [instance: 525f37d5-2262-40c8-a339-fc262c53ce6e] self._vmops.spawn(context, instance, image_meta, injected_files, [ 644.867022] env[61923]: ERROR nova.compute.manager [instance: 525f37d5-2262-40c8-a339-fc262c53ce6e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 644.867022] env[61923]: ERROR nova.compute.manager [instance: 525f37d5-2262-40c8-a339-fc262c53ce6e] vm_ref = self.build_virtual_machine(instance, [ 644.867487] env[61923]: ERROR nova.compute.manager [instance: 525f37d5-2262-40c8-a339-fc262c53ce6e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 644.867487] env[61923]: ERROR nova.compute.manager [instance: 525f37d5-2262-40c8-a339-fc262c53ce6e] vif_infos = vmwarevif.get_vif_info(self._session, [ 644.867487] env[61923]: ERROR nova.compute.manager [instance: 525f37d5-2262-40c8-a339-fc262c53ce6e] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 644.867487] env[61923]: ERROR nova.compute.manager [instance: 525f37d5-2262-40c8-a339-fc262c53ce6e] for vif in network_info: [ 644.867487] env[61923]: ERROR nova.compute.manager [instance: 525f37d5-2262-40c8-a339-fc262c53ce6e] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 644.867487] env[61923]: ERROR nova.compute.manager [instance: 525f37d5-2262-40c8-a339-fc262c53ce6e] return self._sync_wrapper(fn, *args, **kwargs) [ 644.867487] env[61923]: ERROR nova.compute.manager [instance: 525f37d5-2262-40c8-a339-fc262c53ce6e] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 644.867487] env[61923]: ERROR nova.compute.manager [instance: 525f37d5-2262-40c8-a339-fc262c53ce6e] self.wait() [ 644.867487] env[61923]: ERROR nova.compute.manager [instance: 525f37d5-2262-40c8-a339-fc262c53ce6e] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 644.867487] env[61923]: ERROR nova.compute.manager [instance: 525f37d5-2262-40c8-a339-fc262c53ce6e] self[:] = self._gt.wait() [ 644.867487] env[61923]: ERROR nova.compute.manager [instance: 525f37d5-2262-40c8-a339-fc262c53ce6e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 644.867487] env[61923]: ERROR nova.compute.manager [instance: 525f37d5-2262-40c8-a339-fc262c53ce6e] return self._exit_event.wait() [ 644.867487] env[61923]: ERROR nova.compute.manager [instance: 525f37d5-2262-40c8-a339-fc262c53ce6e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 644.867872] env[61923]: ERROR nova.compute.manager [instance: 525f37d5-2262-40c8-a339-fc262c53ce6e] result = hub.switch() [ 644.867872] env[61923]: ERROR nova.compute.manager [instance: 525f37d5-2262-40c8-a339-fc262c53ce6e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 644.867872] env[61923]: ERROR nova.compute.manager [instance: 525f37d5-2262-40c8-a339-fc262c53ce6e] return self.greenlet.switch() [ 644.867872] env[61923]: ERROR nova.compute.manager [instance: 525f37d5-2262-40c8-a339-fc262c53ce6e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 644.867872] env[61923]: ERROR nova.compute.manager [instance: 525f37d5-2262-40c8-a339-fc262c53ce6e] result = function(*args, **kwargs) [ 644.867872] env[61923]: ERROR nova.compute.manager [instance: 525f37d5-2262-40c8-a339-fc262c53ce6e] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 644.867872] env[61923]: ERROR nova.compute.manager [instance: 525f37d5-2262-40c8-a339-fc262c53ce6e] return func(*args, **kwargs) [ 644.867872] env[61923]: ERROR nova.compute.manager [instance: 525f37d5-2262-40c8-a339-fc262c53ce6e] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 644.867872] env[61923]: ERROR nova.compute.manager [instance: 525f37d5-2262-40c8-a339-fc262c53ce6e] raise e [ 644.867872] env[61923]: ERROR nova.compute.manager [instance: 525f37d5-2262-40c8-a339-fc262c53ce6e] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 644.867872] env[61923]: ERROR nova.compute.manager [instance: 525f37d5-2262-40c8-a339-fc262c53ce6e] nwinfo = self.network_api.allocate_for_instance( [ 644.867872] env[61923]: ERROR nova.compute.manager [instance: 525f37d5-2262-40c8-a339-fc262c53ce6e] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 644.867872] env[61923]: ERROR nova.compute.manager [instance: 525f37d5-2262-40c8-a339-fc262c53ce6e] created_port_ids = self._update_ports_for_instance( [ 644.868325] env[61923]: ERROR nova.compute.manager [instance: 525f37d5-2262-40c8-a339-fc262c53ce6e] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 644.868325] env[61923]: ERROR nova.compute.manager [instance: 525f37d5-2262-40c8-a339-fc262c53ce6e] with excutils.save_and_reraise_exception(): [ 644.868325] env[61923]: ERROR nova.compute.manager [instance: 525f37d5-2262-40c8-a339-fc262c53ce6e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 644.868325] env[61923]: ERROR nova.compute.manager [instance: 525f37d5-2262-40c8-a339-fc262c53ce6e] self.force_reraise() [ 644.868325] env[61923]: ERROR nova.compute.manager [instance: 525f37d5-2262-40c8-a339-fc262c53ce6e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 644.868325] env[61923]: ERROR nova.compute.manager [instance: 525f37d5-2262-40c8-a339-fc262c53ce6e] raise self.value [ 644.868325] env[61923]: ERROR nova.compute.manager [instance: 525f37d5-2262-40c8-a339-fc262c53ce6e] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 644.868325] env[61923]: ERROR nova.compute.manager [instance: 525f37d5-2262-40c8-a339-fc262c53ce6e] updated_port = self._update_port( [ 644.868325] env[61923]: ERROR nova.compute.manager [instance: 525f37d5-2262-40c8-a339-fc262c53ce6e] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 644.868325] env[61923]: ERROR nova.compute.manager [instance: 525f37d5-2262-40c8-a339-fc262c53ce6e] _ensure_no_port_binding_failure(port) [ 644.868325] env[61923]: ERROR nova.compute.manager [instance: 525f37d5-2262-40c8-a339-fc262c53ce6e] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 644.868325] env[61923]: ERROR nova.compute.manager [instance: 525f37d5-2262-40c8-a339-fc262c53ce6e] raise exception.PortBindingFailed(port_id=port['id']) [ 644.869212] env[61923]: ERROR nova.compute.manager [instance: 525f37d5-2262-40c8-a339-fc262c53ce6e] nova.exception.PortBindingFailed: Binding failed for port 0aa60830-cb1d-45ad-8800-f2352dd43766, please check neutron logs for more information. [ 644.869212] env[61923]: ERROR nova.compute.manager [instance: 525f37d5-2262-40c8-a339-fc262c53ce6e] [ 644.869212] env[61923]: INFO nova.compute.manager [None req-85ac5bbf-a6d0-481b-b0bf-1428265b44a9 tempest-AttachInterfacesV270Test-411696101 tempest-AttachInterfacesV270Test-411696101-project-member] [instance: 525f37d5-2262-40c8-a339-fc262c53ce6e] Terminating instance [ 644.869935] env[61923]: DEBUG oslo_concurrency.lockutils [None req-85ac5bbf-a6d0-481b-b0bf-1428265b44a9 tempest-AttachInterfacesV270Test-411696101 tempest-AttachInterfacesV270Test-411696101-project-member] Acquiring lock "refresh_cache-525f37d5-2262-40c8-a339-fc262c53ce6e" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 644.870070] env[61923]: DEBUG oslo_concurrency.lockutils [None req-85ac5bbf-a6d0-481b-b0bf-1428265b44a9 tempest-AttachInterfacesV270Test-411696101 tempest-AttachInterfacesV270Test-411696101-project-member] Acquired lock "refresh_cache-525f37d5-2262-40c8-a339-fc262c53ce6e" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 644.870933] env[61923]: DEBUG nova.network.neutron [None req-85ac5bbf-a6d0-481b-b0bf-1428265b44a9 tempest-AttachInterfacesV270Test-411696101 tempest-AttachInterfacesV270Test-411696101-project-member] [instance: 525f37d5-2262-40c8-a339-fc262c53ce6e] Building network info cache for instance {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 644.946973] env[61923]: DEBUG oslo_concurrency.lockutils [None req-4bd5c12b-d52a-4c54-93f7-87894f09013a tempest-ServersTestJSON-483293076 tempest-ServersTestJSON-483293076-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.450s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 644.947562] env[61923]: DEBUG nova.compute.manager [None req-4bd5c12b-d52a-4c54-93f7-87894f09013a tempest-ServersTestJSON-483293076 tempest-ServersTestJSON-483293076-project-member] [instance: 55b60fc2-6484-4fad-aa40-3e73e020841a] Start building networks asynchronously for instance. {{(pid=61923) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 644.950343] env[61923]: DEBUG oslo_concurrency.lockutils [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 17.376s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 644.950521] env[61923]: DEBUG oslo_concurrency.lockutils [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 644.950670] env[61923]: DEBUG nova.compute.resource_tracker [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61923) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 644.950948] env[61923]: DEBUG oslo_concurrency.lockutils [None req-aa9c464b-ac6b-44fe-9384-28cd9bdfecbc tempest-ImagesOneServerTestJSON-63657725 tempest-ImagesOneServerTestJSON-63657725-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.553s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 644.954174] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bccbcc5a-78ea-4fbf-90ea-13159a5dc4df {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.962674] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb5760d0-04c7-40c9-84d5-699ab0d7231c {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.977209] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dce39b6e-d36c-4fec-bd27-e310752eb27d {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.983830] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c445ae6-7e05-4c0b-8b69-419ee125d3da {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.013383] env[61923]: DEBUG nova.compute.resource_tracker [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181444MB free_disk=178GB free_vcpus=48 pci_devices=None {{(pid=61923) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 645.013542] env[61923]: DEBUG oslo_concurrency.lockutils [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 645.025122] env[61923]: DEBUG nova.network.neutron [-] [instance: c0f87aa5-447b-4920-8251-354d89a14fbb] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 645.387096] env[61923]: DEBUG nova.network.neutron [None req-85ac5bbf-a6d0-481b-b0bf-1428265b44a9 tempest-AttachInterfacesV270Test-411696101 tempest-AttachInterfacesV270Test-411696101-project-member] [instance: 525f37d5-2262-40c8-a339-fc262c53ce6e] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 645.455385] env[61923]: DEBUG nova.network.neutron [None req-85ac5bbf-a6d0-481b-b0bf-1428265b44a9 tempest-AttachInterfacesV270Test-411696101 tempest-AttachInterfacesV270Test-411696101-project-member] [instance: 525f37d5-2262-40c8-a339-fc262c53ce6e] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 645.458028] env[61923]: DEBUG nova.compute.utils [None req-4bd5c12b-d52a-4c54-93f7-87894f09013a tempest-ServersTestJSON-483293076 tempest-ServersTestJSON-483293076-project-member] Using /dev/sd instead of None {{(pid=61923) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 645.461589] env[61923]: DEBUG nova.compute.manager [None req-4bd5c12b-d52a-4c54-93f7-87894f09013a tempest-ServersTestJSON-483293076 tempest-ServersTestJSON-483293076-project-member] [instance: 55b60fc2-6484-4fad-aa40-3e73e020841a] Allocating IP information in the background. {{(pid=61923) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 645.461759] env[61923]: DEBUG nova.network.neutron [None req-4bd5c12b-d52a-4c54-93f7-87894f09013a tempest-ServersTestJSON-483293076 tempest-ServersTestJSON-483293076-project-member] [instance: 55b60fc2-6484-4fad-aa40-3e73e020841a] allocate_for_instance() {{(pid=61923) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 645.512713] env[61923]: DEBUG nova.policy [None req-4bd5c12b-d52a-4c54-93f7-87894f09013a tempest-ServersTestJSON-483293076 tempest-ServersTestJSON-483293076-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5fcc223f2ded4f4b9cea8c61bcf99679', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f0e689d37f794bdc97a52601a52f2fc4', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61923) authorize /opt/stack/nova/nova/policy.py:201}} [ 645.527597] env[61923]: INFO nova.compute.manager [-] [instance: c0f87aa5-447b-4920-8251-354d89a14fbb] Took 1.05 seconds to deallocate network for instance. [ 645.531610] env[61923]: DEBUG nova.compute.claims [None req-76010439-9fb6-4296-9b5a-107842db76db tempest-ServerExternalEventsTest-1479796458 tempest-ServerExternalEventsTest-1479796458-project-member] [instance: c0f87aa5-447b-4920-8251-354d89a14fbb] Aborting claim: {{(pid=61923) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 645.531818] env[61923]: DEBUG oslo_concurrency.lockutils [None req-76010439-9fb6-4296-9b5a-107842db76db tempest-ServerExternalEventsTest-1479796458 tempest-ServerExternalEventsTest-1479796458-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 645.787022] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01d9bdfc-9429-442e-84e6-88f3b764891a {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.800992] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7fac997-529f-48e3-a362-a3e96bfc12bc {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.835745] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc632237-6afd-433d-abc1-92dce03f7199 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.844046] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b3074e9-1290-470d-8355-31752872a1a0 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.857992] env[61923]: DEBUG nova.compute.provider_tree [None req-aa9c464b-ac6b-44fe-9384-28cd9bdfecbc tempest-ImagesOneServerTestJSON-63657725 tempest-ImagesOneServerTestJSON-63657725-project-member] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 645.902343] env[61923]: DEBUG nova.network.neutron [None req-4bd5c12b-d52a-4c54-93f7-87894f09013a tempest-ServersTestJSON-483293076 tempest-ServersTestJSON-483293076-project-member] [instance: 55b60fc2-6484-4fad-aa40-3e73e020841a] Successfully created port: 445a37d4-005a-49e8-afb5-816754610ddc {{(pid=61923) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 645.960419] env[61923]: DEBUG oslo_concurrency.lockutils [None req-85ac5bbf-a6d0-481b-b0bf-1428265b44a9 tempest-AttachInterfacesV270Test-411696101 tempest-AttachInterfacesV270Test-411696101-project-member] Releasing lock "refresh_cache-525f37d5-2262-40c8-a339-fc262c53ce6e" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 645.960865] env[61923]: DEBUG nova.compute.manager [None req-85ac5bbf-a6d0-481b-b0bf-1428265b44a9 tempest-AttachInterfacesV270Test-411696101 tempest-AttachInterfacesV270Test-411696101-project-member] [instance: 525f37d5-2262-40c8-a339-fc262c53ce6e] Start destroying the instance on the hypervisor. {{(pid=61923) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 645.961087] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-85ac5bbf-a6d0-481b-b0bf-1428265b44a9 tempest-AttachInterfacesV270Test-411696101 tempest-AttachInterfacesV270Test-411696101-project-member] [instance: 525f37d5-2262-40c8-a339-fc262c53ce6e] Destroying instance {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 645.961380] env[61923]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-6855a4cf-4687-4a28-a4df-e085ad197cf7 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.964515] env[61923]: DEBUG nova.compute.manager [None req-4bd5c12b-d52a-4c54-93f7-87894f09013a tempest-ServersTestJSON-483293076 tempest-ServersTestJSON-483293076-project-member] [instance: 55b60fc2-6484-4fad-aa40-3e73e020841a] Start building block device mappings for instance. {{(pid=61923) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 645.971303] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4cb77140-35b7-4121-90f6-ca1b4b53de94 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 645.995431] env[61923]: WARNING nova.virt.vmwareapi.vmops [None req-85ac5bbf-a6d0-481b-b0bf-1428265b44a9 tempest-AttachInterfacesV270Test-411696101 tempest-AttachInterfacesV270Test-411696101-project-member] [instance: 525f37d5-2262-40c8-a339-fc262c53ce6e] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 525f37d5-2262-40c8-a339-fc262c53ce6e could not be found. [ 645.995431] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-85ac5bbf-a6d0-481b-b0bf-1428265b44a9 tempest-AttachInterfacesV270Test-411696101 tempest-AttachInterfacesV270Test-411696101-project-member] [instance: 525f37d5-2262-40c8-a339-fc262c53ce6e] Instance destroyed {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 645.995431] env[61923]: INFO nova.compute.manager [None req-85ac5bbf-a6d0-481b-b0bf-1428265b44a9 tempest-AttachInterfacesV270Test-411696101 tempest-AttachInterfacesV270Test-411696101-project-member] [instance: 525f37d5-2262-40c8-a339-fc262c53ce6e] Took 0.03 seconds to destroy the instance on the hypervisor. [ 645.995580] env[61923]: DEBUG oslo.service.loopingcall [None req-85ac5bbf-a6d0-481b-b0bf-1428265b44a9 tempest-AttachInterfacesV270Test-411696101 tempest-AttachInterfacesV270Test-411696101-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61923) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 645.995844] env[61923]: DEBUG nova.compute.manager [-] [instance: 525f37d5-2262-40c8-a339-fc262c53ce6e] Deallocating network for instance {{(pid=61923) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 645.995844] env[61923]: DEBUG nova.network.neutron [-] [instance: 525f37d5-2262-40c8-a339-fc262c53ce6e] deallocate_for_instance() {{(pid=61923) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 646.017569] env[61923]: DEBUG nova.network.neutron [-] [instance: 525f37d5-2262-40c8-a339-fc262c53ce6e] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 646.348392] env[61923]: DEBUG nova.compute.manager [req-df70278a-71cc-469b-bf5f-5bd9e52bb3ae req-037306bf-5792-49cc-85aa-ef64ad16018f service nova] [instance: 525f37d5-2262-40c8-a339-fc262c53ce6e] Received event network-changed-0aa60830-cb1d-45ad-8800-f2352dd43766 {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 646.348694] env[61923]: DEBUG nova.compute.manager [req-df70278a-71cc-469b-bf5f-5bd9e52bb3ae req-037306bf-5792-49cc-85aa-ef64ad16018f service nova] [instance: 525f37d5-2262-40c8-a339-fc262c53ce6e] Refreshing instance network info cache due to event network-changed-0aa60830-cb1d-45ad-8800-f2352dd43766. {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 646.348857] env[61923]: DEBUG oslo_concurrency.lockutils [req-df70278a-71cc-469b-bf5f-5bd9e52bb3ae req-037306bf-5792-49cc-85aa-ef64ad16018f service nova] Acquiring lock "refresh_cache-525f37d5-2262-40c8-a339-fc262c53ce6e" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 646.348921] env[61923]: DEBUG oslo_concurrency.lockutils [req-df70278a-71cc-469b-bf5f-5bd9e52bb3ae req-037306bf-5792-49cc-85aa-ef64ad16018f service nova] Acquired lock "refresh_cache-525f37d5-2262-40c8-a339-fc262c53ce6e" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 646.349094] env[61923]: DEBUG nova.network.neutron [req-df70278a-71cc-469b-bf5f-5bd9e52bb3ae req-037306bf-5792-49cc-85aa-ef64ad16018f service nova] [instance: 525f37d5-2262-40c8-a339-fc262c53ce6e] Refreshing network info cache for port 0aa60830-cb1d-45ad-8800-f2352dd43766 {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 646.360986] env[61923]: DEBUG nova.scheduler.client.report [None req-aa9c464b-ac6b-44fe-9384-28cd9bdfecbc tempest-ImagesOneServerTestJSON-63657725 tempest-ImagesOneServerTestJSON-63657725-project-member] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 646.524025] env[61923]: DEBUG nova.network.neutron [-] [instance: 525f37d5-2262-40c8-a339-fc262c53ce6e] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 646.865370] env[61923]: DEBUG oslo_concurrency.lockutils [None req-aa9c464b-ac6b-44fe-9384-28cd9bdfecbc tempest-ImagesOneServerTestJSON-63657725 tempest-ImagesOneServerTestJSON-63657725-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.914s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 646.866015] env[61923]: ERROR nova.compute.manager [None req-aa9c464b-ac6b-44fe-9384-28cd9bdfecbc tempest-ImagesOneServerTestJSON-63657725 tempest-ImagesOneServerTestJSON-63657725-project-member] [instance: 0c1b8117-199f-4101-ab81-4ae7ef0d1251] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 9ba992af-f024-4728-b3b6-b6cf6af4e319, please check neutron logs for more information. [ 646.866015] env[61923]: ERROR nova.compute.manager [instance: 0c1b8117-199f-4101-ab81-4ae7ef0d1251] Traceback (most recent call last): [ 646.866015] env[61923]: ERROR nova.compute.manager [instance: 0c1b8117-199f-4101-ab81-4ae7ef0d1251] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 646.866015] env[61923]: ERROR nova.compute.manager [instance: 0c1b8117-199f-4101-ab81-4ae7ef0d1251] self.driver.spawn(context, instance, image_meta, [ 646.866015] env[61923]: ERROR nova.compute.manager [instance: 0c1b8117-199f-4101-ab81-4ae7ef0d1251] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 646.866015] env[61923]: ERROR nova.compute.manager [instance: 0c1b8117-199f-4101-ab81-4ae7ef0d1251] self._vmops.spawn(context, instance, image_meta, injected_files, [ 646.866015] env[61923]: ERROR nova.compute.manager [instance: 0c1b8117-199f-4101-ab81-4ae7ef0d1251] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 646.866015] env[61923]: ERROR nova.compute.manager [instance: 0c1b8117-199f-4101-ab81-4ae7ef0d1251] vm_ref = self.build_virtual_machine(instance, [ 646.866015] env[61923]: ERROR nova.compute.manager [instance: 0c1b8117-199f-4101-ab81-4ae7ef0d1251] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 646.866015] env[61923]: ERROR nova.compute.manager [instance: 0c1b8117-199f-4101-ab81-4ae7ef0d1251] vif_infos = vmwarevif.get_vif_info(self._session, [ 646.866015] env[61923]: ERROR nova.compute.manager [instance: 0c1b8117-199f-4101-ab81-4ae7ef0d1251] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 646.866304] env[61923]: ERROR nova.compute.manager [instance: 0c1b8117-199f-4101-ab81-4ae7ef0d1251] for vif in network_info: [ 646.866304] env[61923]: ERROR nova.compute.manager [instance: 0c1b8117-199f-4101-ab81-4ae7ef0d1251] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 646.866304] env[61923]: ERROR nova.compute.manager [instance: 0c1b8117-199f-4101-ab81-4ae7ef0d1251] return self._sync_wrapper(fn, *args, **kwargs) [ 646.866304] env[61923]: ERROR nova.compute.manager [instance: 0c1b8117-199f-4101-ab81-4ae7ef0d1251] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 646.866304] env[61923]: ERROR nova.compute.manager [instance: 0c1b8117-199f-4101-ab81-4ae7ef0d1251] self.wait() [ 646.866304] env[61923]: ERROR nova.compute.manager [instance: 0c1b8117-199f-4101-ab81-4ae7ef0d1251] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 646.866304] env[61923]: ERROR nova.compute.manager [instance: 0c1b8117-199f-4101-ab81-4ae7ef0d1251] self[:] = self._gt.wait() [ 646.866304] env[61923]: ERROR nova.compute.manager [instance: 0c1b8117-199f-4101-ab81-4ae7ef0d1251] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 646.866304] env[61923]: ERROR nova.compute.manager [instance: 0c1b8117-199f-4101-ab81-4ae7ef0d1251] return self._exit_event.wait() [ 646.866304] env[61923]: ERROR nova.compute.manager [instance: 0c1b8117-199f-4101-ab81-4ae7ef0d1251] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 646.866304] env[61923]: ERROR nova.compute.manager [instance: 0c1b8117-199f-4101-ab81-4ae7ef0d1251] result = hub.switch() [ 646.866304] env[61923]: ERROR nova.compute.manager [instance: 0c1b8117-199f-4101-ab81-4ae7ef0d1251] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 646.866304] env[61923]: ERROR nova.compute.manager [instance: 0c1b8117-199f-4101-ab81-4ae7ef0d1251] return self.greenlet.switch() [ 646.866633] env[61923]: ERROR nova.compute.manager [instance: 0c1b8117-199f-4101-ab81-4ae7ef0d1251] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 646.866633] env[61923]: ERROR nova.compute.manager [instance: 0c1b8117-199f-4101-ab81-4ae7ef0d1251] result = function(*args, **kwargs) [ 646.866633] env[61923]: ERROR nova.compute.manager [instance: 0c1b8117-199f-4101-ab81-4ae7ef0d1251] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 646.866633] env[61923]: ERROR nova.compute.manager [instance: 0c1b8117-199f-4101-ab81-4ae7ef0d1251] return func(*args, **kwargs) [ 646.866633] env[61923]: ERROR nova.compute.manager [instance: 0c1b8117-199f-4101-ab81-4ae7ef0d1251] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 646.866633] env[61923]: ERROR nova.compute.manager [instance: 0c1b8117-199f-4101-ab81-4ae7ef0d1251] raise e [ 646.866633] env[61923]: ERROR nova.compute.manager [instance: 0c1b8117-199f-4101-ab81-4ae7ef0d1251] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 646.866633] env[61923]: ERROR nova.compute.manager [instance: 0c1b8117-199f-4101-ab81-4ae7ef0d1251] nwinfo = self.network_api.allocate_for_instance( [ 646.866633] env[61923]: ERROR nova.compute.manager [instance: 0c1b8117-199f-4101-ab81-4ae7ef0d1251] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 646.866633] env[61923]: ERROR nova.compute.manager [instance: 0c1b8117-199f-4101-ab81-4ae7ef0d1251] created_port_ids = self._update_ports_for_instance( [ 646.866633] env[61923]: ERROR nova.compute.manager [instance: 0c1b8117-199f-4101-ab81-4ae7ef0d1251] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 646.866633] env[61923]: ERROR nova.compute.manager [instance: 0c1b8117-199f-4101-ab81-4ae7ef0d1251] with excutils.save_and_reraise_exception(): [ 646.866633] env[61923]: ERROR nova.compute.manager [instance: 0c1b8117-199f-4101-ab81-4ae7ef0d1251] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 646.866937] env[61923]: ERROR nova.compute.manager [instance: 0c1b8117-199f-4101-ab81-4ae7ef0d1251] self.force_reraise() [ 646.866937] env[61923]: ERROR nova.compute.manager [instance: 0c1b8117-199f-4101-ab81-4ae7ef0d1251] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 646.866937] env[61923]: ERROR nova.compute.manager [instance: 0c1b8117-199f-4101-ab81-4ae7ef0d1251] raise self.value [ 646.866937] env[61923]: ERROR nova.compute.manager [instance: 0c1b8117-199f-4101-ab81-4ae7ef0d1251] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 646.866937] env[61923]: ERROR nova.compute.manager [instance: 0c1b8117-199f-4101-ab81-4ae7ef0d1251] updated_port = self._update_port( [ 646.866937] env[61923]: ERROR nova.compute.manager [instance: 0c1b8117-199f-4101-ab81-4ae7ef0d1251] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 646.866937] env[61923]: ERROR nova.compute.manager [instance: 0c1b8117-199f-4101-ab81-4ae7ef0d1251] _ensure_no_port_binding_failure(port) [ 646.866937] env[61923]: ERROR nova.compute.manager [instance: 0c1b8117-199f-4101-ab81-4ae7ef0d1251] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 646.866937] env[61923]: ERROR nova.compute.manager [instance: 0c1b8117-199f-4101-ab81-4ae7ef0d1251] raise exception.PortBindingFailed(port_id=port['id']) [ 646.866937] env[61923]: ERROR nova.compute.manager [instance: 0c1b8117-199f-4101-ab81-4ae7ef0d1251] nova.exception.PortBindingFailed: Binding failed for port 9ba992af-f024-4728-b3b6-b6cf6af4e319, please check neutron logs for more information. [ 646.866937] env[61923]: ERROR nova.compute.manager [instance: 0c1b8117-199f-4101-ab81-4ae7ef0d1251] [ 646.867218] env[61923]: DEBUG nova.compute.utils [None req-aa9c464b-ac6b-44fe-9384-28cd9bdfecbc tempest-ImagesOneServerTestJSON-63657725 tempest-ImagesOneServerTestJSON-63657725-project-member] [instance: 0c1b8117-199f-4101-ab81-4ae7ef0d1251] Binding failed for port 9ba992af-f024-4728-b3b6-b6cf6af4e319, please check neutron logs for more information. {{(pid=61923) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 646.867901] env[61923]: DEBUG oslo_concurrency.lockutils [None req-0c55849d-766e-4fac-b1c4-e4e97ba688a6 tempest-DeleteServersAdminTestJSON-590116201 tempest-DeleteServersAdminTestJSON-590116201-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.975s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 646.870800] env[61923]: DEBUG nova.compute.manager [None req-aa9c464b-ac6b-44fe-9384-28cd9bdfecbc tempest-ImagesOneServerTestJSON-63657725 tempest-ImagesOneServerTestJSON-63657725-project-member] [instance: 0c1b8117-199f-4101-ab81-4ae7ef0d1251] Build of instance 0c1b8117-199f-4101-ab81-4ae7ef0d1251 was re-scheduled: Binding failed for port 9ba992af-f024-4728-b3b6-b6cf6af4e319, please check neutron logs for more information. {{(pid=61923) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 646.871234] env[61923]: DEBUG nova.compute.manager [None req-aa9c464b-ac6b-44fe-9384-28cd9bdfecbc tempest-ImagesOneServerTestJSON-63657725 tempest-ImagesOneServerTestJSON-63657725-project-member] [instance: 0c1b8117-199f-4101-ab81-4ae7ef0d1251] Unplugging VIFs for instance {{(pid=61923) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 646.871469] env[61923]: DEBUG oslo_concurrency.lockutils [None req-aa9c464b-ac6b-44fe-9384-28cd9bdfecbc tempest-ImagesOneServerTestJSON-63657725 tempest-ImagesOneServerTestJSON-63657725-project-member] Acquiring lock "refresh_cache-0c1b8117-199f-4101-ab81-4ae7ef0d1251" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 646.871621] env[61923]: DEBUG oslo_concurrency.lockutils [None req-aa9c464b-ac6b-44fe-9384-28cd9bdfecbc tempest-ImagesOneServerTestJSON-63657725 tempest-ImagesOneServerTestJSON-63657725-project-member] Acquired lock "refresh_cache-0c1b8117-199f-4101-ab81-4ae7ef0d1251" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 646.871779] env[61923]: DEBUG nova.network.neutron [None req-aa9c464b-ac6b-44fe-9384-28cd9bdfecbc tempest-ImagesOneServerTestJSON-63657725 tempest-ImagesOneServerTestJSON-63657725-project-member] [instance: 0c1b8117-199f-4101-ab81-4ae7ef0d1251] Building network info cache for instance {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 646.881676] env[61923]: DEBUG nova.network.neutron [req-df70278a-71cc-469b-bf5f-5bd9e52bb3ae req-037306bf-5792-49cc-85aa-ef64ad16018f service nova] [instance: 525f37d5-2262-40c8-a339-fc262c53ce6e] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 646.974267] env[61923]: DEBUG nova.compute.manager [None req-4bd5c12b-d52a-4c54-93f7-87894f09013a tempest-ServersTestJSON-483293076 tempest-ServersTestJSON-483293076-project-member] [instance: 55b60fc2-6484-4fad-aa40-3e73e020841a] Start spawning the instance on the hypervisor. {{(pid=61923) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 647.009100] env[61923]: DEBUG nova.virt.hardware [None req-4bd5c12b-d52a-4c54-93f7-87894f09013a tempest-ServersTestJSON-483293076 tempest-ServersTestJSON-483293076-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-29T20:07:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-29T20:07:35Z,direct_url=,disk_format='vmdk',id=0f153f63-ae0a-45b1-b7f5-7f9b673c947f,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='4e7b5e3b3c3443c8bd5c70f8a15739f5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-29T20:07:36Z,virtual_size=,visibility=), allow threads: False {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 647.009100] env[61923]: DEBUG nova.virt.hardware [None req-4bd5c12b-d52a-4c54-93f7-87894f09013a tempest-ServersTestJSON-483293076 tempest-ServersTestJSON-483293076-project-member] Flavor limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 647.009100] env[61923]: DEBUG nova.virt.hardware [None req-4bd5c12b-d52a-4c54-93f7-87894f09013a tempest-ServersTestJSON-483293076 tempest-ServersTestJSON-483293076-project-member] Image limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 647.009434] env[61923]: DEBUG nova.virt.hardware [None req-4bd5c12b-d52a-4c54-93f7-87894f09013a tempest-ServersTestJSON-483293076 tempest-ServersTestJSON-483293076-project-member] Flavor pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 647.009434] env[61923]: DEBUG nova.virt.hardware [None req-4bd5c12b-d52a-4c54-93f7-87894f09013a tempest-ServersTestJSON-483293076 tempest-ServersTestJSON-483293076-project-member] Image pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 647.009564] env[61923]: DEBUG nova.virt.hardware [None req-4bd5c12b-d52a-4c54-93f7-87894f09013a tempest-ServersTestJSON-483293076 tempest-ServersTestJSON-483293076-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 647.009765] env[61923]: DEBUG nova.virt.hardware [None req-4bd5c12b-d52a-4c54-93f7-87894f09013a tempest-ServersTestJSON-483293076 tempest-ServersTestJSON-483293076-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 647.009954] env[61923]: DEBUG nova.virt.hardware [None req-4bd5c12b-d52a-4c54-93f7-87894f09013a tempest-ServersTestJSON-483293076 tempest-ServersTestJSON-483293076-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 647.010159] env[61923]: DEBUG nova.virt.hardware [None req-4bd5c12b-d52a-4c54-93f7-87894f09013a tempest-ServersTestJSON-483293076 tempest-ServersTestJSON-483293076-project-member] Got 1 possible topologies {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 647.010322] env[61923]: DEBUG nova.virt.hardware [None req-4bd5c12b-d52a-4c54-93f7-87894f09013a tempest-ServersTestJSON-483293076 tempest-ServersTestJSON-483293076-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 647.010488] env[61923]: DEBUG nova.virt.hardware [None req-4bd5c12b-d52a-4c54-93f7-87894f09013a tempest-ServersTestJSON-483293076 tempest-ServersTestJSON-483293076-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 647.011402] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0f485f7-0fdb-4793-bf6e-a14efd552d9e {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 647.022129] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14d544bd-594a-4acb-ac49-db6bd944b403 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 647.037858] env[61923]: INFO nova.compute.manager [-] [instance: 525f37d5-2262-40c8-a339-fc262c53ce6e] Took 1.04 seconds to deallocate network for instance. [ 647.040561] env[61923]: DEBUG nova.compute.claims [None req-85ac5bbf-a6d0-481b-b0bf-1428265b44a9 tempest-AttachInterfacesV270Test-411696101 tempest-AttachInterfacesV270Test-411696101-project-member] [instance: 525f37d5-2262-40c8-a339-fc262c53ce6e] Aborting claim: {{(pid=61923) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 647.040741] env[61923]: DEBUG oslo_concurrency.lockutils [None req-85ac5bbf-a6d0-481b-b0bf-1428265b44a9 tempest-AttachInterfacesV270Test-411696101 tempest-AttachInterfacesV270Test-411696101-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 647.053227] env[61923]: DEBUG nova.network.neutron [req-df70278a-71cc-469b-bf5f-5bd9e52bb3ae req-037306bf-5792-49cc-85aa-ef64ad16018f service nova] [instance: 525f37d5-2262-40c8-a339-fc262c53ce6e] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 647.074828] env[61923]: ERROR nova.compute.manager [None req-4bd5c12b-d52a-4c54-93f7-87894f09013a tempest-ServersTestJSON-483293076 tempest-ServersTestJSON-483293076-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 445a37d4-005a-49e8-afb5-816754610ddc, please check neutron logs for more information. [ 647.074828] env[61923]: ERROR nova.compute.manager Traceback (most recent call last): [ 647.074828] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 647.074828] env[61923]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 647.074828] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 647.074828] env[61923]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 647.074828] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 647.074828] env[61923]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 647.074828] env[61923]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 647.074828] env[61923]: ERROR nova.compute.manager self.force_reraise() [ 647.074828] env[61923]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 647.074828] env[61923]: ERROR nova.compute.manager raise self.value [ 647.074828] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 647.074828] env[61923]: ERROR nova.compute.manager updated_port = self._update_port( [ 647.074828] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 647.074828] env[61923]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 647.075566] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 647.075566] env[61923]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 647.075566] env[61923]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 445a37d4-005a-49e8-afb5-816754610ddc, please check neutron logs for more information. [ 647.075566] env[61923]: ERROR nova.compute.manager [ 647.075566] env[61923]: Traceback (most recent call last): [ 647.075566] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 647.075566] env[61923]: listener.cb(fileno) [ 647.075566] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 647.075566] env[61923]: result = function(*args, **kwargs) [ 647.075566] env[61923]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 647.075566] env[61923]: return func(*args, **kwargs) [ 647.075566] env[61923]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 647.075566] env[61923]: raise e [ 647.075566] env[61923]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 647.075566] env[61923]: nwinfo = self.network_api.allocate_for_instance( [ 647.075566] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 647.075566] env[61923]: created_port_ids = self._update_ports_for_instance( [ 647.075566] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 647.075566] env[61923]: with excutils.save_and_reraise_exception(): [ 647.075566] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 647.075566] env[61923]: self.force_reraise() [ 647.075566] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 647.075566] env[61923]: raise self.value [ 647.075566] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 647.075566] env[61923]: updated_port = self._update_port( [ 647.075566] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 647.075566] env[61923]: _ensure_no_port_binding_failure(port) [ 647.075566] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 647.075566] env[61923]: raise exception.PortBindingFailed(port_id=port['id']) [ 647.076661] env[61923]: nova.exception.PortBindingFailed: Binding failed for port 445a37d4-005a-49e8-afb5-816754610ddc, please check neutron logs for more information. [ 647.076661] env[61923]: Removing descriptor: 17 [ 647.076661] env[61923]: ERROR nova.compute.manager [None req-4bd5c12b-d52a-4c54-93f7-87894f09013a tempest-ServersTestJSON-483293076 tempest-ServersTestJSON-483293076-project-member] [instance: 55b60fc2-6484-4fad-aa40-3e73e020841a] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 445a37d4-005a-49e8-afb5-816754610ddc, please check neutron logs for more information. [ 647.076661] env[61923]: ERROR nova.compute.manager [instance: 55b60fc2-6484-4fad-aa40-3e73e020841a] Traceback (most recent call last): [ 647.076661] env[61923]: ERROR nova.compute.manager [instance: 55b60fc2-6484-4fad-aa40-3e73e020841a] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 647.076661] env[61923]: ERROR nova.compute.manager [instance: 55b60fc2-6484-4fad-aa40-3e73e020841a] yield resources [ 647.076661] env[61923]: ERROR nova.compute.manager [instance: 55b60fc2-6484-4fad-aa40-3e73e020841a] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 647.076661] env[61923]: ERROR nova.compute.manager [instance: 55b60fc2-6484-4fad-aa40-3e73e020841a] self.driver.spawn(context, instance, image_meta, [ 647.076661] env[61923]: ERROR nova.compute.manager [instance: 55b60fc2-6484-4fad-aa40-3e73e020841a] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 647.076661] env[61923]: ERROR nova.compute.manager [instance: 55b60fc2-6484-4fad-aa40-3e73e020841a] self._vmops.spawn(context, instance, image_meta, injected_files, [ 647.076661] env[61923]: ERROR nova.compute.manager [instance: 55b60fc2-6484-4fad-aa40-3e73e020841a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 647.076661] env[61923]: ERROR nova.compute.manager [instance: 55b60fc2-6484-4fad-aa40-3e73e020841a] vm_ref = self.build_virtual_machine(instance, [ 647.077288] env[61923]: ERROR nova.compute.manager [instance: 55b60fc2-6484-4fad-aa40-3e73e020841a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 647.077288] env[61923]: ERROR nova.compute.manager [instance: 55b60fc2-6484-4fad-aa40-3e73e020841a] vif_infos = vmwarevif.get_vif_info(self._session, [ 647.077288] env[61923]: ERROR nova.compute.manager [instance: 55b60fc2-6484-4fad-aa40-3e73e020841a] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 647.077288] env[61923]: ERROR nova.compute.manager [instance: 55b60fc2-6484-4fad-aa40-3e73e020841a] for vif in network_info: [ 647.077288] env[61923]: ERROR nova.compute.manager [instance: 55b60fc2-6484-4fad-aa40-3e73e020841a] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 647.077288] env[61923]: ERROR nova.compute.manager [instance: 55b60fc2-6484-4fad-aa40-3e73e020841a] return self._sync_wrapper(fn, *args, **kwargs) [ 647.077288] env[61923]: ERROR nova.compute.manager [instance: 55b60fc2-6484-4fad-aa40-3e73e020841a] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 647.077288] env[61923]: ERROR nova.compute.manager [instance: 55b60fc2-6484-4fad-aa40-3e73e020841a] self.wait() [ 647.077288] env[61923]: ERROR nova.compute.manager [instance: 55b60fc2-6484-4fad-aa40-3e73e020841a] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 647.077288] env[61923]: ERROR nova.compute.manager [instance: 55b60fc2-6484-4fad-aa40-3e73e020841a] self[:] = self._gt.wait() [ 647.077288] env[61923]: ERROR nova.compute.manager [instance: 55b60fc2-6484-4fad-aa40-3e73e020841a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 647.077288] env[61923]: ERROR nova.compute.manager [instance: 55b60fc2-6484-4fad-aa40-3e73e020841a] return self._exit_event.wait() [ 647.077288] env[61923]: ERROR nova.compute.manager [instance: 55b60fc2-6484-4fad-aa40-3e73e020841a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 647.078123] env[61923]: ERROR nova.compute.manager [instance: 55b60fc2-6484-4fad-aa40-3e73e020841a] result = hub.switch() [ 647.078123] env[61923]: ERROR nova.compute.manager [instance: 55b60fc2-6484-4fad-aa40-3e73e020841a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 647.078123] env[61923]: ERROR nova.compute.manager [instance: 55b60fc2-6484-4fad-aa40-3e73e020841a] return self.greenlet.switch() [ 647.078123] env[61923]: ERROR nova.compute.manager [instance: 55b60fc2-6484-4fad-aa40-3e73e020841a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 647.078123] env[61923]: ERROR nova.compute.manager [instance: 55b60fc2-6484-4fad-aa40-3e73e020841a] result = function(*args, **kwargs) [ 647.078123] env[61923]: ERROR nova.compute.manager [instance: 55b60fc2-6484-4fad-aa40-3e73e020841a] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 647.078123] env[61923]: ERROR nova.compute.manager [instance: 55b60fc2-6484-4fad-aa40-3e73e020841a] return func(*args, **kwargs) [ 647.078123] env[61923]: ERROR nova.compute.manager [instance: 55b60fc2-6484-4fad-aa40-3e73e020841a] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 647.078123] env[61923]: ERROR nova.compute.manager [instance: 55b60fc2-6484-4fad-aa40-3e73e020841a] raise e [ 647.078123] env[61923]: ERROR nova.compute.manager [instance: 55b60fc2-6484-4fad-aa40-3e73e020841a] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 647.078123] env[61923]: ERROR nova.compute.manager [instance: 55b60fc2-6484-4fad-aa40-3e73e020841a] nwinfo = self.network_api.allocate_for_instance( [ 647.078123] env[61923]: ERROR nova.compute.manager [instance: 55b60fc2-6484-4fad-aa40-3e73e020841a] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 647.078123] env[61923]: ERROR nova.compute.manager [instance: 55b60fc2-6484-4fad-aa40-3e73e020841a] created_port_ids = self._update_ports_for_instance( [ 647.078811] env[61923]: ERROR nova.compute.manager [instance: 55b60fc2-6484-4fad-aa40-3e73e020841a] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 647.078811] env[61923]: ERROR nova.compute.manager [instance: 55b60fc2-6484-4fad-aa40-3e73e020841a] with excutils.save_and_reraise_exception(): [ 647.078811] env[61923]: ERROR nova.compute.manager [instance: 55b60fc2-6484-4fad-aa40-3e73e020841a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 647.078811] env[61923]: ERROR nova.compute.manager [instance: 55b60fc2-6484-4fad-aa40-3e73e020841a] self.force_reraise() [ 647.078811] env[61923]: ERROR nova.compute.manager [instance: 55b60fc2-6484-4fad-aa40-3e73e020841a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 647.078811] env[61923]: ERROR nova.compute.manager [instance: 55b60fc2-6484-4fad-aa40-3e73e020841a] raise self.value [ 647.078811] env[61923]: ERROR nova.compute.manager [instance: 55b60fc2-6484-4fad-aa40-3e73e020841a] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 647.078811] env[61923]: ERROR nova.compute.manager [instance: 55b60fc2-6484-4fad-aa40-3e73e020841a] updated_port = self._update_port( [ 647.078811] env[61923]: ERROR nova.compute.manager [instance: 55b60fc2-6484-4fad-aa40-3e73e020841a] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 647.078811] env[61923]: ERROR nova.compute.manager [instance: 55b60fc2-6484-4fad-aa40-3e73e020841a] _ensure_no_port_binding_failure(port) [ 647.078811] env[61923]: ERROR nova.compute.manager [instance: 55b60fc2-6484-4fad-aa40-3e73e020841a] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 647.078811] env[61923]: ERROR nova.compute.manager [instance: 55b60fc2-6484-4fad-aa40-3e73e020841a] raise exception.PortBindingFailed(port_id=port['id']) [ 647.079307] env[61923]: ERROR nova.compute.manager [instance: 55b60fc2-6484-4fad-aa40-3e73e020841a] nova.exception.PortBindingFailed: Binding failed for port 445a37d4-005a-49e8-afb5-816754610ddc, please check neutron logs for more information. [ 647.079307] env[61923]: ERROR nova.compute.manager [instance: 55b60fc2-6484-4fad-aa40-3e73e020841a] [ 647.079307] env[61923]: INFO nova.compute.manager [None req-4bd5c12b-d52a-4c54-93f7-87894f09013a tempest-ServersTestJSON-483293076 tempest-ServersTestJSON-483293076-project-member] [instance: 55b60fc2-6484-4fad-aa40-3e73e020841a] Terminating instance [ 647.079307] env[61923]: DEBUG oslo_concurrency.lockutils [None req-4bd5c12b-d52a-4c54-93f7-87894f09013a tempest-ServersTestJSON-483293076 tempest-ServersTestJSON-483293076-project-member] Acquiring lock "refresh_cache-55b60fc2-6484-4fad-aa40-3e73e020841a" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 647.079307] env[61923]: DEBUG oslo_concurrency.lockutils [None req-4bd5c12b-d52a-4c54-93f7-87894f09013a tempest-ServersTestJSON-483293076 tempest-ServersTestJSON-483293076-project-member] Acquired lock "refresh_cache-55b60fc2-6484-4fad-aa40-3e73e020841a" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 647.079307] env[61923]: DEBUG nova.network.neutron [None req-4bd5c12b-d52a-4c54-93f7-87894f09013a tempest-ServersTestJSON-483293076 tempest-ServersTestJSON-483293076-project-member] [instance: 55b60fc2-6484-4fad-aa40-3e73e020841a] Building network info cache for instance {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 647.395266] env[61923]: DEBUG nova.network.neutron [None req-aa9c464b-ac6b-44fe-9384-28cd9bdfecbc tempest-ImagesOneServerTestJSON-63657725 tempest-ImagesOneServerTestJSON-63657725-project-member] [instance: 0c1b8117-199f-4101-ab81-4ae7ef0d1251] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 647.490416] env[61923]: DEBUG nova.network.neutron [None req-aa9c464b-ac6b-44fe-9384-28cd9bdfecbc tempest-ImagesOneServerTestJSON-63657725 tempest-ImagesOneServerTestJSON-63657725-project-member] [instance: 0c1b8117-199f-4101-ab81-4ae7ef0d1251] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 647.555787] env[61923]: DEBUG oslo_concurrency.lockutils [req-df70278a-71cc-469b-bf5f-5bd9e52bb3ae req-037306bf-5792-49cc-85aa-ef64ad16018f service nova] Releasing lock "refresh_cache-525f37d5-2262-40c8-a339-fc262c53ce6e" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 647.556092] env[61923]: DEBUG nova.compute.manager [req-df70278a-71cc-469b-bf5f-5bd9e52bb3ae req-037306bf-5792-49cc-85aa-ef64ad16018f service nova] [instance: 525f37d5-2262-40c8-a339-fc262c53ce6e] Received event network-vif-deleted-0aa60830-cb1d-45ad-8800-f2352dd43766 {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 647.595220] env[61923]: DEBUG nova.network.neutron [None req-4bd5c12b-d52a-4c54-93f7-87894f09013a tempest-ServersTestJSON-483293076 tempest-ServersTestJSON-483293076-project-member] [instance: 55b60fc2-6484-4fad-aa40-3e73e020841a] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 647.662163] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ceb60e8-b085-4614-8fb2-2ed2beef9e1b {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 647.670489] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98454847-e4a5-4543-91d8-d673a8f06869 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 647.701949] env[61923]: DEBUG nova.network.neutron [None req-4bd5c12b-d52a-4c54-93f7-87894f09013a tempest-ServersTestJSON-483293076 tempest-ServersTestJSON-483293076-project-member] [instance: 55b60fc2-6484-4fad-aa40-3e73e020841a] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 647.703602] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f6fa215-2acd-4c04-b326-f71effeca69e {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 647.710672] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-673c8b12-c8c1-4cc5-b43a-b0ef7dfc2b03 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 647.724057] env[61923]: DEBUG nova.compute.provider_tree [None req-0c55849d-766e-4fac-b1c4-e4e97ba688a6 tempest-DeleteServersAdminTestJSON-590116201 tempest-DeleteServersAdminTestJSON-590116201-project-member] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 647.994035] env[61923]: DEBUG oslo_concurrency.lockutils [None req-aa9c464b-ac6b-44fe-9384-28cd9bdfecbc tempest-ImagesOneServerTestJSON-63657725 tempest-ImagesOneServerTestJSON-63657725-project-member] Releasing lock "refresh_cache-0c1b8117-199f-4101-ab81-4ae7ef0d1251" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 647.994035] env[61923]: DEBUG nova.compute.manager [None req-aa9c464b-ac6b-44fe-9384-28cd9bdfecbc tempest-ImagesOneServerTestJSON-63657725 tempest-ImagesOneServerTestJSON-63657725-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61923) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 647.994035] env[61923]: DEBUG nova.compute.manager [None req-aa9c464b-ac6b-44fe-9384-28cd9bdfecbc tempest-ImagesOneServerTestJSON-63657725 tempest-ImagesOneServerTestJSON-63657725-project-member] [instance: 0c1b8117-199f-4101-ab81-4ae7ef0d1251] Deallocating network for instance {{(pid=61923) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 647.994035] env[61923]: DEBUG nova.network.neutron [None req-aa9c464b-ac6b-44fe-9384-28cd9bdfecbc tempest-ImagesOneServerTestJSON-63657725 tempest-ImagesOneServerTestJSON-63657725-project-member] [instance: 0c1b8117-199f-4101-ab81-4ae7ef0d1251] deallocate_for_instance() {{(pid=61923) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 648.017622] env[61923]: DEBUG nova.network.neutron [None req-aa9c464b-ac6b-44fe-9384-28cd9bdfecbc tempest-ImagesOneServerTestJSON-63657725 tempest-ImagesOneServerTestJSON-63657725-project-member] [instance: 0c1b8117-199f-4101-ab81-4ae7ef0d1251] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 648.207623] env[61923]: DEBUG oslo_concurrency.lockutils [None req-4bd5c12b-d52a-4c54-93f7-87894f09013a tempest-ServersTestJSON-483293076 tempest-ServersTestJSON-483293076-project-member] Releasing lock "refresh_cache-55b60fc2-6484-4fad-aa40-3e73e020841a" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 648.208435] env[61923]: DEBUG nova.compute.manager [None req-4bd5c12b-d52a-4c54-93f7-87894f09013a tempest-ServersTestJSON-483293076 tempest-ServersTestJSON-483293076-project-member] [instance: 55b60fc2-6484-4fad-aa40-3e73e020841a] Start destroying the instance on the hypervisor. {{(pid=61923) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 648.208679] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-4bd5c12b-d52a-4c54-93f7-87894f09013a tempest-ServersTestJSON-483293076 tempest-ServersTestJSON-483293076-project-member] [instance: 55b60fc2-6484-4fad-aa40-3e73e020841a] Destroying instance {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 648.208992] env[61923]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-992cc267-97d9-4532-ac13-4163f850bed1 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 648.218359] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae607ff1-4793-451e-ad81-1049db35e98c {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 648.229041] env[61923]: DEBUG nova.scheduler.client.report [None req-0c55849d-766e-4fac-b1c4-e4e97ba688a6 tempest-DeleteServersAdminTestJSON-590116201 tempest-DeleteServersAdminTestJSON-590116201-project-member] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 648.248225] env[61923]: WARNING nova.virt.vmwareapi.vmops [None req-4bd5c12b-d52a-4c54-93f7-87894f09013a tempest-ServersTestJSON-483293076 tempest-ServersTestJSON-483293076-project-member] [instance: 55b60fc2-6484-4fad-aa40-3e73e020841a] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 55b60fc2-6484-4fad-aa40-3e73e020841a could not be found. [ 648.248533] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-4bd5c12b-d52a-4c54-93f7-87894f09013a tempest-ServersTestJSON-483293076 tempest-ServersTestJSON-483293076-project-member] [instance: 55b60fc2-6484-4fad-aa40-3e73e020841a] Instance destroyed {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 648.248940] env[61923]: INFO nova.compute.manager [None req-4bd5c12b-d52a-4c54-93f7-87894f09013a tempest-ServersTestJSON-483293076 tempest-ServersTestJSON-483293076-project-member] [instance: 55b60fc2-6484-4fad-aa40-3e73e020841a] Took 0.04 seconds to destroy the instance on the hypervisor. [ 648.249236] env[61923]: DEBUG oslo.service.loopingcall [None req-4bd5c12b-d52a-4c54-93f7-87894f09013a tempest-ServersTestJSON-483293076 tempest-ServersTestJSON-483293076-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61923) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 648.250099] env[61923]: DEBUG nova.compute.manager [-] [instance: 55b60fc2-6484-4fad-aa40-3e73e020841a] Deallocating network for instance {{(pid=61923) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 648.250202] env[61923]: DEBUG nova.network.neutron [-] [instance: 55b60fc2-6484-4fad-aa40-3e73e020841a] deallocate_for_instance() {{(pid=61923) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 648.263973] env[61923]: DEBUG nova.network.neutron [-] [instance: 55b60fc2-6484-4fad-aa40-3e73e020841a] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 648.391751] env[61923]: DEBUG nova.compute.manager [req-75d0b83c-c537-4675-a73a-8388131b50a4 req-f075f610-19da-483b-b8bc-95396f922996 service nova] [instance: 55b60fc2-6484-4fad-aa40-3e73e020841a] Received event network-changed-445a37d4-005a-49e8-afb5-816754610ddc {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 648.393141] env[61923]: DEBUG nova.compute.manager [req-75d0b83c-c537-4675-a73a-8388131b50a4 req-f075f610-19da-483b-b8bc-95396f922996 service nova] [instance: 55b60fc2-6484-4fad-aa40-3e73e020841a] Refreshing instance network info cache due to event network-changed-445a37d4-005a-49e8-afb5-816754610ddc. {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 648.393141] env[61923]: DEBUG oslo_concurrency.lockutils [req-75d0b83c-c537-4675-a73a-8388131b50a4 req-f075f610-19da-483b-b8bc-95396f922996 service nova] Acquiring lock "refresh_cache-55b60fc2-6484-4fad-aa40-3e73e020841a" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 648.393141] env[61923]: DEBUG oslo_concurrency.lockutils [req-75d0b83c-c537-4675-a73a-8388131b50a4 req-f075f610-19da-483b-b8bc-95396f922996 service nova] Acquired lock "refresh_cache-55b60fc2-6484-4fad-aa40-3e73e020841a" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 648.393141] env[61923]: DEBUG nova.network.neutron [req-75d0b83c-c537-4675-a73a-8388131b50a4 req-f075f610-19da-483b-b8bc-95396f922996 service nova] [instance: 55b60fc2-6484-4fad-aa40-3e73e020841a] Refreshing network info cache for port 445a37d4-005a-49e8-afb5-816754610ddc {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 648.520364] env[61923]: DEBUG nova.network.neutron [None req-aa9c464b-ac6b-44fe-9384-28cd9bdfecbc tempest-ImagesOneServerTestJSON-63657725 tempest-ImagesOneServerTestJSON-63657725-project-member] [instance: 0c1b8117-199f-4101-ab81-4ae7ef0d1251] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 648.735178] env[61923]: DEBUG oslo_concurrency.lockutils [None req-0c55849d-766e-4fac-b1c4-e4e97ba688a6 tempest-DeleteServersAdminTestJSON-590116201 tempest-DeleteServersAdminTestJSON-590116201-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.867s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 648.735870] env[61923]: ERROR nova.compute.manager [None req-0c55849d-766e-4fac-b1c4-e4e97ba688a6 tempest-DeleteServersAdminTestJSON-590116201 tempest-DeleteServersAdminTestJSON-590116201-project-member] [instance: 1b509288-2df3-4927-afcf-cb76918d3da3] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port dbf4bfa8-cc4a-4be3-a70c-7f709e040750, please check neutron logs for more information. [ 648.735870] env[61923]: ERROR nova.compute.manager [instance: 1b509288-2df3-4927-afcf-cb76918d3da3] Traceback (most recent call last): [ 648.735870] env[61923]: ERROR nova.compute.manager [instance: 1b509288-2df3-4927-afcf-cb76918d3da3] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 648.735870] env[61923]: ERROR nova.compute.manager [instance: 1b509288-2df3-4927-afcf-cb76918d3da3] self.driver.spawn(context, instance, image_meta, [ 648.735870] env[61923]: ERROR nova.compute.manager [instance: 1b509288-2df3-4927-afcf-cb76918d3da3] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 648.735870] env[61923]: ERROR nova.compute.manager [instance: 1b509288-2df3-4927-afcf-cb76918d3da3] self._vmops.spawn(context, instance, image_meta, injected_files, [ 648.735870] env[61923]: ERROR nova.compute.manager [instance: 1b509288-2df3-4927-afcf-cb76918d3da3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 648.735870] env[61923]: ERROR nova.compute.manager [instance: 1b509288-2df3-4927-afcf-cb76918d3da3] vm_ref = self.build_virtual_machine(instance, [ 648.735870] env[61923]: ERROR nova.compute.manager [instance: 1b509288-2df3-4927-afcf-cb76918d3da3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 648.735870] env[61923]: ERROR nova.compute.manager [instance: 1b509288-2df3-4927-afcf-cb76918d3da3] vif_infos = vmwarevif.get_vif_info(self._session, [ 648.735870] env[61923]: ERROR nova.compute.manager [instance: 1b509288-2df3-4927-afcf-cb76918d3da3] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 648.736224] env[61923]: ERROR nova.compute.manager [instance: 1b509288-2df3-4927-afcf-cb76918d3da3] for vif in network_info: [ 648.736224] env[61923]: ERROR nova.compute.manager [instance: 1b509288-2df3-4927-afcf-cb76918d3da3] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 648.736224] env[61923]: ERROR nova.compute.manager [instance: 1b509288-2df3-4927-afcf-cb76918d3da3] return self._sync_wrapper(fn, *args, **kwargs) [ 648.736224] env[61923]: ERROR nova.compute.manager [instance: 1b509288-2df3-4927-afcf-cb76918d3da3] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 648.736224] env[61923]: ERROR nova.compute.manager [instance: 1b509288-2df3-4927-afcf-cb76918d3da3] self.wait() [ 648.736224] env[61923]: ERROR nova.compute.manager [instance: 1b509288-2df3-4927-afcf-cb76918d3da3] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 648.736224] env[61923]: ERROR nova.compute.manager [instance: 1b509288-2df3-4927-afcf-cb76918d3da3] self[:] = self._gt.wait() [ 648.736224] env[61923]: ERROR nova.compute.manager [instance: 1b509288-2df3-4927-afcf-cb76918d3da3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 648.736224] env[61923]: ERROR nova.compute.manager [instance: 1b509288-2df3-4927-afcf-cb76918d3da3] return self._exit_event.wait() [ 648.736224] env[61923]: ERROR nova.compute.manager [instance: 1b509288-2df3-4927-afcf-cb76918d3da3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 648.736224] env[61923]: ERROR nova.compute.manager [instance: 1b509288-2df3-4927-afcf-cb76918d3da3] result = hub.switch() [ 648.736224] env[61923]: ERROR nova.compute.manager [instance: 1b509288-2df3-4927-afcf-cb76918d3da3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 648.736224] env[61923]: ERROR nova.compute.manager [instance: 1b509288-2df3-4927-afcf-cb76918d3da3] return self.greenlet.switch() [ 648.736587] env[61923]: ERROR nova.compute.manager [instance: 1b509288-2df3-4927-afcf-cb76918d3da3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 648.736587] env[61923]: ERROR nova.compute.manager [instance: 1b509288-2df3-4927-afcf-cb76918d3da3] result = function(*args, **kwargs) [ 648.736587] env[61923]: ERROR nova.compute.manager [instance: 1b509288-2df3-4927-afcf-cb76918d3da3] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 648.736587] env[61923]: ERROR nova.compute.manager [instance: 1b509288-2df3-4927-afcf-cb76918d3da3] return func(*args, **kwargs) [ 648.736587] env[61923]: ERROR nova.compute.manager [instance: 1b509288-2df3-4927-afcf-cb76918d3da3] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 648.736587] env[61923]: ERROR nova.compute.manager [instance: 1b509288-2df3-4927-afcf-cb76918d3da3] raise e [ 648.736587] env[61923]: ERROR nova.compute.manager [instance: 1b509288-2df3-4927-afcf-cb76918d3da3] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 648.736587] env[61923]: ERROR nova.compute.manager [instance: 1b509288-2df3-4927-afcf-cb76918d3da3] nwinfo = self.network_api.allocate_for_instance( [ 648.736587] env[61923]: ERROR nova.compute.manager [instance: 1b509288-2df3-4927-afcf-cb76918d3da3] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 648.736587] env[61923]: ERROR nova.compute.manager [instance: 1b509288-2df3-4927-afcf-cb76918d3da3] created_port_ids = self._update_ports_for_instance( [ 648.736587] env[61923]: ERROR nova.compute.manager [instance: 1b509288-2df3-4927-afcf-cb76918d3da3] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 648.736587] env[61923]: ERROR nova.compute.manager [instance: 1b509288-2df3-4927-afcf-cb76918d3da3] with excutils.save_and_reraise_exception(): [ 648.736587] env[61923]: ERROR nova.compute.manager [instance: 1b509288-2df3-4927-afcf-cb76918d3da3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 648.736946] env[61923]: ERROR nova.compute.manager [instance: 1b509288-2df3-4927-afcf-cb76918d3da3] self.force_reraise() [ 648.736946] env[61923]: ERROR nova.compute.manager [instance: 1b509288-2df3-4927-afcf-cb76918d3da3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 648.736946] env[61923]: ERROR nova.compute.manager [instance: 1b509288-2df3-4927-afcf-cb76918d3da3] raise self.value [ 648.736946] env[61923]: ERROR nova.compute.manager [instance: 1b509288-2df3-4927-afcf-cb76918d3da3] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 648.736946] env[61923]: ERROR nova.compute.manager [instance: 1b509288-2df3-4927-afcf-cb76918d3da3] updated_port = self._update_port( [ 648.736946] env[61923]: ERROR nova.compute.manager [instance: 1b509288-2df3-4927-afcf-cb76918d3da3] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 648.736946] env[61923]: ERROR nova.compute.manager [instance: 1b509288-2df3-4927-afcf-cb76918d3da3] _ensure_no_port_binding_failure(port) [ 648.736946] env[61923]: ERROR nova.compute.manager [instance: 1b509288-2df3-4927-afcf-cb76918d3da3] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 648.736946] env[61923]: ERROR nova.compute.manager [instance: 1b509288-2df3-4927-afcf-cb76918d3da3] raise exception.PortBindingFailed(port_id=port['id']) [ 648.736946] env[61923]: ERROR nova.compute.manager [instance: 1b509288-2df3-4927-afcf-cb76918d3da3] nova.exception.PortBindingFailed: Binding failed for port dbf4bfa8-cc4a-4be3-a70c-7f709e040750, please check neutron logs for more information. [ 648.736946] env[61923]: ERROR nova.compute.manager [instance: 1b509288-2df3-4927-afcf-cb76918d3da3] [ 648.737283] env[61923]: DEBUG nova.compute.utils [None req-0c55849d-766e-4fac-b1c4-e4e97ba688a6 tempest-DeleteServersAdminTestJSON-590116201 tempest-DeleteServersAdminTestJSON-590116201-project-member] [instance: 1b509288-2df3-4927-afcf-cb76918d3da3] Binding failed for port dbf4bfa8-cc4a-4be3-a70c-7f709e040750, please check neutron logs for more information. {{(pid=61923) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 648.737823] env[61923]: DEBUG oslo_concurrency.lockutils [None req-56950b6c-0b87-48bf-a575-f699fcd1a5b0 tempest-FloatingIPsAssociationTestJSON-1125124803 tempest-FloatingIPsAssociationTestJSON-1125124803-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.516s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 648.740822] env[61923]: INFO nova.compute.claims [None req-56950b6c-0b87-48bf-a575-f699fcd1a5b0 tempest-FloatingIPsAssociationTestJSON-1125124803 tempest-FloatingIPsAssociationTestJSON-1125124803-project-member] [instance: aeca6a94-478d-435a-9d93-59936f26654e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 648.744047] env[61923]: DEBUG nova.compute.manager [None req-0c55849d-766e-4fac-b1c4-e4e97ba688a6 tempest-DeleteServersAdminTestJSON-590116201 tempest-DeleteServersAdminTestJSON-590116201-project-member] [instance: 1b509288-2df3-4927-afcf-cb76918d3da3] Build of instance 1b509288-2df3-4927-afcf-cb76918d3da3 was re-scheduled: Binding failed for port dbf4bfa8-cc4a-4be3-a70c-7f709e040750, please check neutron logs for more information. {{(pid=61923) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 648.744304] env[61923]: DEBUG nova.compute.manager [None req-0c55849d-766e-4fac-b1c4-e4e97ba688a6 tempest-DeleteServersAdminTestJSON-590116201 tempest-DeleteServersAdminTestJSON-590116201-project-member] [instance: 1b509288-2df3-4927-afcf-cb76918d3da3] Unplugging VIFs for instance {{(pid=61923) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 648.744528] env[61923]: DEBUG oslo_concurrency.lockutils [None req-0c55849d-766e-4fac-b1c4-e4e97ba688a6 tempest-DeleteServersAdminTestJSON-590116201 tempest-DeleteServersAdminTestJSON-590116201-project-member] Acquiring lock "refresh_cache-1b509288-2df3-4927-afcf-cb76918d3da3" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 648.744672] env[61923]: DEBUG oslo_concurrency.lockutils [None req-0c55849d-766e-4fac-b1c4-e4e97ba688a6 tempest-DeleteServersAdminTestJSON-590116201 tempest-DeleteServersAdminTestJSON-590116201-project-member] Acquired lock "refresh_cache-1b509288-2df3-4927-afcf-cb76918d3da3" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 648.744829] env[61923]: DEBUG nova.network.neutron [None req-0c55849d-766e-4fac-b1c4-e4e97ba688a6 tempest-DeleteServersAdminTestJSON-590116201 tempest-DeleteServersAdminTestJSON-590116201-project-member] [instance: 1b509288-2df3-4927-afcf-cb76918d3da3] Building network info cache for instance {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 648.766752] env[61923]: DEBUG nova.network.neutron [-] [instance: 55b60fc2-6484-4fad-aa40-3e73e020841a] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 649.024744] env[61923]: INFO nova.compute.manager [None req-aa9c464b-ac6b-44fe-9384-28cd9bdfecbc tempest-ImagesOneServerTestJSON-63657725 tempest-ImagesOneServerTestJSON-63657725-project-member] [instance: 0c1b8117-199f-4101-ab81-4ae7ef0d1251] Took 1.03 seconds to deallocate network for instance. [ 649.034751] env[61923]: DEBUG nova.network.neutron [req-75d0b83c-c537-4675-a73a-8388131b50a4 req-f075f610-19da-483b-b8bc-95396f922996 service nova] [instance: 55b60fc2-6484-4fad-aa40-3e73e020841a] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 649.150174] env[61923]: DEBUG nova.network.neutron [req-75d0b83c-c537-4675-a73a-8388131b50a4 req-f075f610-19da-483b-b8bc-95396f922996 service nova] [instance: 55b60fc2-6484-4fad-aa40-3e73e020841a] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 649.269386] env[61923]: INFO nova.compute.manager [-] [instance: 55b60fc2-6484-4fad-aa40-3e73e020841a] Took 1.02 seconds to deallocate network for instance. [ 649.271640] env[61923]: DEBUG nova.compute.claims [None req-4bd5c12b-d52a-4c54-93f7-87894f09013a tempest-ServersTestJSON-483293076 tempest-ServersTestJSON-483293076-project-member] [instance: 55b60fc2-6484-4fad-aa40-3e73e020841a] Aborting claim: {{(pid=61923) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 649.271729] env[61923]: DEBUG oslo_concurrency.lockutils [None req-4bd5c12b-d52a-4c54-93f7-87894f09013a tempest-ServersTestJSON-483293076 tempest-ServersTestJSON-483293076-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 649.273362] env[61923]: DEBUG nova.network.neutron [None req-0c55849d-766e-4fac-b1c4-e4e97ba688a6 tempest-DeleteServersAdminTestJSON-590116201 tempest-DeleteServersAdminTestJSON-590116201-project-member] [instance: 1b509288-2df3-4927-afcf-cb76918d3da3] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 649.337533] env[61923]: DEBUG nova.network.neutron [None req-0c55849d-766e-4fac-b1c4-e4e97ba688a6 tempest-DeleteServersAdminTestJSON-590116201 tempest-DeleteServersAdminTestJSON-590116201-project-member] [instance: 1b509288-2df3-4927-afcf-cb76918d3da3] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 649.652434] env[61923]: DEBUG oslo_concurrency.lockutils [req-75d0b83c-c537-4675-a73a-8388131b50a4 req-f075f610-19da-483b-b8bc-95396f922996 service nova] Releasing lock "refresh_cache-55b60fc2-6484-4fad-aa40-3e73e020841a" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 649.652662] env[61923]: DEBUG nova.compute.manager [req-75d0b83c-c537-4675-a73a-8388131b50a4 req-f075f610-19da-483b-b8bc-95396f922996 service nova] [instance: 55b60fc2-6484-4fad-aa40-3e73e020841a] Received event network-vif-deleted-445a37d4-005a-49e8-afb5-816754610ddc {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 649.841893] env[61923]: DEBUG oslo_concurrency.lockutils [None req-0c55849d-766e-4fac-b1c4-e4e97ba688a6 tempest-DeleteServersAdminTestJSON-590116201 tempest-DeleteServersAdminTestJSON-590116201-project-member] Releasing lock "refresh_cache-1b509288-2df3-4927-afcf-cb76918d3da3" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 649.842159] env[61923]: DEBUG nova.compute.manager [None req-0c55849d-766e-4fac-b1c4-e4e97ba688a6 tempest-DeleteServersAdminTestJSON-590116201 tempest-DeleteServersAdminTestJSON-590116201-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61923) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 649.842296] env[61923]: DEBUG nova.compute.manager [None req-0c55849d-766e-4fac-b1c4-e4e97ba688a6 tempest-DeleteServersAdminTestJSON-590116201 tempest-DeleteServersAdminTestJSON-590116201-project-member] [instance: 1b509288-2df3-4927-afcf-cb76918d3da3] Deallocating network for instance {{(pid=61923) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 649.842459] env[61923]: DEBUG nova.network.neutron [None req-0c55849d-766e-4fac-b1c4-e4e97ba688a6 tempest-DeleteServersAdminTestJSON-590116201 tempest-DeleteServersAdminTestJSON-590116201-project-member] [instance: 1b509288-2df3-4927-afcf-cb76918d3da3] deallocate_for_instance() {{(pid=61923) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 649.856812] env[61923]: DEBUG nova.network.neutron [None req-0c55849d-766e-4fac-b1c4-e4e97ba688a6 tempest-DeleteServersAdminTestJSON-590116201 tempest-DeleteServersAdminTestJSON-590116201-project-member] [instance: 1b509288-2df3-4927-afcf-cb76918d3da3] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 650.006754] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f826ca4a-f05a-4aab-8f4e-a9fff71a7223 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 650.014097] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75b3e73d-88da-4f4b-8240-2234655ab80b {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 650.044955] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee42d1b4-a41e-499d-8857-0e9265003d15 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 650.052116] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70099e7a-4482-4af1-a49c-733b804a20ee {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 650.067125] env[61923]: DEBUG nova.compute.provider_tree [None req-56950b6c-0b87-48bf-a575-f699fcd1a5b0 tempest-FloatingIPsAssociationTestJSON-1125124803 tempest-FloatingIPsAssociationTestJSON-1125124803-project-member] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 650.069134] env[61923]: INFO nova.scheduler.client.report [None req-aa9c464b-ac6b-44fe-9384-28cd9bdfecbc tempest-ImagesOneServerTestJSON-63657725 tempest-ImagesOneServerTestJSON-63657725-project-member] Deleted allocations for instance 0c1b8117-199f-4101-ab81-4ae7ef0d1251 [ 650.359337] env[61923]: DEBUG nova.network.neutron [None req-0c55849d-766e-4fac-b1c4-e4e97ba688a6 tempest-DeleteServersAdminTestJSON-590116201 tempest-DeleteServersAdminTestJSON-590116201-project-member] [instance: 1b509288-2df3-4927-afcf-cb76918d3da3] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 650.579021] env[61923]: DEBUG nova.scheduler.client.report [None req-56950b6c-0b87-48bf-a575-f699fcd1a5b0 tempest-FloatingIPsAssociationTestJSON-1125124803 tempest-FloatingIPsAssociationTestJSON-1125124803-project-member] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 650.580574] env[61923]: DEBUG oslo_concurrency.lockutils [None req-aa9c464b-ac6b-44fe-9384-28cd9bdfecbc tempest-ImagesOneServerTestJSON-63657725 tempest-ImagesOneServerTestJSON-63657725-project-member] Lock "0c1b8117-199f-4101-ab81-4ae7ef0d1251" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 105.660s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 650.861807] env[61923]: INFO nova.compute.manager [None req-0c55849d-766e-4fac-b1c4-e4e97ba688a6 tempest-DeleteServersAdminTestJSON-590116201 tempest-DeleteServersAdminTestJSON-590116201-project-member] [instance: 1b509288-2df3-4927-afcf-cb76918d3da3] Took 1.02 seconds to deallocate network for instance. [ 651.083627] env[61923]: DEBUG oslo_concurrency.lockutils [None req-56950b6c-0b87-48bf-a575-f699fcd1a5b0 tempest-FloatingIPsAssociationTestJSON-1125124803 tempest-FloatingIPsAssociationTestJSON-1125124803-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.346s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 651.084157] env[61923]: DEBUG nova.compute.manager [None req-56950b6c-0b87-48bf-a575-f699fcd1a5b0 tempest-FloatingIPsAssociationTestJSON-1125124803 tempest-FloatingIPsAssociationTestJSON-1125124803-project-member] [instance: aeca6a94-478d-435a-9d93-59936f26654e] Start building networks asynchronously for instance. {{(pid=61923) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 651.086811] env[61923]: DEBUG nova.compute.manager [None req-1f24f04c-cf4b-4940-9447-eacc11fa5e66 tempest-ServersTestFqdnHostnames-345291565 tempest-ServersTestFqdnHostnames-345291565-project-member] [instance: c94e8e46-7697-426c-ae2f-aece493fa8f4] Starting instance... {{(pid=61923) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 651.090800] env[61923]: DEBUG oslo_concurrency.lockutils [None req-95b9b70a-a835-4901-9a6d-c0fa1b2217f3 tempest-MigrationsAdminTest-1311988845 tempest-MigrationsAdminTest-1311988845-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.783s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 651.598392] env[61923]: DEBUG nova.compute.utils [None req-56950b6c-0b87-48bf-a575-f699fcd1a5b0 tempest-FloatingIPsAssociationTestJSON-1125124803 tempest-FloatingIPsAssociationTestJSON-1125124803-project-member] Using /dev/sd instead of None {{(pid=61923) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 651.603963] env[61923]: DEBUG nova.compute.manager [None req-56950b6c-0b87-48bf-a575-f699fcd1a5b0 tempest-FloatingIPsAssociationTestJSON-1125124803 tempest-FloatingIPsAssociationTestJSON-1125124803-project-member] [instance: aeca6a94-478d-435a-9d93-59936f26654e] Allocating IP information in the background. {{(pid=61923) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 651.604262] env[61923]: DEBUG nova.network.neutron [None req-56950b6c-0b87-48bf-a575-f699fcd1a5b0 tempest-FloatingIPsAssociationTestJSON-1125124803 tempest-FloatingIPsAssociationTestJSON-1125124803-project-member] [instance: aeca6a94-478d-435a-9d93-59936f26654e] allocate_for_instance() {{(pid=61923) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 651.621765] env[61923]: DEBUG oslo_concurrency.lockutils [None req-1f24f04c-cf4b-4940-9447-eacc11fa5e66 tempest-ServersTestFqdnHostnames-345291565 tempest-ServersTestFqdnHostnames-345291565-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 651.705421] env[61923]: DEBUG nova.policy [None req-56950b6c-0b87-48bf-a575-f699fcd1a5b0 tempest-FloatingIPsAssociationTestJSON-1125124803 tempest-FloatingIPsAssociationTestJSON-1125124803-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '22957897a47942b0a145ae71698d1fc3', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '356aace930ba4a8a95e0832300136c25', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61923) authorize /opt/stack/nova/nova/policy.py:201}} [ 651.897130] env[61923]: INFO nova.scheduler.client.report [None req-0c55849d-766e-4fac-b1c4-e4e97ba688a6 tempest-DeleteServersAdminTestJSON-590116201 tempest-DeleteServersAdminTestJSON-590116201-project-member] Deleted allocations for instance 1b509288-2df3-4927-afcf-cb76918d3da3 [ 651.998263] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c31935c-7d05-4a78-96ec-97ea4c17225e {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.007251] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7ad10de-8911-4428-abc9-3e32e5f1aa90 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.037615] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53a83c63-c601-4b0e-9db3-d426c84c87bf {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.047296] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b998ed26-a2a5-4fa9-96f9-129a75e5e394 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.062487] env[61923]: DEBUG nova.compute.provider_tree [None req-95b9b70a-a835-4901-9a6d-c0fa1b2217f3 tempest-MigrationsAdminTest-1311988845 tempest-MigrationsAdminTest-1311988845-project-member] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 652.103257] env[61923]: DEBUG nova.compute.manager [None req-56950b6c-0b87-48bf-a575-f699fcd1a5b0 tempest-FloatingIPsAssociationTestJSON-1125124803 tempest-FloatingIPsAssociationTestJSON-1125124803-project-member] [instance: aeca6a94-478d-435a-9d93-59936f26654e] Start building block device mappings for instance. {{(pid=61923) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 652.122121] env[61923]: DEBUG nova.network.neutron [None req-56950b6c-0b87-48bf-a575-f699fcd1a5b0 tempest-FloatingIPsAssociationTestJSON-1125124803 tempest-FloatingIPsAssociationTestJSON-1125124803-project-member] [instance: aeca6a94-478d-435a-9d93-59936f26654e] Successfully created port: 61e9a5e4-10ed-40b2-a961-2d5cbae80f02 {{(pid=61923) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 652.409046] env[61923]: DEBUG oslo_concurrency.lockutils [None req-0c55849d-766e-4fac-b1c4-e4e97ba688a6 tempest-DeleteServersAdminTestJSON-590116201 tempest-DeleteServersAdminTestJSON-590116201-project-member] Lock "1b509288-2df3-4927-afcf-cb76918d3da3" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 105.936s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 652.565777] env[61923]: DEBUG nova.scheduler.client.report [None req-95b9b70a-a835-4901-9a6d-c0fa1b2217f3 tempest-MigrationsAdminTest-1311988845 tempest-MigrationsAdminTest-1311988845-project-member] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 652.913669] env[61923]: DEBUG nova.compute.manager [None req-aadda1bb-f770-4742-8c34-cd25e3f7e7cb tempest-InstanceActionsNegativeTestJSON-1700332362 tempest-InstanceActionsNegativeTestJSON-1700332362-project-member] [instance: 64584976-b3f3-4da5-a76b-a05cf2ed6aa4] Starting instance... {{(pid=61923) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 653.070559] env[61923]: DEBUG oslo_concurrency.lockutils [None req-95b9b70a-a835-4901-9a6d-c0fa1b2217f3 tempest-MigrationsAdminTest-1311988845 tempest-MigrationsAdminTest-1311988845-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.980s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 653.071289] env[61923]: ERROR nova.compute.manager [None req-95b9b70a-a835-4901-9a6d-c0fa1b2217f3 tempest-MigrationsAdminTest-1311988845 tempest-MigrationsAdminTest-1311988845-project-member] [instance: a4b54a7a-3f3d-49ae-ba43-5a10c262d4e5] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 2110c7b3-1dff-4f09-b631-618ae3b8467e, please check neutron logs for more information. [ 653.071289] env[61923]: ERROR nova.compute.manager [instance: a4b54a7a-3f3d-49ae-ba43-5a10c262d4e5] Traceback (most recent call last): [ 653.071289] env[61923]: ERROR nova.compute.manager [instance: a4b54a7a-3f3d-49ae-ba43-5a10c262d4e5] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 653.071289] env[61923]: ERROR nova.compute.manager [instance: a4b54a7a-3f3d-49ae-ba43-5a10c262d4e5] self.driver.spawn(context, instance, image_meta, [ 653.071289] env[61923]: ERROR nova.compute.manager [instance: a4b54a7a-3f3d-49ae-ba43-5a10c262d4e5] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 653.071289] env[61923]: ERROR nova.compute.manager [instance: a4b54a7a-3f3d-49ae-ba43-5a10c262d4e5] self._vmops.spawn(context, instance, image_meta, injected_files, [ 653.071289] env[61923]: ERROR nova.compute.manager [instance: a4b54a7a-3f3d-49ae-ba43-5a10c262d4e5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 653.071289] env[61923]: ERROR nova.compute.manager [instance: a4b54a7a-3f3d-49ae-ba43-5a10c262d4e5] vm_ref = self.build_virtual_machine(instance, [ 653.071289] env[61923]: ERROR nova.compute.manager [instance: a4b54a7a-3f3d-49ae-ba43-5a10c262d4e5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 653.071289] env[61923]: ERROR nova.compute.manager [instance: a4b54a7a-3f3d-49ae-ba43-5a10c262d4e5] vif_infos = vmwarevif.get_vif_info(self._session, [ 653.071289] env[61923]: ERROR nova.compute.manager [instance: a4b54a7a-3f3d-49ae-ba43-5a10c262d4e5] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 653.071575] env[61923]: ERROR nova.compute.manager [instance: a4b54a7a-3f3d-49ae-ba43-5a10c262d4e5] for vif in network_info: [ 653.071575] env[61923]: ERROR nova.compute.manager [instance: a4b54a7a-3f3d-49ae-ba43-5a10c262d4e5] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 653.071575] env[61923]: ERROR nova.compute.manager [instance: a4b54a7a-3f3d-49ae-ba43-5a10c262d4e5] return self._sync_wrapper(fn, *args, **kwargs) [ 653.071575] env[61923]: ERROR nova.compute.manager [instance: a4b54a7a-3f3d-49ae-ba43-5a10c262d4e5] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 653.071575] env[61923]: ERROR nova.compute.manager [instance: a4b54a7a-3f3d-49ae-ba43-5a10c262d4e5] self.wait() [ 653.071575] env[61923]: ERROR nova.compute.manager [instance: a4b54a7a-3f3d-49ae-ba43-5a10c262d4e5] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 653.071575] env[61923]: ERROR nova.compute.manager [instance: a4b54a7a-3f3d-49ae-ba43-5a10c262d4e5] self[:] = self._gt.wait() [ 653.071575] env[61923]: ERROR nova.compute.manager [instance: a4b54a7a-3f3d-49ae-ba43-5a10c262d4e5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 653.071575] env[61923]: ERROR nova.compute.manager [instance: a4b54a7a-3f3d-49ae-ba43-5a10c262d4e5] return self._exit_event.wait() [ 653.071575] env[61923]: ERROR nova.compute.manager [instance: a4b54a7a-3f3d-49ae-ba43-5a10c262d4e5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 653.071575] env[61923]: ERROR nova.compute.manager [instance: a4b54a7a-3f3d-49ae-ba43-5a10c262d4e5] result = hub.switch() [ 653.071575] env[61923]: ERROR nova.compute.manager [instance: a4b54a7a-3f3d-49ae-ba43-5a10c262d4e5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 653.071575] env[61923]: ERROR nova.compute.manager [instance: a4b54a7a-3f3d-49ae-ba43-5a10c262d4e5] return self.greenlet.switch() [ 653.072157] env[61923]: ERROR nova.compute.manager [instance: a4b54a7a-3f3d-49ae-ba43-5a10c262d4e5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 653.072157] env[61923]: ERROR nova.compute.manager [instance: a4b54a7a-3f3d-49ae-ba43-5a10c262d4e5] result = function(*args, **kwargs) [ 653.072157] env[61923]: ERROR nova.compute.manager [instance: a4b54a7a-3f3d-49ae-ba43-5a10c262d4e5] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 653.072157] env[61923]: ERROR nova.compute.manager [instance: a4b54a7a-3f3d-49ae-ba43-5a10c262d4e5] return func(*args, **kwargs) [ 653.072157] env[61923]: ERROR nova.compute.manager [instance: a4b54a7a-3f3d-49ae-ba43-5a10c262d4e5] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 653.072157] env[61923]: ERROR nova.compute.manager [instance: a4b54a7a-3f3d-49ae-ba43-5a10c262d4e5] raise e [ 653.072157] env[61923]: ERROR nova.compute.manager [instance: a4b54a7a-3f3d-49ae-ba43-5a10c262d4e5] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 653.072157] env[61923]: ERROR nova.compute.manager [instance: a4b54a7a-3f3d-49ae-ba43-5a10c262d4e5] nwinfo = self.network_api.allocate_for_instance( [ 653.072157] env[61923]: ERROR nova.compute.manager [instance: a4b54a7a-3f3d-49ae-ba43-5a10c262d4e5] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 653.072157] env[61923]: ERROR nova.compute.manager [instance: a4b54a7a-3f3d-49ae-ba43-5a10c262d4e5] created_port_ids = self._update_ports_for_instance( [ 653.072157] env[61923]: ERROR nova.compute.manager [instance: a4b54a7a-3f3d-49ae-ba43-5a10c262d4e5] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 653.072157] env[61923]: ERROR nova.compute.manager [instance: a4b54a7a-3f3d-49ae-ba43-5a10c262d4e5] with excutils.save_and_reraise_exception(): [ 653.072157] env[61923]: ERROR nova.compute.manager [instance: a4b54a7a-3f3d-49ae-ba43-5a10c262d4e5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 653.072480] env[61923]: ERROR nova.compute.manager [instance: a4b54a7a-3f3d-49ae-ba43-5a10c262d4e5] self.force_reraise() [ 653.072480] env[61923]: ERROR nova.compute.manager [instance: a4b54a7a-3f3d-49ae-ba43-5a10c262d4e5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 653.072480] env[61923]: ERROR nova.compute.manager [instance: a4b54a7a-3f3d-49ae-ba43-5a10c262d4e5] raise self.value [ 653.072480] env[61923]: ERROR nova.compute.manager [instance: a4b54a7a-3f3d-49ae-ba43-5a10c262d4e5] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 653.072480] env[61923]: ERROR nova.compute.manager [instance: a4b54a7a-3f3d-49ae-ba43-5a10c262d4e5] updated_port = self._update_port( [ 653.072480] env[61923]: ERROR nova.compute.manager [instance: a4b54a7a-3f3d-49ae-ba43-5a10c262d4e5] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 653.072480] env[61923]: ERROR nova.compute.manager [instance: a4b54a7a-3f3d-49ae-ba43-5a10c262d4e5] _ensure_no_port_binding_failure(port) [ 653.072480] env[61923]: ERROR nova.compute.manager [instance: a4b54a7a-3f3d-49ae-ba43-5a10c262d4e5] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 653.072480] env[61923]: ERROR nova.compute.manager [instance: a4b54a7a-3f3d-49ae-ba43-5a10c262d4e5] raise exception.PortBindingFailed(port_id=port['id']) [ 653.072480] env[61923]: ERROR nova.compute.manager [instance: a4b54a7a-3f3d-49ae-ba43-5a10c262d4e5] nova.exception.PortBindingFailed: Binding failed for port 2110c7b3-1dff-4f09-b631-618ae3b8467e, please check neutron logs for more information. [ 653.072480] env[61923]: ERROR nova.compute.manager [instance: a4b54a7a-3f3d-49ae-ba43-5a10c262d4e5] [ 653.073058] env[61923]: DEBUG nova.compute.utils [None req-95b9b70a-a835-4901-9a6d-c0fa1b2217f3 tempest-MigrationsAdminTest-1311988845 tempest-MigrationsAdminTest-1311988845-project-member] [instance: a4b54a7a-3f3d-49ae-ba43-5a10c262d4e5] Binding failed for port 2110c7b3-1dff-4f09-b631-618ae3b8467e, please check neutron logs for more information. {{(pid=61923) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 653.074076] env[61923]: DEBUG oslo_concurrency.lockutils [None req-cb394400-3d4b-4149-a290-413641bb6f18 tempest-InstanceActionsV221TestJSON-1730893270 tempest-InstanceActionsV221TestJSON-1730893270-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.403s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 653.077796] env[61923]: DEBUG nova.compute.manager [None req-95b9b70a-a835-4901-9a6d-c0fa1b2217f3 tempest-MigrationsAdminTest-1311988845 tempest-MigrationsAdminTest-1311988845-project-member] [instance: a4b54a7a-3f3d-49ae-ba43-5a10c262d4e5] Build of instance a4b54a7a-3f3d-49ae-ba43-5a10c262d4e5 was re-scheduled: Binding failed for port 2110c7b3-1dff-4f09-b631-618ae3b8467e, please check neutron logs for more information. {{(pid=61923) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 653.078470] env[61923]: DEBUG nova.compute.manager [None req-95b9b70a-a835-4901-9a6d-c0fa1b2217f3 tempest-MigrationsAdminTest-1311988845 tempest-MigrationsAdminTest-1311988845-project-member] [instance: a4b54a7a-3f3d-49ae-ba43-5a10c262d4e5] Unplugging VIFs for instance {{(pid=61923) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 653.078797] env[61923]: DEBUG oslo_concurrency.lockutils [None req-95b9b70a-a835-4901-9a6d-c0fa1b2217f3 tempest-MigrationsAdminTest-1311988845 tempest-MigrationsAdminTest-1311988845-project-member] Acquiring lock "refresh_cache-a4b54a7a-3f3d-49ae-ba43-5a10c262d4e5" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 653.079017] env[61923]: DEBUG oslo_concurrency.lockutils [None req-95b9b70a-a835-4901-9a6d-c0fa1b2217f3 tempest-MigrationsAdminTest-1311988845 tempest-MigrationsAdminTest-1311988845-project-member] Acquired lock "refresh_cache-a4b54a7a-3f3d-49ae-ba43-5a10c262d4e5" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 653.079231] env[61923]: DEBUG nova.network.neutron [None req-95b9b70a-a835-4901-9a6d-c0fa1b2217f3 tempest-MigrationsAdminTest-1311988845 tempest-MigrationsAdminTest-1311988845-project-member] [instance: a4b54a7a-3f3d-49ae-ba43-5a10c262d4e5] Building network info cache for instance {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 653.110302] env[61923]: DEBUG nova.compute.manager [None req-56950b6c-0b87-48bf-a575-f699fcd1a5b0 tempest-FloatingIPsAssociationTestJSON-1125124803 tempest-FloatingIPsAssociationTestJSON-1125124803-project-member] [instance: aeca6a94-478d-435a-9d93-59936f26654e] Start spawning the instance on the hypervisor. {{(pid=61923) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 653.138288] env[61923]: DEBUG nova.virt.hardware [None req-56950b6c-0b87-48bf-a575-f699fcd1a5b0 tempest-FloatingIPsAssociationTestJSON-1125124803 tempest-FloatingIPsAssociationTestJSON-1125124803-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-29T20:07:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-29T20:07:35Z,direct_url=,disk_format='vmdk',id=0f153f63-ae0a-45b1-b7f5-7f9b673c947f,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='4e7b5e3b3c3443c8bd5c70f8a15739f5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-29T20:07:36Z,virtual_size=,visibility=), allow threads: False {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 653.138549] env[61923]: DEBUG nova.virt.hardware [None req-56950b6c-0b87-48bf-a575-f699fcd1a5b0 tempest-FloatingIPsAssociationTestJSON-1125124803 tempest-FloatingIPsAssociationTestJSON-1125124803-project-member] Flavor limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 653.138706] env[61923]: DEBUG nova.virt.hardware [None req-56950b6c-0b87-48bf-a575-f699fcd1a5b0 tempest-FloatingIPsAssociationTestJSON-1125124803 tempest-FloatingIPsAssociationTestJSON-1125124803-project-member] Image limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 653.138886] env[61923]: DEBUG nova.virt.hardware [None req-56950b6c-0b87-48bf-a575-f699fcd1a5b0 tempest-FloatingIPsAssociationTestJSON-1125124803 tempest-FloatingIPsAssociationTestJSON-1125124803-project-member] Flavor pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 653.139040] env[61923]: DEBUG nova.virt.hardware [None req-56950b6c-0b87-48bf-a575-f699fcd1a5b0 tempest-FloatingIPsAssociationTestJSON-1125124803 tempest-FloatingIPsAssociationTestJSON-1125124803-project-member] Image pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 653.139195] env[61923]: DEBUG nova.virt.hardware [None req-56950b6c-0b87-48bf-a575-f699fcd1a5b0 tempest-FloatingIPsAssociationTestJSON-1125124803 tempest-FloatingIPsAssociationTestJSON-1125124803-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 653.139417] env[61923]: DEBUG nova.virt.hardware [None req-56950b6c-0b87-48bf-a575-f699fcd1a5b0 tempest-FloatingIPsAssociationTestJSON-1125124803 tempest-FloatingIPsAssociationTestJSON-1125124803-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 653.139574] env[61923]: DEBUG nova.virt.hardware [None req-56950b6c-0b87-48bf-a575-f699fcd1a5b0 tempest-FloatingIPsAssociationTestJSON-1125124803 tempest-FloatingIPsAssociationTestJSON-1125124803-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 653.139737] env[61923]: DEBUG nova.virt.hardware [None req-56950b6c-0b87-48bf-a575-f699fcd1a5b0 tempest-FloatingIPsAssociationTestJSON-1125124803 tempest-FloatingIPsAssociationTestJSON-1125124803-project-member] Got 1 possible topologies {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 653.139901] env[61923]: DEBUG nova.virt.hardware [None req-56950b6c-0b87-48bf-a575-f699fcd1a5b0 tempest-FloatingIPsAssociationTestJSON-1125124803 tempest-FloatingIPsAssociationTestJSON-1125124803-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 653.140086] env[61923]: DEBUG nova.virt.hardware [None req-56950b6c-0b87-48bf-a575-f699fcd1a5b0 tempest-FloatingIPsAssociationTestJSON-1125124803 tempest-FloatingIPsAssociationTestJSON-1125124803-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 653.141220] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c6d0601-499f-4eee-ab7e-69a3a45bf153 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 653.150582] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-077ad011-3baa-4aa4-a75a-386d6f99d111 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 653.155417] env[61923]: DEBUG nova.compute.manager [req-323549fd-92bd-4a6a-acc7-727efb9abdc7 req-24705cbb-f7a4-453c-ab3c-d0f1b4252f6e service nova] [instance: aeca6a94-478d-435a-9d93-59936f26654e] Received event network-changed-61e9a5e4-10ed-40b2-a961-2d5cbae80f02 {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 653.155605] env[61923]: DEBUG nova.compute.manager [req-323549fd-92bd-4a6a-acc7-727efb9abdc7 req-24705cbb-f7a4-453c-ab3c-d0f1b4252f6e service nova] [instance: aeca6a94-478d-435a-9d93-59936f26654e] Refreshing instance network info cache due to event network-changed-61e9a5e4-10ed-40b2-a961-2d5cbae80f02. {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 653.155816] env[61923]: DEBUG oslo_concurrency.lockutils [req-323549fd-92bd-4a6a-acc7-727efb9abdc7 req-24705cbb-f7a4-453c-ab3c-d0f1b4252f6e service nova] Acquiring lock "refresh_cache-aeca6a94-478d-435a-9d93-59936f26654e" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 653.155962] env[61923]: DEBUG oslo_concurrency.lockutils [req-323549fd-92bd-4a6a-acc7-727efb9abdc7 req-24705cbb-f7a4-453c-ab3c-d0f1b4252f6e service nova] Acquired lock "refresh_cache-aeca6a94-478d-435a-9d93-59936f26654e" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 653.156131] env[61923]: DEBUG nova.network.neutron [req-323549fd-92bd-4a6a-acc7-727efb9abdc7 req-24705cbb-f7a4-453c-ab3c-d0f1b4252f6e service nova] [instance: aeca6a94-478d-435a-9d93-59936f26654e] Refreshing network info cache for port 61e9a5e4-10ed-40b2-a961-2d5cbae80f02 {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 653.334420] env[61923]: ERROR nova.compute.manager [None req-56950b6c-0b87-48bf-a575-f699fcd1a5b0 tempest-FloatingIPsAssociationTestJSON-1125124803 tempest-FloatingIPsAssociationTestJSON-1125124803-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 61e9a5e4-10ed-40b2-a961-2d5cbae80f02, please check neutron logs for more information. [ 653.334420] env[61923]: ERROR nova.compute.manager Traceback (most recent call last): [ 653.334420] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 653.334420] env[61923]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 653.334420] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 653.334420] env[61923]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 653.334420] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 653.334420] env[61923]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 653.334420] env[61923]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 653.334420] env[61923]: ERROR nova.compute.manager self.force_reraise() [ 653.334420] env[61923]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 653.334420] env[61923]: ERROR nova.compute.manager raise self.value [ 653.334420] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 653.334420] env[61923]: ERROR nova.compute.manager updated_port = self._update_port( [ 653.334420] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 653.334420] env[61923]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 653.334936] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 653.334936] env[61923]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 653.334936] env[61923]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 61e9a5e4-10ed-40b2-a961-2d5cbae80f02, please check neutron logs for more information. [ 653.334936] env[61923]: ERROR nova.compute.manager [ 653.334936] env[61923]: Traceback (most recent call last): [ 653.334936] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 653.334936] env[61923]: listener.cb(fileno) [ 653.334936] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 653.334936] env[61923]: result = function(*args, **kwargs) [ 653.334936] env[61923]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 653.334936] env[61923]: return func(*args, **kwargs) [ 653.334936] env[61923]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 653.334936] env[61923]: raise e [ 653.334936] env[61923]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 653.334936] env[61923]: nwinfo = self.network_api.allocate_for_instance( [ 653.334936] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 653.334936] env[61923]: created_port_ids = self._update_ports_for_instance( [ 653.334936] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 653.334936] env[61923]: with excutils.save_and_reraise_exception(): [ 653.334936] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 653.334936] env[61923]: self.force_reraise() [ 653.334936] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 653.334936] env[61923]: raise self.value [ 653.334936] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 653.334936] env[61923]: updated_port = self._update_port( [ 653.334936] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 653.334936] env[61923]: _ensure_no_port_binding_failure(port) [ 653.334936] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 653.334936] env[61923]: raise exception.PortBindingFailed(port_id=port['id']) [ 653.335735] env[61923]: nova.exception.PortBindingFailed: Binding failed for port 61e9a5e4-10ed-40b2-a961-2d5cbae80f02, please check neutron logs for more information. [ 653.335735] env[61923]: Removing descriptor: 17 [ 653.335735] env[61923]: ERROR nova.compute.manager [None req-56950b6c-0b87-48bf-a575-f699fcd1a5b0 tempest-FloatingIPsAssociationTestJSON-1125124803 tempest-FloatingIPsAssociationTestJSON-1125124803-project-member] [instance: aeca6a94-478d-435a-9d93-59936f26654e] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 61e9a5e4-10ed-40b2-a961-2d5cbae80f02, please check neutron logs for more information. [ 653.335735] env[61923]: ERROR nova.compute.manager [instance: aeca6a94-478d-435a-9d93-59936f26654e] Traceback (most recent call last): [ 653.335735] env[61923]: ERROR nova.compute.manager [instance: aeca6a94-478d-435a-9d93-59936f26654e] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 653.335735] env[61923]: ERROR nova.compute.manager [instance: aeca6a94-478d-435a-9d93-59936f26654e] yield resources [ 653.335735] env[61923]: ERROR nova.compute.manager [instance: aeca6a94-478d-435a-9d93-59936f26654e] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 653.335735] env[61923]: ERROR nova.compute.manager [instance: aeca6a94-478d-435a-9d93-59936f26654e] self.driver.spawn(context, instance, image_meta, [ 653.335735] env[61923]: ERROR nova.compute.manager [instance: aeca6a94-478d-435a-9d93-59936f26654e] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 653.335735] env[61923]: ERROR nova.compute.manager [instance: aeca6a94-478d-435a-9d93-59936f26654e] self._vmops.spawn(context, instance, image_meta, injected_files, [ 653.335735] env[61923]: ERROR nova.compute.manager [instance: aeca6a94-478d-435a-9d93-59936f26654e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 653.335735] env[61923]: ERROR nova.compute.manager [instance: aeca6a94-478d-435a-9d93-59936f26654e] vm_ref = self.build_virtual_machine(instance, [ 653.336074] env[61923]: ERROR nova.compute.manager [instance: aeca6a94-478d-435a-9d93-59936f26654e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 653.336074] env[61923]: ERROR nova.compute.manager [instance: aeca6a94-478d-435a-9d93-59936f26654e] vif_infos = vmwarevif.get_vif_info(self._session, [ 653.336074] env[61923]: ERROR nova.compute.manager [instance: aeca6a94-478d-435a-9d93-59936f26654e] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 653.336074] env[61923]: ERROR nova.compute.manager [instance: aeca6a94-478d-435a-9d93-59936f26654e] for vif in network_info: [ 653.336074] env[61923]: ERROR nova.compute.manager [instance: aeca6a94-478d-435a-9d93-59936f26654e] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 653.336074] env[61923]: ERROR nova.compute.manager [instance: aeca6a94-478d-435a-9d93-59936f26654e] return self._sync_wrapper(fn, *args, **kwargs) [ 653.336074] env[61923]: ERROR nova.compute.manager [instance: aeca6a94-478d-435a-9d93-59936f26654e] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 653.336074] env[61923]: ERROR nova.compute.manager [instance: aeca6a94-478d-435a-9d93-59936f26654e] self.wait() [ 653.336074] env[61923]: ERROR nova.compute.manager [instance: aeca6a94-478d-435a-9d93-59936f26654e] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 653.336074] env[61923]: ERROR nova.compute.manager [instance: aeca6a94-478d-435a-9d93-59936f26654e] self[:] = self._gt.wait() [ 653.336074] env[61923]: ERROR nova.compute.manager [instance: aeca6a94-478d-435a-9d93-59936f26654e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 653.336074] env[61923]: ERROR nova.compute.manager [instance: aeca6a94-478d-435a-9d93-59936f26654e] return self._exit_event.wait() [ 653.336074] env[61923]: ERROR nova.compute.manager [instance: aeca6a94-478d-435a-9d93-59936f26654e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 653.336446] env[61923]: ERROR nova.compute.manager [instance: aeca6a94-478d-435a-9d93-59936f26654e] result = hub.switch() [ 653.336446] env[61923]: ERROR nova.compute.manager [instance: aeca6a94-478d-435a-9d93-59936f26654e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 653.336446] env[61923]: ERROR nova.compute.manager [instance: aeca6a94-478d-435a-9d93-59936f26654e] return self.greenlet.switch() [ 653.336446] env[61923]: ERROR nova.compute.manager [instance: aeca6a94-478d-435a-9d93-59936f26654e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 653.336446] env[61923]: ERROR nova.compute.manager [instance: aeca6a94-478d-435a-9d93-59936f26654e] result = function(*args, **kwargs) [ 653.336446] env[61923]: ERROR nova.compute.manager [instance: aeca6a94-478d-435a-9d93-59936f26654e] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 653.336446] env[61923]: ERROR nova.compute.manager [instance: aeca6a94-478d-435a-9d93-59936f26654e] return func(*args, **kwargs) [ 653.336446] env[61923]: ERROR nova.compute.manager [instance: aeca6a94-478d-435a-9d93-59936f26654e] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 653.336446] env[61923]: ERROR nova.compute.manager [instance: aeca6a94-478d-435a-9d93-59936f26654e] raise e [ 653.336446] env[61923]: ERROR nova.compute.manager [instance: aeca6a94-478d-435a-9d93-59936f26654e] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 653.336446] env[61923]: ERROR nova.compute.manager [instance: aeca6a94-478d-435a-9d93-59936f26654e] nwinfo = self.network_api.allocate_for_instance( [ 653.336446] env[61923]: ERROR nova.compute.manager [instance: aeca6a94-478d-435a-9d93-59936f26654e] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 653.336446] env[61923]: ERROR nova.compute.manager [instance: aeca6a94-478d-435a-9d93-59936f26654e] created_port_ids = self._update_ports_for_instance( [ 653.336802] env[61923]: ERROR nova.compute.manager [instance: aeca6a94-478d-435a-9d93-59936f26654e] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 653.336802] env[61923]: ERROR nova.compute.manager [instance: aeca6a94-478d-435a-9d93-59936f26654e] with excutils.save_and_reraise_exception(): [ 653.336802] env[61923]: ERROR nova.compute.manager [instance: aeca6a94-478d-435a-9d93-59936f26654e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 653.336802] env[61923]: ERROR nova.compute.manager [instance: aeca6a94-478d-435a-9d93-59936f26654e] self.force_reraise() [ 653.336802] env[61923]: ERROR nova.compute.manager [instance: aeca6a94-478d-435a-9d93-59936f26654e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 653.336802] env[61923]: ERROR nova.compute.manager [instance: aeca6a94-478d-435a-9d93-59936f26654e] raise self.value [ 653.336802] env[61923]: ERROR nova.compute.manager [instance: aeca6a94-478d-435a-9d93-59936f26654e] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 653.336802] env[61923]: ERROR nova.compute.manager [instance: aeca6a94-478d-435a-9d93-59936f26654e] updated_port = self._update_port( [ 653.336802] env[61923]: ERROR nova.compute.manager [instance: aeca6a94-478d-435a-9d93-59936f26654e] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 653.336802] env[61923]: ERROR nova.compute.manager [instance: aeca6a94-478d-435a-9d93-59936f26654e] _ensure_no_port_binding_failure(port) [ 653.336802] env[61923]: ERROR nova.compute.manager [instance: aeca6a94-478d-435a-9d93-59936f26654e] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 653.336802] env[61923]: ERROR nova.compute.manager [instance: aeca6a94-478d-435a-9d93-59936f26654e] raise exception.PortBindingFailed(port_id=port['id']) [ 653.337186] env[61923]: ERROR nova.compute.manager [instance: aeca6a94-478d-435a-9d93-59936f26654e] nova.exception.PortBindingFailed: Binding failed for port 61e9a5e4-10ed-40b2-a961-2d5cbae80f02, please check neutron logs for more information. [ 653.337186] env[61923]: ERROR nova.compute.manager [instance: aeca6a94-478d-435a-9d93-59936f26654e] [ 653.337186] env[61923]: INFO nova.compute.manager [None req-56950b6c-0b87-48bf-a575-f699fcd1a5b0 tempest-FloatingIPsAssociationTestJSON-1125124803 tempest-FloatingIPsAssociationTestJSON-1125124803-project-member] [instance: aeca6a94-478d-435a-9d93-59936f26654e] Terminating instance [ 653.340037] env[61923]: DEBUG oslo_concurrency.lockutils [None req-56950b6c-0b87-48bf-a575-f699fcd1a5b0 tempest-FloatingIPsAssociationTestJSON-1125124803 tempest-FloatingIPsAssociationTestJSON-1125124803-project-member] Acquiring lock "refresh_cache-aeca6a94-478d-435a-9d93-59936f26654e" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 653.438268] env[61923]: DEBUG oslo_concurrency.lockutils [None req-aadda1bb-f770-4742-8c34-cd25e3f7e7cb tempest-InstanceActionsNegativeTestJSON-1700332362 tempest-InstanceActionsNegativeTestJSON-1700332362-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 653.669379] env[61923]: DEBUG nova.network.neutron [None req-95b9b70a-a835-4901-9a6d-c0fa1b2217f3 tempest-MigrationsAdminTest-1311988845 tempest-MigrationsAdminTest-1311988845-project-member] [instance: a4b54a7a-3f3d-49ae-ba43-5a10c262d4e5] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 653.694969] env[61923]: DEBUG nova.network.neutron [req-323549fd-92bd-4a6a-acc7-727efb9abdc7 req-24705cbb-f7a4-453c-ab3c-d0f1b4252f6e service nova] [instance: aeca6a94-478d-435a-9d93-59936f26654e] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 653.838861] env[61923]: DEBUG nova.network.neutron [None req-95b9b70a-a835-4901-9a6d-c0fa1b2217f3 tempest-MigrationsAdminTest-1311988845 tempest-MigrationsAdminTest-1311988845-project-member] [instance: a4b54a7a-3f3d-49ae-ba43-5a10c262d4e5] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 653.909666] env[61923]: DEBUG nova.network.neutron [req-323549fd-92bd-4a6a-acc7-727efb9abdc7 req-24705cbb-f7a4-453c-ab3c-d0f1b4252f6e service nova] [instance: aeca6a94-478d-435a-9d93-59936f26654e] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 654.035935] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f99b5217-4850-4456-bccd-2257fe22cd73 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 654.044082] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ab2133d-d452-482d-a218-ea4f8292166f {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 654.082652] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2721b3b-66aa-4b4c-927f-c6c2ed88dede {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 654.090769] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b210f588-a2c3-4f5d-8867-b39355370ed4 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 654.109345] env[61923]: DEBUG nova.compute.provider_tree [None req-cb394400-3d4b-4149-a290-413641bb6f18 tempest-InstanceActionsV221TestJSON-1730893270 tempest-InstanceActionsV221TestJSON-1730893270-project-member] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 654.338787] env[61923]: DEBUG oslo_concurrency.lockutils [None req-95b9b70a-a835-4901-9a6d-c0fa1b2217f3 tempest-MigrationsAdminTest-1311988845 tempest-MigrationsAdminTest-1311988845-project-member] Releasing lock "refresh_cache-a4b54a7a-3f3d-49ae-ba43-5a10c262d4e5" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 654.339046] env[61923]: DEBUG nova.compute.manager [None req-95b9b70a-a835-4901-9a6d-c0fa1b2217f3 tempest-MigrationsAdminTest-1311988845 tempest-MigrationsAdminTest-1311988845-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61923) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 654.339225] env[61923]: DEBUG nova.compute.manager [None req-95b9b70a-a835-4901-9a6d-c0fa1b2217f3 tempest-MigrationsAdminTest-1311988845 tempest-MigrationsAdminTest-1311988845-project-member] [instance: a4b54a7a-3f3d-49ae-ba43-5a10c262d4e5] Deallocating network for instance {{(pid=61923) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 654.339393] env[61923]: DEBUG nova.network.neutron [None req-95b9b70a-a835-4901-9a6d-c0fa1b2217f3 tempest-MigrationsAdminTest-1311988845 tempest-MigrationsAdminTest-1311988845-project-member] [instance: a4b54a7a-3f3d-49ae-ba43-5a10c262d4e5] deallocate_for_instance() {{(pid=61923) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 654.359213] env[61923]: DEBUG nova.network.neutron [None req-95b9b70a-a835-4901-9a6d-c0fa1b2217f3 tempest-MigrationsAdminTest-1311988845 tempest-MigrationsAdminTest-1311988845-project-member] [instance: a4b54a7a-3f3d-49ae-ba43-5a10c262d4e5] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 654.412815] env[61923]: DEBUG oslo_concurrency.lockutils [req-323549fd-92bd-4a6a-acc7-727efb9abdc7 req-24705cbb-f7a4-453c-ab3c-d0f1b4252f6e service nova] Releasing lock "refresh_cache-aeca6a94-478d-435a-9d93-59936f26654e" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 654.414681] env[61923]: DEBUG oslo_concurrency.lockutils [None req-56950b6c-0b87-48bf-a575-f699fcd1a5b0 tempest-FloatingIPsAssociationTestJSON-1125124803 tempest-FloatingIPsAssociationTestJSON-1125124803-project-member] Acquired lock "refresh_cache-aeca6a94-478d-435a-9d93-59936f26654e" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 654.414890] env[61923]: DEBUG nova.network.neutron [None req-56950b6c-0b87-48bf-a575-f699fcd1a5b0 tempest-FloatingIPsAssociationTestJSON-1125124803 tempest-FloatingIPsAssociationTestJSON-1125124803-project-member] [instance: aeca6a94-478d-435a-9d93-59936f26654e] Building network info cache for instance {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 654.613876] env[61923]: DEBUG nova.scheduler.client.report [None req-cb394400-3d4b-4149-a290-413641bb6f18 tempest-InstanceActionsV221TestJSON-1730893270 tempest-InstanceActionsV221TestJSON-1730893270-project-member] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 654.864626] env[61923]: DEBUG nova.network.neutron [None req-95b9b70a-a835-4901-9a6d-c0fa1b2217f3 tempest-MigrationsAdminTest-1311988845 tempest-MigrationsAdminTest-1311988845-project-member] [instance: a4b54a7a-3f3d-49ae-ba43-5a10c262d4e5] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 654.957681] env[61923]: DEBUG nova.network.neutron [None req-56950b6c-0b87-48bf-a575-f699fcd1a5b0 tempest-FloatingIPsAssociationTestJSON-1125124803 tempest-FloatingIPsAssociationTestJSON-1125124803-project-member] [instance: aeca6a94-478d-435a-9d93-59936f26654e] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 655.122026] env[61923]: DEBUG nova.network.neutron [None req-56950b6c-0b87-48bf-a575-f699fcd1a5b0 tempest-FloatingIPsAssociationTestJSON-1125124803 tempest-FloatingIPsAssociationTestJSON-1125124803-project-member] [instance: aeca6a94-478d-435a-9d93-59936f26654e] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 655.126186] env[61923]: DEBUG oslo_concurrency.lockutils [None req-cb394400-3d4b-4149-a290-413641bb6f18 tempest-InstanceActionsV221TestJSON-1730893270 tempest-InstanceActionsV221TestJSON-1730893270-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.050s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 655.126186] env[61923]: ERROR nova.compute.manager [None req-cb394400-3d4b-4149-a290-413641bb6f18 tempest-InstanceActionsV221TestJSON-1730893270 tempest-InstanceActionsV221TestJSON-1730893270-project-member] [instance: dc8cc22d-aba9-4dba-b456-1b323013d234] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 07f299cc-2ca7-4beb-9f20-7509c49ed4f0, please check neutron logs for more information. [ 655.126186] env[61923]: ERROR nova.compute.manager [instance: dc8cc22d-aba9-4dba-b456-1b323013d234] Traceback (most recent call last): [ 655.126186] env[61923]: ERROR nova.compute.manager [instance: dc8cc22d-aba9-4dba-b456-1b323013d234] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 655.126186] env[61923]: ERROR nova.compute.manager [instance: dc8cc22d-aba9-4dba-b456-1b323013d234] self.driver.spawn(context, instance, image_meta, [ 655.126186] env[61923]: ERROR nova.compute.manager [instance: dc8cc22d-aba9-4dba-b456-1b323013d234] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 655.126186] env[61923]: ERROR nova.compute.manager [instance: dc8cc22d-aba9-4dba-b456-1b323013d234] self._vmops.spawn(context, instance, image_meta, injected_files, [ 655.126186] env[61923]: ERROR nova.compute.manager [instance: dc8cc22d-aba9-4dba-b456-1b323013d234] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 655.126186] env[61923]: ERROR nova.compute.manager [instance: dc8cc22d-aba9-4dba-b456-1b323013d234] vm_ref = self.build_virtual_machine(instance, [ 655.126468] env[61923]: ERROR nova.compute.manager [instance: dc8cc22d-aba9-4dba-b456-1b323013d234] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 655.126468] env[61923]: ERROR nova.compute.manager [instance: dc8cc22d-aba9-4dba-b456-1b323013d234] vif_infos = vmwarevif.get_vif_info(self._session, [ 655.126468] env[61923]: ERROR nova.compute.manager [instance: dc8cc22d-aba9-4dba-b456-1b323013d234] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 655.126468] env[61923]: ERROR nova.compute.manager [instance: dc8cc22d-aba9-4dba-b456-1b323013d234] for vif in network_info: [ 655.126468] env[61923]: ERROR nova.compute.manager [instance: dc8cc22d-aba9-4dba-b456-1b323013d234] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 655.126468] env[61923]: ERROR nova.compute.manager [instance: dc8cc22d-aba9-4dba-b456-1b323013d234] return self._sync_wrapper(fn, *args, **kwargs) [ 655.126468] env[61923]: ERROR nova.compute.manager [instance: dc8cc22d-aba9-4dba-b456-1b323013d234] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 655.126468] env[61923]: ERROR nova.compute.manager [instance: dc8cc22d-aba9-4dba-b456-1b323013d234] self.wait() [ 655.126468] env[61923]: ERROR nova.compute.manager [instance: dc8cc22d-aba9-4dba-b456-1b323013d234] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 655.126468] env[61923]: ERROR nova.compute.manager [instance: dc8cc22d-aba9-4dba-b456-1b323013d234] self[:] = self._gt.wait() [ 655.126468] env[61923]: ERROR nova.compute.manager [instance: dc8cc22d-aba9-4dba-b456-1b323013d234] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 655.126468] env[61923]: ERROR nova.compute.manager [instance: dc8cc22d-aba9-4dba-b456-1b323013d234] return self._exit_event.wait() [ 655.126468] env[61923]: ERROR nova.compute.manager [instance: dc8cc22d-aba9-4dba-b456-1b323013d234] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 655.126785] env[61923]: ERROR nova.compute.manager [instance: dc8cc22d-aba9-4dba-b456-1b323013d234] result = hub.switch() [ 655.126785] env[61923]: ERROR nova.compute.manager [instance: dc8cc22d-aba9-4dba-b456-1b323013d234] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 655.126785] env[61923]: ERROR nova.compute.manager [instance: dc8cc22d-aba9-4dba-b456-1b323013d234] return self.greenlet.switch() [ 655.126785] env[61923]: ERROR nova.compute.manager [instance: dc8cc22d-aba9-4dba-b456-1b323013d234] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 655.126785] env[61923]: ERROR nova.compute.manager [instance: dc8cc22d-aba9-4dba-b456-1b323013d234] result = function(*args, **kwargs) [ 655.126785] env[61923]: ERROR nova.compute.manager [instance: dc8cc22d-aba9-4dba-b456-1b323013d234] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 655.126785] env[61923]: ERROR nova.compute.manager [instance: dc8cc22d-aba9-4dba-b456-1b323013d234] return func(*args, **kwargs) [ 655.126785] env[61923]: ERROR nova.compute.manager [instance: dc8cc22d-aba9-4dba-b456-1b323013d234] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 655.126785] env[61923]: ERROR nova.compute.manager [instance: dc8cc22d-aba9-4dba-b456-1b323013d234] raise e [ 655.126785] env[61923]: ERROR nova.compute.manager [instance: dc8cc22d-aba9-4dba-b456-1b323013d234] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 655.126785] env[61923]: ERROR nova.compute.manager [instance: dc8cc22d-aba9-4dba-b456-1b323013d234] nwinfo = self.network_api.allocate_for_instance( [ 655.126785] env[61923]: ERROR nova.compute.manager [instance: dc8cc22d-aba9-4dba-b456-1b323013d234] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 655.126785] env[61923]: ERROR nova.compute.manager [instance: dc8cc22d-aba9-4dba-b456-1b323013d234] created_port_ids = self._update_ports_for_instance( [ 655.128114] env[61923]: ERROR nova.compute.manager [instance: dc8cc22d-aba9-4dba-b456-1b323013d234] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 655.128114] env[61923]: ERROR nova.compute.manager [instance: dc8cc22d-aba9-4dba-b456-1b323013d234] with excutils.save_and_reraise_exception(): [ 655.128114] env[61923]: ERROR nova.compute.manager [instance: dc8cc22d-aba9-4dba-b456-1b323013d234] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 655.128114] env[61923]: ERROR nova.compute.manager [instance: dc8cc22d-aba9-4dba-b456-1b323013d234] self.force_reraise() [ 655.128114] env[61923]: ERROR nova.compute.manager [instance: dc8cc22d-aba9-4dba-b456-1b323013d234] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 655.128114] env[61923]: ERROR nova.compute.manager [instance: dc8cc22d-aba9-4dba-b456-1b323013d234] raise self.value [ 655.128114] env[61923]: ERROR nova.compute.manager [instance: dc8cc22d-aba9-4dba-b456-1b323013d234] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 655.128114] env[61923]: ERROR nova.compute.manager [instance: dc8cc22d-aba9-4dba-b456-1b323013d234] updated_port = self._update_port( [ 655.128114] env[61923]: ERROR nova.compute.manager [instance: dc8cc22d-aba9-4dba-b456-1b323013d234] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 655.128114] env[61923]: ERROR nova.compute.manager [instance: dc8cc22d-aba9-4dba-b456-1b323013d234] _ensure_no_port_binding_failure(port) [ 655.128114] env[61923]: ERROR nova.compute.manager [instance: dc8cc22d-aba9-4dba-b456-1b323013d234] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 655.128114] env[61923]: ERROR nova.compute.manager [instance: dc8cc22d-aba9-4dba-b456-1b323013d234] raise exception.PortBindingFailed(port_id=port['id']) [ 655.128504] env[61923]: ERROR nova.compute.manager [instance: dc8cc22d-aba9-4dba-b456-1b323013d234] nova.exception.PortBindingFailed: Binding failed for port 07f299cc-2ca7-4beb-9f20-7509c49ed4f0, please check neutron logs for more information. [ 655.128504] env[61923]: ERROR nova.compute.manager [instance: dc8cc22d-aba9-4dba-b456-1b323013d234] [ 655.128504] env[61923]: DEBUG nova.compute.utils [None req-cb394400-3d4b-4149-a290-413641bb6f18 tempest-InstanceActionsV221TestJSON-1730893270 tempest-InstanceActionsV221TestJSON-1730893270-project-member] [instance: dc8cc22d-aba9-4dba-b456-1b323013d234] Binding failed for port 07f299cc-2ca7-4beb-9f20-7509c49ed4f0, please check neutron logs for more information. {{(pid=61923) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 655.128504] env[61923]: DEBUG oslo_concurrency.lockutils [None req-753531fa-39a1-4c82-9915-d195bf0fe1b2 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.838s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 655.132640] env[61923]: DEBUG nova.compute.manager [None req-cb394400-3d4b-4149-a290-413641bb6f18 tempest-InstanceActionsV221TestJSON-1730893270 tempest-InstanceActionsV221TestJSON-1730893270-project-member] [instance: dc8cc22d-aba9-4dba-b456-1b323013d234] Build of instance dc8cc22d-aba9-4dba-b456-1b323013d234 was re-scheduled: Binding failed for port 07f299cc-2ca7-4beb-9f20-7509c49ed4f0, please check neutron logs for more information. {{(pid=61923) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 655.133233] env[61923]: DEBUG nova.compute.manager [None req-cb394400-3d4b-4149-a290-413641bb6f18 tempest-InstanceActionsV221TestJSON-1730893270 tempest-InstanceActionsV221TestJSON-1730893270-project-member] [instance: dc8cc22d-aba9-4dba-b456-1b323013d234] Unplugging VIFs for instance {{(pid=61923) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 655.133499] env[61923]: DEBUG oslo_concurrency.lockutils [None req-cb394400-3d4b-4149-a290-413641bb6f18 tempest-InstanceActionsV221TestJSON-1730893270 tempest-InstanceActionsV221TestJSON-1730893270-project-member] Acquiring lock "refresh_cache-dc8cc22d-aba9-4dba-b456-1b323013d234" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 655.133861] env[61923]: DEBUG oslo_concurrency.lockutils [None req-cb394400-3d4b-4149-a290-413641bb6f18 tempest-InstanceActionsV221TestJSON-1730893270 tempest-InstanceActionsV221TestJSON-1730893270-project-member] Acquired lock "refresh_cache-dc8cc22d-aba9-4dba-b456-1b323013d234" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 655.134109] env[61923]: DEBUG nova.network.neutron [None req-cb394400-3d4b-4149-a290-413641bb6f18 tempest-InstanceActionsV221TestJSON-1730893270 tempest-InstanceActionsV221TestJSON-1730893270-project-member] [instance: dc8cc22d-aba9-4dba-b456-1b323013d234] Building network info cache for instance {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 655.350893] env[61923]: DEBUG nova.compute.manager [req-89794378-c8a6-48ea-bc3f-8318683f45e7 req-4866d7ff-5189-4434-9bda-0c18eacdce68 service nova] [instance: aeca6a94-478d-435a-9d93-59936f26654e] Received event network-vif-deleted-61e9a5e4-10ed-40b2-a961-2d5cbae80f02 {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 655.367613] env[61923]: INFO nova.compute.manager [None req-95b9b70a-a835-4901-9a6d-c0fa1b2217f3 tempest-MigrationsAdminTest-1311988845 tempest-MigrationsAdminTest-1311988845-project-member] [instance: a4b54a7a-3f3d-49ae-ba43-5a10c262d4e5] Took 1.03 seconds to deallocate network for instance. [ 655.625408] env[61923]: DEBUG oslo_concurrency.lockutils [None req-56950b6c-0b87-48bf-a575-f699fcd1a5b0 tempest-FloatingIPsAssociationTestJSON-1125124803 tempest-FloatingIPsAssociationTestJSON-1125124803-project-member] Releasing lock "refresh_cache-aeca6a94-478d-435a-9d93-59936f26654e" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 655.625408] env[61923]: DEBUG nova.compute.manager [None req-56950b6c-0b87-48bf-a575-f699fcd1a5b0 tempest-FloatingIPsAssociationTestJSON-1125124803 tempest-FloatingIPsAssociationTestJSON-1125124803-project-member] [instance: aeca6a94-478d-435a-9d93-59936f26654e] Start destroying the instance on the hypervisor. {{(pid=61923) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 655.625408] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-56950b6c-0b87-48bf-a575-f699fcd1a5b0 tempest-FloatingIPsAssociationTestJSON-1125124803 tempest-FloatingIPsAssociationTestJSON-1125124803-project-member] [instance: aeca6a94-478d-435a-9d93-59936f26654e] Destroying instance {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 655.625723] env[61923]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-929be888-2598-4dc8-bafa-33c80c01dcdd {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 655.637748] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-208f3619-efd1-4f03-9ae6-d3d6ba05b4ee {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 655.665871] env[61923]: WARNING nova.virt.vmwareapi.vmops [None req-56950b6c-0b87-48bf-a575-f699fcd1a5b0 tempest-FloatingIPsAssociationTestJSON-1125124803 tempest-FloatingIPsAssociationTestJSON-1125124803-project-member] [instance: aeca6a94-478d-435a-9d93-59936f26654e] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance aeca6a94-478d-435a-9d93-59936f26654e could not be found. [ 655.666125] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-56950b6c-0b87-48bf-a575-f699fcd1a5b0 tempest-FloatingIPsAssociationTestJSON-1125124803 tempest-FloatingIPsAssociationTestJSON-1125124803-project-member] [instance: aeca6a94-478d-435a-9d93-59936f26654e] Instance destroyed {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 655.667023] env[61923]: INFO nova.compute.manager [None req-56950b6c-0b87-48bf-a575-f699fcd1a5b0 tempest-FloatingIPsAssociationTestJSON-1125124803 tempest-FloatingIPsAssociationTestJSON-1125124803-project-member] [instance: aeca6a94-478d-435a-9d93-59936f26654e] Took 0.04 seconds to destroy the instance on the hypervisor. [ 655.667023] env[61923]: DEBUG oslo.service.loopingcall [None req-56950b6c-0b87-48bf-a575-f699fcd1a5b0 tempest-FloatingIPsAssociationTestJSON-1125124803 tempest-FloatingIPsAssociationTestJSON-1125124803-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61923) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 655.667023] env[61923]: DEBUG nova.compute.manager [-] [instance: aeca6a94-478d-435a-9d93-59936f26654e] Deallocating network for instance {{(pid=61923) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 655.667023] env[61923]: DEBUG nova.network.neutron [-] [instance: aeca6a94-478d-435a-9d93-59936f26654e] deallocate_for_instance() {{(pid=61923) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 655.677270] env[61923]: DEBUG nova.network.neutron [None req-cb394400-3d4b-4149-a290-413641bb6f18 tempest-InstanceActionsV221TestJSON-1730893270 tempest-InstanceActionsV221TestJSON-1730893270-project-member] [instance: dc8cc22d-aba9-4dba-b456-1b323013d234] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 655.703035] env[61923]: DEBUG nova.network.neutron [-] [instance: aeca6a94-478d-435a-9d93-59936f26654e] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 655.853822] env[61923]: DEBUG nova.network.neutron [None req-cb394400-3d4b-4149-a290-413641bb6f18 tempest-InstanceActionsV221TestJSON-1730893270 tempest-InstanceActionsV221TestJSON-1730893270-project-member] [instance: dc8cc22d-aba9-4dba-b456-1b323013d234] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 656.013806] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e365b5e3-f8fd-49bc-8c28-b8186bca8073 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 656.022277] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f232b0ab-6d87-496d-a6c7-541e1f812a93 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 656.053944] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6631bfda-82a1-46b6-af78-99eca60baa19 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 656.059365] env[61923]: DEBUG oslo_concurrency.lockutils [None req-ab7b3da0-3a01-44d4-9dfd-3e5d2ad27432 tempest-TenantUsagesTestJSON-871253541 tempest-TenantUsagesTestJSON-871253541-project-member] Acquiring lock "b6712878-3270-43f4-9518-d41eb77124b7" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 656.059480] env[61923]: DEBUG oslo_concurrency.lockutils [None req-ab7b3da0-3a01-44d4-9dfd-3e5d2ad27432 tempest-TenantUsagesTestJSON-871253541 tempest-TenantUsagesTestJSON-871253541-project-member] Lock "b6712878-3270-43f4-9518-d41eb77124b7" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 656.064547] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7089a5fa-7c85-46de-bd76-e58cc3ecfcaa {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 656.078558] env[61923]: DEBUG nova.compute.provider_tree [None req-753531fa-39a1-4c82-9915-d195bf0fe1b2 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 656.206082] env[61923]: DEBUG nova.network.neutron [-] [instance: aeca6a94-478d-435a-9d93-59936f26654e] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 656.355862] env[61923]: DEBUG oslo_concurrency.lockutils [None req-cb394400-3d4b-4149-a290-413641bb6f18 tempest-InstanceActionsV221TestJSON-1730893270 tempest-InstanceActionsV221TestJSON-1730893270-project-member] Releasing lock "refresh_cache-dc8cc22d-aba9-4dba-b456-1b323013d234" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 656.356320] env[61923]: DEBUG nova.compute.manager [None req-cb394400-3d4b-4149-a290-413641bb6f18 tempest-InstanceActionsV221TestJSON-1730893270 tempest-InstanceActionsV221TestJSON-1730893270-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61923) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 656.356532] env[61923]: DEBUG nova.compute.manager [None req-cb394400-3d4b-4149-a290-413641bb6f18 tempest-InstanceActionsV221TestJSON-1730893270 tempest-InstanceActionsV221TestJSON-1730893270-project-member] [instance: dc8cc22d-aba9-4dba-b456-1b323013d234] Deallocating network for instance {{(pid=61923) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 656.356704] env[61923]: DEBUG nova.network.neutron [None req-cb394400-3d4b-4149-a290-413641bb6f18 tempest-InstanceActionsV221TestJSON-1730893270 tempest-InstanceActionsV221TestJSON-1730893270-project-member] [instance: dc8cc22d-aba9-4dba-b456-1b323013d234] deallocate_for_instance() {{(pid=61923) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 656.383438] env[61923]: DEBUG nova.network.neutron [None req-cb394400-3d4b-4149-a290-413641bb6f18 tempest-InstanceActionsV221TestJSON-1730893270 tempest-InstanceActionsV221TestJSON-1730893270-project-member] [instance: dc8cc22d-aba9-4dba-b456-1b323013d234] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 656.409568] env[61923]: INFO nova.scheduler.client.report [None req-95b9b70a-a835-4901-9a6d-c0fa1b2217f3 tempest-MigrationsAdminTest-1311988845 tempest-MigrationsAdminTest-1311988845-project-member] Deleted allocations for instance a4b54a7a-3f3d-49ae-ba43-5a10c262d4e5 [ 656.581943] env[61923]: DEBUG nova.scheduler.client.report [None req-753531fa-39a1-4c82-9915-d195bf0fe1b2 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 656.709201] env[61923]: INFO nova.compute.manager [-] [instance: aeca6a94-478d-435a-9d93-59936f26654e] Took 1.04 seconds to deallocate network for instance. [ 656.711586] env[61923]: DEBUG nova.compute.claims [None req-56950b6c-0b87-48bf-a575-f699fcd1a5b0 tempest-FloatingIPsAssociationTestJSON-1125124803 tempest-FloatingIPsAssociationTestJSON-1125124803-project-member] [instance: aeca6a94-478d-435a-9d93-59936f26654e] Aborting claim: {{(pid=61923) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 656.711743] env[61923]: DEBUG oslo_concurrency.lockutils [None req-56950b6c-0b87-48bf-a575-f699fcd1a5b0 tempest-FloatingIPsAssociationTestJSON-1125124803 tempest-FloatingIPsAssociationTestJSON-1125124803-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 656.887887] env[61923]: DEBUG nova.network.neutron [None req-cb394400-3d4b-4149-a290-413641bb6f18 tempest-InstanceActionsV221TestJSON-1730893270 tempest-InstanceActionsV221TestJSON-1730893270-project-member] [instance: dc8cc22d-aba9-4dba-b456-1b323013d234] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 656.917635] env[61923]: DEBUG oslo_concurrency.lockutils [None req-95b9b70a-a835-4901-9a6d-c0fa1b2217f3 tempest-MigrationsAdminTest-1311988845 tempest-MigrationsAdminTest-1311988845-project-member] Lock "a4b54a7a-3f3d-49ae-ba43-5a10c262d4e5" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 110.045s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 657.086797] env[61923]: DEBUG oslo_concurrency.lockutils [None req-753531fa-39a1-4c82-9915-d195bf0fe1b2 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.960s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 657.087482] env[61923]: ERROR nova.compute.manager [None req-753531fa-39a1-4c82-9915-d195bf0fe1b2 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] [instance: 03100310-86ce-4739-8636-cd2aa1b542ff] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 5a9feebd-c914-422f-af20-45e4709470e8, please check neutron logs for more information. [ 657.087482] env[61923]: ERROR nova.compute.manager [instance: 03100310-86ce-4739-8636-cd2aa1b542ff] Traceback (most recent call last): [ 657.087482] env[61923]: ERROR nova.compute.manager [instance: 03100310-86ce-4739-8636-cd2aa1b542ff] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 657.087482] env[61923]: ERROR nova.compute.manager [instance: 03100310-86ce-4739-8636-cd2aa1b542ff] self.driver.spawn(context, instance, image_meta, [ 657.087482] env[61923]: ERROR nova.compute.manager [instance: 03100310-86ce-4739-8636-cd2aa1b542ff] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 657.087482] env[61923]: ERROR nova.compute.manager [instance: 03100310-86ce-4739-8636-cd2aa1b542ff] self._vmops.spawn(context, instance, image_meta, injected_files, [ 657.087482] env[61923]: ERROR nova.compute.manager [instance: 03100310-86ce-4739-8636-cd2aa1b542ff] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 657.087482] env[61923]: ERROR nova.compute.manager [instance: 03100310-86ce-4739-8636-cd2aa1b542ff] vm_ref = self.build_virtual_machine(instance, [ 657.087482] env[61923]: ERROR nova.compute.manager [instance: 03100310-86ce-4739-8636-cd2aa1b542ff] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 657.087482] env[61923]: ERROR nova.compute.manager [instance: 03100310-86ce-4739-8636-cd2aa1b542ff] vif_infos = vmwarevif.get_vif_info(self._session, [ 657.087482] env[61923]: ERROR nova.compute.manager [instance: 03100310-86ce-4739-8636-cd2aa1b542ff] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 657.087905] env[61923]: ERROR nova.compute.manager [instance: 03100310-86ce-4739-8636-cd2aa1b542ff] for vif in network_info: [ 657.087905] env[61923]: ERROR nova.compute.manager [instance: 03100310-86ce-4739-8636-cd2aa1b542ff] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 657.087905] env[61923]: ERROR nova.compute.manager [instance: 03100310-86ce-4739-8636-cd2aa1b542ff] return self._sync_wrapper(fn, *args, **kwargs) [ 657.087905] env[61923]: ERROR nova.compute.manager [instance: 03100310-86ce-4739-8636-cd2aa1b542ff] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 657.087905] env[61923]: ERROR nova.compute.manager [instance: 03100310-86ce-4739-8636-cd2aa1b542ff] self.wait() [ 657.087905] env[61923]: ERROR nova.compute.manager [instance: 03100310-86ce-4739-8636-cd2aa1b542ff] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 657.087905] env[61923]: ERROR nova.compute.manager [instance: 03100310-86ce-4739-8636-cd2aa1b542ff] self[:] = self._gt.wait() [ 657.087905] env[61923]: ERROR nova.compute.manager [instance: 03100310-86ce-4739-8636-cd2aa1b542ff] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 657.087905] env[61923]: ERROR nova.compute.manager [instance: 03100310-86ce-4739-8636-cd2aa1b542ff] return self._exit_event.wait() [ 657.087905] env[61923]: ERROR nova.compute.manager [instance: 03100310-86ce-4739-8636-cd2aa1b542ff] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 657.087905] env[61923]: ERROR nova.compute.manager [instance: 03100310-86ce-4739-8636-cd2aa1b542ff] result = hub.switch() [ 657.087905] env[61923]: ERROR nova.compute.manager [instance: 03100310-86ce-4739-8636-cd2aa1b542ff] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 657.087905] env[61923]: ERROR nova.compute.manager [instance: 03100310-86ce-4739-8636-cd2aa1b542ff] return self.greenlet.switch() [ 657.088284] env[61923]: ERROR nova.compute.manager [instance: 03100310-86ce-4739-8636-cd2aa1b542ff] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 657.088284] env[61923]: ERROR nova.compute.manager [instance: 03100310-86ce-4739-8636-cd2aa1b542ff] result = function(*args, **kwargs) [ 657.088284] env[61923]: ERROR nova.compute.manager [instance: 03100310-86ce-4739-8636-cd2aa1b542ff] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 657.088284] env[61923]: ERROR nova.compute.manager [instance: 03100310-86ce-4739-8636-cd2aa1b542ff] return func(*args, **kwargs) [ 657.088284] env[61923]: ERROR nova.compute.manager [instance: 03100310-86ce-4739-8636-cd2aa1b542ff] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 657.088284] env[61923]: ERROR nova.compute.manager [instance: 03100310-86ce-4739-8636-cd2aa1b542ff] raise e [ 657.088284] env[61923]: ERROR nova.compute.manager [instance: 03100310-86ce-4739-8636-cd2aa1b542ff] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 657.088284] env[61923]: ERROR nova.compute.manager [instance: 03100310-86ce-4739-8636-cd2aa1b542ff] nwinfo = self.network_api.allocate_for_instance( [ 657.088284] env[61923]: ERROR nova.compute.manager [instance: 03100310-86ce-4739-8636-cd2aa1b542ff] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 657.088284] env[61923]: ERROR nova.compute.manager [instance: 03100310-86ce-4739-8636-cd2aa1b542ff] created_port_ids = self._update_ports_for_instance( [ 657.088284] env[61923]: ERROR nova.compute.manager [instance: 03100310-86ce-4739-8636-cd2aa1b542ff] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 657.088284] env[61923]: ERROR nova.compute.manager [instance: 03100310-86ce-4739-8636-cd2aa1b542ff] with excutils.save_and_reraise_exception(): [ 657.088284] env[61923]: ERROR nova.compute.manager [instance: 03100310-86ce-4739-8636-cd2aa1b542ff] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 657.089708] env[61923]: ERROR nova.compute.manager [instance: 03100310-86ce-4739-8636-cd2aa1b542ff] self.force_reraise() [ 657.089708] env[61923]: ERROR nova.compute.manager [instance: 03100310-86ce-4739-8636-cd2aa1b542ff] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 657.089708] env[61923]: ERROR nova.compute.manager [instance: 03100310-86ce-4739-8636-cd2aa1b542ff] raise self.value [ 657.089708] env[61923]: ERROR nova.compute.manager [instance: 03100310-86ce-4739-8636-cd2aa1b542ff] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 657.089708] env[61923]: ERROR nova.compute.manager [instance: 03100310-86ce-4739-8636-cd2aa1b542ff] updated_port = self._update_port( [ 657.089708] env[61923]: ERROR nova.compute.manager [instance: 03100310-86ce-4739-8636-cd2aa1b542ff] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 657.089708] env[61923]: ERROR nova.compute.manager [instance: 03100310-86ce-4739-8636-cd2aa1b542ff] _ensure_no_port_binding_failure(port) [ 657.089708] env[61923]: ERROR nova.compute.manager [instance: 03100310-86ce-4739-8636-cd2aa1b542ff] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 657.089708] env[61923]: ERROR nova.compute.manager [instance: 03100310-86ce-4739-8636-cd2aa1b542ff] raise exception.PortBindingFailed(port_id=port['id']) [ 657.089708] env[61923]: ERROR nova.compute.manager [instance: 03100310-86ce-4739-8636-cd2aa1b542ff] nova.exception.PortBindingFailed: Binding failed for port 5a9feebd-c914-422f-af20-45e4709470e8, please check neutron logs for more information. [ 657.089708] env[61923]: ERROR nova.compute.manager [instance: 03100310-86ce-4739-8636-cd2aa1b542ff] [ 657.090055] env[61923]: DEBUG nova.compute.utils [None req-753531fa-39a1-4c82-9915-d195bf0fe1b2 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] [instance: 03100310-86ce-4739-8636-cd2aa1b542ff] Binding failed for port 5a9feebd-c914-422f-af20-45e4709470e8, please check neutron logs for more information. {{(pid=61923) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 657.090055] env[61923]: DEBUG oslo_concurrency.lockutils [None req-859ffdff-aaee-46b4-bdd6-0a15cce9792f tempest-ServerRescueNegativeTestJSON-1435265126 tempest-ServerRescueNegativeTestJSON-1435265126-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.707s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 657.091154] env[61923]: INFO nova.compute.claims [None req-859ffdff-aaee-46b4-bdd6-0a15cce9792f tempest-ServerRescueNegativeTestJSON-1435265126 tempest-ServerRescueNegativeTestJSON-1435265126-project-member] [instance: 41b6744e-d548-442c-b6cc-e12bd403f229] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 657.097019] env[61923]: DEBUG nova.compute.manager [None req-753531fa-39a1-4c82-9915-d195bf0fe1b2 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] [instance: 03100310-86ce-4739-8636-cd2aa1b542ff] Build of instance 03100310-86ce-4739-8636-cd2aa1b542ff was re-scheduled: Binding failed for port 5a9feebd-c914-422f-af20-45e4709470e8, please check neutron logs for more information. {{(pid=61923) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 657.097019] env[61923]: DEBUG nova.compute.manager [None req-753531fa-39a1-4c82-9915-d195bf0fe1b2 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] [instance: 03100310-86ce-4739-8636-cd2aa1b542ff] Unplugging VIFs for instance {{(pid=61923) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 657.097019] env[61923]: DEBUG oslo_concurrency.lockutils [None req-753531fa-39a1-4c82-9915-d195bf0fe1b2 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] Acquiring lock "refresh_cache-03100310-86ce-4739-8636-cd2aa1b542ff" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 657.097019] env[61923]: DEBUG oslo_concurrency.lockutils [None req-753531fa-39a1-4c82-9915-d195bf0fe1b2 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] Acquired lock "refresh_cache-03100310-86ce-4739-8636-cd2aa1b542ff" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 657.097513] env[61923]: DEBUG nova.network.neutron [None req-753531fa-39a1-4c82-9915-d195bf0fe1b2 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] [instance: 03100310-86ce-4739-8636-cd2aa1b542ff] Building network info cache for instance {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 657.287272] env[61923]: DEBUG oslo_concurrency.lockutils [None req-5138192d-6349-432b-b7de-64ec5d391f8d tempest-ServerActionsTestOtherB-987045973 tempest-ServerActionsTestOtherB-987045973-project-member] Acquiring lock "b0bc08c1-38d5-45ee-b521-1cffda41c77c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 657.287583] env[61923]: DEBUG oslo_concurrency.lockutils [None req-5138192d-6349-432b-b7de-64ec5d391f8d tempest-ServerActionsTestOtherB-987045973 tempest-ServerActionsTestOtherB-987045973-project-member] Lock "b0bc08c1-38d5-45ee-b521-1cffda41c77c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 657.390104] env[61923]: INFO nova.compute.manager [None req-cb394400-3d4b-4149-a290-413641bb6f18 tempest-InstanceActionsV221TestJSON-1730893270 tempest-InstanceActionsV221TestJSON-1730893270-project-member] [instance: dc8cc22d-aba9-4dba-b456-1b323013d234] Took 1.03 seconds to deallocate network for instance. [ 657.422064] env[61923]: DEBUG nova.compute.manager [None req-5f08bfba-63c5-4161-be1b-a884c0ae71af tempest-VolumesAdminNegativeTest-1422120048 tempest-VolumesAdminNegativeTest-1422120048-project-member] [instance: 590907a7-b6cc-48fc-a4bf-e4a2e48b05b8] Starting instance... {{(pid=61923) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 657.620715] env[61923]: DEBUG nova.network.neutron [None req-753531fa-39a1-4c82-9915-d195bf0fe1b2 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] [instance: 03100310-86ce-4739-8636-cd2aa1b542ff] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 657.736099] env[61923]: DEBUG nova.network.neutron [None req-753531fa-39a1-4c82-9915-d195bf0fe1b2 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] [instance: 03100310-86ce-4739-8636-cd2aa1b542ff] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 657.947715] env[61923]: DEBUG oslo_concurrency.lockutils [None req-5f08bfba-63c5-4161-be1b-a884c0ae71af tempest-VolumesAdminNegativeTest-1422120048 tempest-VolumesAdminNegativeTest-1422120048-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 658.238961] env[61923]: DEBUG oslo_concurrency.lockutils [None req-753531fa-39a1-4c82-9915-d195bf0fe1b2 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] Releasing lock "refresh_cache-03100310-86ce-4739-8636-cd2aa1b542ff" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 658.240058] env[61923]: DEBUG nova.compute.manager [None req-753531fa-39a1-4c82-9915-d195bf0fe1b2 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61923) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 658.240058] env[61923]: DEBUG nova.compute.manager [None req-753531fa-39a1-4c82-9915-d195bf0fe1b2 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] [instance: 03100310-86ce-4739-8636-cd2aa1b542ff] Deallocating network for instance {{(pid=61923) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 658.240058] env[61923]: DEBUG nova.network.neutron [None req-753531fa-39a1-4c82-9915-d195bf0fe1b2 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] [instance: 03100310-86ce-4739-8636-cd2aa1b542ff] deallocate_for_instance() {{(pid=61923) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 658.254437] env[61923]: DEBUG nova.network.neutron [None req-753531fa-39a1-4c82-9915-d195bf0fe1b2 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] [instance: 03100310-86ce-4739-8636-cd2aa1b542ff] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 658.416822] env[61923]: INFO nova.scheduler.client.report [None req-cb394400-3d4b-4149-a290-413641bb6f18 tempest-InstanceActionsV221TestJSON-1730893270 tempest-InstanceActionsV221TestJSON-1730893270-project-member] Deleted allocations for instance dc8cc22d-aba9-4dba-b456-1b323013d234 [ 658.516090] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a581aeef-10c4-4dbd-a0bd-22c06e98b43f {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.524034] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f8d766c-8bb5-4a20-8fc8-3011d6f3caea {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.554426] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4cb81fc4-f369-4d60-8622-79d1ec28bc92 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.561516] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-585d503b-a222-4a8e-91a3-4b61abcf1202 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.574507] env[61923]: DEBUG nova.compute.provider_tree [None req-859ffdff-aaee-46b4-bdd6-0a15cce9792f tempest-ServerRescueNegativeTestJSON-1435265126 tempest-ServerRescueNegativeTestJSON-1435265126-project-member] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 658.757618] env[61923]: DEBUG nova.network.neutron [None req-753531fa-39a1-4c82-9915-d195bf0fe1b2 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] [instance: 03100310-86ce-4739-8636-cd2aa1b542ff] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 658.932026] env[61923]: DEBUG oslo_concurrency.lockutils [None req-cb394400-3d4b-4149-a290-413641bb6f18 tempest-InstanceActionsV221TestJSON-1730893270 tempest-InstanceActionsV221TestJSON-1730893270-project-member] Lock "dc8cc22d-aba9-4dba-b456-1b323013d234" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 110.995s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 659.077550] env[61923]: DEBUG nova.scheduler.client.report [None req-859ffdff-aaee-46b4-bdd6-0a15cce9792f tempest-ServerRescueNegativeTestJSON-1435265126 tempest-ServerRescueNegativeTestJSON-1435265126-project-member] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 659.263036] env[61923]: INFO nova.compute.manager [None req-753531fa-39a1-4c82-9915-d195bf0fe1b2 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] [instance: 03100310-86ce-4739-8636-cd2aa1b542ff] Took 1.02 seconds to deallocate network for instance. [ 659.434813] env[61923]: DEBUG nova.compute.manager [None req-08382130-d70f-42e8-a5b5-93e86f5fa49a tempest-AttachVolumeShelveTestJSON-1252849694 tempest-AttachVolumeShelveTestJSON-1252849694-project-member] [instance: 60ffc8b0-4a7b-4e2a-8774-d16e502d8a06] Starting instance... {{(pid=61923) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 659.539222] env[61923]: DEBUG oslo_concurrency.lockutils [None req-f1ee1de3-0f72-403f-9698-da4536b6f37f tempest-ServerDiagnosticsV248Test-1822646244 tempest-ServerDiagnosticsV248Test-1822646244-project-member] Acquiring lock "655b01ae-76a5-4c09-9245-e629fec8fd4b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 659.539456] env[61923]: DEBUG oslo_concurrency.lockutils [None req-f1ee1de3-0f72-403f-9698-da4536b6f37f tempest-ServerDiagnosticsV248Test-1822646244 tempest-ServerDiagnosticsV248Test-1822646244-project-member] Lock "655b01ae-76a5-4c09-9245-e629fec8fd4b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 659.582623] env[61923]: DEBUG oslo_concurrency.lockutils [None req-859ffdff-aaee-46b4-bdd6-0a15cce9792f tempest-ServerRescueNegativeTestJSON-1435265126 tempest-ServerRescueNegativeTestJSON-1435265126-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.493s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 659.583447] env[61923]: DEBUG nova.compute.manager [None req-859ffdff-aaee-46b4-bdd6-0a15cce9792f tempest-ServerRescueNegativeTestJSON-1435265126 tempest-ServerRescueNegativeTestJSON-1435265126-project-member] [instance: 41b6744e-d548-442c-b6cc-e12bd403f229] Start building networks asynchronously for instance. {{(pid=61923) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 659.585326] env[61923]: DEBUG oslo_concurrency.lockutils [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 14.572s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 659.960287] env[61923]: DEBUG oslo_concurrency.lockutils [None req-08382130-d70f-42e8-a5b5-93e86f5fa49a tempest-AttachVolumeShelveTestJSON-1252849694 tempest-AttachVolumeShelveTestJSON-1252849694-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 660.094724] env[61923]: DEBUG nova.compute.utils [None req-859ffdff-aaee-46b4-bdd6-0a15cce9792f tempest-ServerRescueNegativeTestJSON-1435265126 tempest-ServerRescueNegativeTestJSON-1435265126-project-member] Using /dev/sd instead of None {{(pid=61923) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 660.095098] env[61923]: DEBUG nova.compute.manager [None req-859ffdff-aaee-46b4-bdd6-0a15cce9792f tempest-ServerRescueNegativeTestJSON-1435265126 tempest-ServerRescueNegativeTestJSON-1435265126-project-member] [instance: 41b6744e-d548-442c-b6cc-e12bd403f229] Allocating IP information in the background. {{(pid=61923) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 660.095462] env[61923]: DEBUG nova.network.neutron [None req-859ffdff-aaee-46b4-bdd6-0a15cce9792f tempest-ServerRescueNegativeTestJSON-1435265126 tempest-ServerRescueNegativeTestJSON-1435265126-project-member] [instance: 41b6744e-d548-442c-b6cc-e12bd403f229] allocate_for_instance() {{(pid=61923) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 660.169812] env[61923]: DEBUG nova.policy [None req-859ffdff-aaee-46b4-bdd6-0a15cce9792f tempest-ServerRescueNegativeTestJSON-1435265126 tempest-ServerRescueNegativeTestJSON-1435265126-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'de8bf84e21494593be329e3dd63da127', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '62d9e2aeadec483e881ea244b439d557', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61923) authorize /opt/stack/nova/nova/policy.py:201}} [ 660.295020] env[61923]: INFO nova.scheduler.client.report [None req-753531fa-39a1-4c82-9915-d195bf0fe1b2 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] Deleted allocations for instance 03100310-86ce-4739-8636-cd2aa1b542ff [ 660.517588] env[61923]: DEBUG nova.network.neutron [None req-859ffdff-aaee-46b4-bdd6-0a15cce9792f tempest-ServerRescueNegativeTestJSON-1435265126 tempest-ServerRescueNegativeTestJSON-1435265126-project-member] [instance: 41b6744e-d548-442c-b6cc-e12bd403f229] Successfully created port: 32a4c72c-9bfd-4f5f-a3e0-6a0ca4c9b6d2 {{(pid=61923) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 660.598862] env[61923]: DEBUG nova.compute.manager [None req-859ffdff-aaee-46b4-bdd6-0a15cce9792f tempest-ServerRescueNegativeTestJSON-1435265126 tempest-ServerRescueNegativeTestJSON-1435265126-project-member] [instance: 41b6744e-d548-442c-b6cc-e12bd403f229] Start building block device mappings for instance. {{(pid=61923) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 660.622579] env[61923]: DEBUG nova.compute.resource_tracker [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Instance c0f87aa5-447b-4920-8251-354d89a14fbb actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61923) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 660.626152] env[61923]: DEBUG nova.compute.resource_tracker [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Instance 525f37d5-2262-40c8-a339-fc262c53ce6e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61923) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 660.626152] env[61923]: DEBUG nova.compute.resource_tracker [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Instance 55b60fc2-6484-4fad-aa40-3e73e020841a actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61923) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 660.626152] env[61923]: DEBUG nova.compute.resource_tracker [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Instance aeca6a94-478d-435a-9d93-59936f26654e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61923) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 660.626152] env[61923]: DEBUG nova.compute.resource_tracker [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Instance 41b6744e-d548-442c-b6cc-e12bd403f229 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61923) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 660.802145] env[61923]: DEBUG oslo_concurrency.lockutils [None req-753531fa-39a1-4c82-9915-d195bf0fe1b2 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] Lock "03100310-86ce-4739-8636-cd2aa1b542ff" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 111.535s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 661.130007] env[61923]: DEBUG nova.compute.resource_tracker [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Instance c94e8e46-7697-426c-ae2f-aece493fa8f4 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61923) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 661.305047] env[61923]: DEBUG nova.compute.manager [None req-6833b601-845b-439d-ad97-e2f9c0927775 tempest-ServerRescueNegativeTestJSON-1435265126 tempest-ServerRescueNegativeTestJSON-1435265126-project-member] [instance: 4f5348ec-f0b6-45e2-be04-31dbf1d49a3a] Starting instance... {{(pid=61923) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 661.620291] env[61923]: DEBUG nova.compute.manager [None req-859ffdff-aaee-46b4-bdd6-0a15cce9792f tempest-ServerRescueNegativeTestJSON-1435265126 tempest-ServerRescueNegativeTestJSON-1435265126-project-member] [instance: 41b6744e-d548-442c-b6cc-e12bd403f229] Start spawning the instance on the hypervisor. {{(pid=61923) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 661.640237] env[61923]: DEBUG nova.compute.resource_tracker [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Instance 64584976-b3f3-4da5-a76b-a05cf2ed6aa4 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61923) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 661.649217] env[61923]: DEBUG nova.virt.hardware [None req-859ffdff-aaee-46b4-bdd6-0a15cce9792f tempest-ServerRescueNegativeTestJSON-1435265126 tempest-ServerRescueNegativeTestJSON-1435265126-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-29T20:07:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-29T20:07:35Z,direct_url=,disk_format='vmdk',id=0f153f63-ae0a-45b1-b7f5-7f9b673c947f,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='4e7b5e3b3c3443c8bd5c70f8a15739f5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-29T20:07:36Z,virtual_size=,visibility=), allow threads: False {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 661.649986] env[61923]: DEBUG nova.virt.hardware [None req-859ffdff-aaee-46b4-bdd6-0a15cce9792f tempest-ServerRescueNegativeTestJSON-1435265126 tempest-ServerRescueNegativeTestJSON-1435265126-project-member] Flavor limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 661.649986] env[61923]: DEBUG nova.virt.hardware [None req-859ffdff-aaee-46b4-bdd6-0a15cce9792f tempest-ServerRescueNegativeTestJSON-1435265126 tempest-ServerRescueNegativeTestJSON-1435265126-project-member] Image limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 661.649986] env[61923]: DEBUG nova.virt.hardware [None req-859ffdff-aaee-46b4-bdd6-0a15cce9792f tempest-ServerRescueNegativeTestJSON-1435265126 tempest-ServerRescueNegativeTestJSON-1435265126-project-member] Flavor pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 661.649986] env[61923]: DEBUG nova.virt.hardware [None req-859ffdff-aaee-46b4-bdd6-0a15cce9792f tempest-ServerRescueNegativeTestJSON-1435265126 tempest-ServerRescueNegativeTestJSON-1435265126-project-member] Image pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 661.650156] env[61923]: DEBUG nova.virt.hardware [None req-859ffdff-aaee-46b4-bdd6-0a15cce9792f tempest-ServerRescueNegativeTestJSON-1435265126 tempest-ServerRescueNegativeTestJSON-1435265126-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 661.650415] env[61923]: DEBUG nova.virt.hardware [None req-859ffdff-aaee-46b4-bdd6-0a15cce9792f tempest-ServerRescueNegativeTestJSON-1435265126 tempest-ServerRescueNegativeTestJSON-1435265126-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 661.650990] env[61923]: DEBUG nova.virt.hardware [None req-859ffdff-aaee-46b4-bdd6-0a15cce9792f tempest-ServerRescueNegativeTestJSON-1435265126 tempest-ServerRescueNegativeTestJSON-1435265126-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 661.650990] env[61923]: DEBUG nova.virt.hardware [None req-859ffdff-aaee-46b4-bdd6-0a15cce9792f tempest-ServerRescueNegativeTestJSON-1435265126 tempest-ServerRescueNegativeTestJSON-1435265126-project-member] Got 1 possible topologies {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 661.651097] env[61923]: DEBUG nova.virt.hardware [None req-859ffdff-aaee-46b4-bdd6-0a15cce9792f tempest-ServerRescueNegativeTestJSON-1435265126 tempest-ServerRescueNegativeTestJSON-1435265126-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 661.651288] env[61923]: DEBUG nova.virt.hardware [None req-859ffdff-aaee-46b4-bdd6-0a15cce9792f tempest-ServerRescueNegativeTestJSON-1435265126 tempest-ServerRescueNegativeTestJSON-1435265126-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 661.652410] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57e6aca4-af6f-40f4-9bc0-2983d78fd4b6 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.662050] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7700674c-828e-406d-9e8e-5a2a13bd923e {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 661.754939] env[61923]: DEBUG nova.compute.manager [req-221a06af-8597-44f5-ae75-4b84bf29b02f req-fa80152e-c305-4db6-93ef-657598365baf service nova] [instance: 41b6744e-d548-442c-b6cc-e12bd403f229] Received event network-changed-32a4c72c-9bfd-4f5f-a3e0-6a0ca4c9b6d2 {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 661.754939] env[61923]: DEBUG nova.compute.manager [req-221a06af-8597-44f5-ae75-4b84bf29b02f req-fa80152e-c305-4db6-93ef-657598365baf service nova] [instance: 41b6744e-d548-442c-b6cc-e12bd403f229] Refreshing instance network info cache due to event network-changed-32a4c72c-9bfd-4f5f-a3e0-6a0ca4c9b6d2. {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 661.754939] env[61923]: DEBUG oslo_concurrency.lockutils [req-221a06af-8597-44f5-ae75-4b84bf29b02f req-fa80152e-c305-4db6-93ef-657598365baf service nova] Acquiring lock "refresh_cache-41b6744e-d548-442c-b6cc-e12bd403f229" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 661.754939] env[61923]: DEBUG oslo_concurrency.lockutils [req-221a06af-8597-44f5-ae75-4b84bf29b02f req-fa80152e-c305-4db6-93ef-657598365baf service nova] Acquired lock "refresh_cache-41b6744e-d548-442c-b6cc-e12bd403f229" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 661.754939] env[61923]: DEBUG nova.network.neutron [req-221a06af-8597-44f5-ae75-4b84bf29b02f req-fa80152e-c305-4db6-93ef-657598365baf service nova] [instance: 41b6744e-d548-442c-b6cc-e12bd403f229] Refreshing network info cache for port 32a4c72c-9bfd-4f5f-a3e0-6a0ca4c9b6d2 {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 661.835946] env[61923]: DEBUG oslo_concurrency.lockutils [None req-6833b601-845b-439d-ad97-e2f9c0927775 tempest-ServerRescueNegativeTestJSON-1435265126 tempest-ServerRescueNegativeTestJSON-1435265126-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 662.042119] env[61923]: ERROR nova.compute.manager [None req-859ffdff-aaee-46b4-bdd6-0a15cce9792f tempest-ServerRescueNegativeTestJSON-1435265126 tempest-ServerRescueNegativeTestJSON-1435265126-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 32a4c72c-9bfd-4f5f-a3e0-6a0ca4c9b6d2, please check neutron logs for more information. [ 662.042119] env[61923]: ERROR nova.compute.manager Traceback (most recent call last): [ 662.042119] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 662.042119] env[61923]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 662.042119] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 662.042119] env[61923]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 662.042119] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 662.042119] env[61923]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 662.042119] env[61923]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 662.042119] env[61923]: ERROR nova.compute.manager self.force_reraise() [ 662.042119] env[61923]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 662.042119] env[61923]: ERROR nova.compute.manager raise self.value [ 662.042119] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 662.042119] env[61923]: ERROR nova.compute.manager updated_port = self._update_port( [ 662.042119] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 662.042119] env[61923]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 662.042878] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 662.042878] env[61923]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 662.042878] env[61923]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 32a4c72c-9bfd-4f5f-a3e0-6a0ca4c9b6d2, please check neutron logs for more information. [ 662.042878] env[61923]: ERROR nova.compute.manager [ 662.042878] env[61923]: Traceback (most recent call last): [ 662.042878] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 662.042878] env[61923]: listener.cb(fileno) [ 662.042878] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 662.042878] env[61923]: result = function(*args, **kwargs) [ 662.042878] env[61923]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 662.042878] env[61923]: return func(*args, **kwargs) [ 662.042878] env[61923]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 662.042878] env[61923]: raise e [ 662.042878] env[61923]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 662.042878] env[61923]: nwinfo = self.network_api.allocate_for_instance( [ 662.042878] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 662.042878] env[61923]: created_port_ids = self._update_ports_for_instance( [ 662.042878] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 662.042878] env[61923]: with excutils.save_and_reraise_exception(): [ 662.042878] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 662.042878] env[61923]: self.force_reraise() [ 662.042878] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 662.042878] env[61923]: raise self.value [ 662.042878] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 662.042878] env[61923]: updated_port = self._update_port( [ 662.042878] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 662.042878] env[61923]: _ensure_no_port_binding_failure(port) [ 662.042878] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 662.042878] env[61923]: raise exception.PortBindingFailed(port_id=port['id']) [ 662.044236] env[61923]: nova.exception.PortBindingFailed: Binding failed for port 32a4c72c-9bfd-4f5f-a3e0-6a0ca4c9b6d2, please check neutron logs for more information. [ 662.044236] env[61923]: Removing descriptor: 14 [ 662.044236] env[61923]: ERROR nova.compute.manager [None req-859ffdff-aaee-46b4-bdd6-0a15cce9792f tempest-ServerRescueNegativeTestJSON-1435265126 tempest-ServerRescueNegativeTestJSON-1435265126-project-member] [instance: 41b6744e-d548-442c-b6cc-e12bd403f229] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 32a4c72c-9bfd-4f5f-a3e0-6a0ca4c9b6d2, please check neutron logs for more information. [ 662.044236] env[61923]: ERROR nova.compute.manager [instance: 41b6744e-d548-442c-b6cc-e12bd403f229] Traceback (most recent call last): [ 662.044236] env[61923]: ERROR nova.compute.manager [instance: 41b6744e-d548-442c-b6cc-e12bd403f229] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 662.044236] env[61923]: ERROR nova.compute.manager [instance: 41b6744e-d548-442c-b6cc-e12bd403f229] yield resources [ 662.044236] env[61923]: ERROR nova.compute.manager [instance: 41b6744e-d548-442c-b6cc-e12bd403f229] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 662.044236] env[61923]: ERROR nova.compute.manager [instance: 41b6744e-d548-442c-b6cc-e12bd403f229] self.driver.spawn(context, instance, image_meta, [ 662.044236] env[61923]: ERROR nova.compute.manager [instance: 41b6744e-d548-442c-b6cc-e12bd403f229] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 662.044236] env[61923]: ERROR nova.compute.manager [instance: 41b6744e-d548-442c-b6cc-e12bd403f229] self._vmops.spawn(context, instance, image_meta, injected_files, [ 662.044236] env[61923]: ERROR nova.compute.manager [instance: 41b6744e-d548-442c-b6cc-e12bd403f229] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 662.044236] env[61923]: ERROR nova.compute.manager [instance: 41b6744e-d548-442c-b6cc-e12bd403f229] vm_ref = self.build_virtual_machine(instance, [ 662.044831] env[61923]: ERROR nova.compute.manager [instance: 41b6744e-d548-442c-b6cc-e12bd403f229] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 662.044831] env[61923]: ERROR nova.compute.manager [instance: 41b6744e-d548-442c-b6cc-e12bd403f229] vif_infos = vmwarevif.get_vif_info(self._session, [ 662.044831] env[61923]: ERROR nova.compute.manager [instance: 41b6744e-d548-442c-b6cc-e12bd403f229] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 662.044831] env[61923]: ERROR nova.compute.manager [instance: 41b6744e-d548-442c-b6cc-e12bd403f229] for vif in network_info: [ 662.044831] env[61923]: ERROR nova.compute.manager [instance: 41b6744e-d548-442c-b6cc-e12bd403f229] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 662.044831] env[61923]: ERROR nova.compute.manager [instance: 41b6744e-d548-442c-b6cc-e12bd403f229] return self._sync_wrapper(fn, *args, **kwargs) [ 662.044831] env[61923]: ERROR nova.compute.manager [instance: 41b6744e-d548-442c-b6cc-e12bd403f229] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 662.044831] env[61923]: ERROR nova.compute.manager [instance: 41b6744e-d548-442c-b6cc-e12bd403f229] self.wait() [ 662.044831] env[61923]: ERROR nova.compute.manager [instance: 41b6744e-d548-442c-b6cc-e12bd403f229] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 662.044831] env[61923]: ERROR nova.compute.manager [instance: 41b6744e-d548-442c-b6cc-e12bd403f229] self[:] = self._gt.wait() [ 662.044831] env[61923]: ERROR nova.compute.manager [instance: 41b6744e-d548-442c-b6cc-e12bd403f229] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 662.044831] env[61923]: ERROR nova.compute.manager [instance: 41b6744e-d548-442c-b6cc-e12bd403f229] return self._exit_event.wait() [ 662.044831] env[61923]: ERROR nova.compute.manager [instance: 41b6744e-d548-442c-b6cc-e12bd403f229] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 662.045787] env[61923]: ERROR nova.compute.manager [instance: 41b6744e-d548-442c-b6cc-e12bd403f229] result = hub.switch() [ 662.045787] env[61923]: ERROR nova.compute.manager [instance: 41b6744e-d548-442c-b6cc-e12bd403f229] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 662.045787] env[61923]: ERROR nova.compute.manager [instance: 41b6744e-d548-442c-b6cc-e12bd403f229] return self.greenlet.switch() [ 662.045787] env[61923]: ERROR nova.compute.manager [instance: 41b6744e-d548-442c-b6cc-e12bd403f229] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 662.045787] env[61923]: ERROR nova.compute.manager [instance: 41b6744e-d548-442c-b6cc-e12bd403f229] result = function(*args, **kwargs) [ 662.045787] env[61923]: ERROR nova.compute.manager [instance: 41b6744e-d548-442c-b6cc-e12bd403f229] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 662.045787] env[61923]: ERROR nova.compute.manager [instance: 41b6744e-d548-442c-b6cc-e12bd403f229] return func(*args, **kwargs) [ 662.045787] env[61923]: ERROR nova.compute.manager [instance: 41b6744e-d548-442c-b6cc-e12bd403f229] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 662.045787] env[61923]: ERROR nova.compute.manager [instance: 41b6744e-d548-442c-b6cc-e12bd403f229] raise e [ 662.045787] env[61923]: ERROR nova.compute.manager [instance: 41b6744e-d548-442c-b6cc-e12bd403f229] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 662.045787] env[61923]: ERROR nova.compute.manager [instance: 41b6744e-d548-442c-b6cc-e12bd403f229] nwinfo = self.network_api.allocate_for_instance( [ 662.045787] env[61923]: ERROR nova.compute.manager [instance: 41b6744e-d548-442c-b6cc-e12bd403f229] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 662.045787] env[61923]: ERROR nova.compute.manager [instance: 41b6744e-d548-442c-b6cc-e12bd403f229] created_port_ids = self._update_ports_for_instance( [ 662.046142] env[61923]: ERROR nova.compute.manager [instance: 41b6744e-d548-442c-b6cc-e12bd403f229] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 662.046142] env[61923]: ERROR nova.compute.manager [instance: 41b6744e-d548-442c-b6cc-e12bd403f229] with excutils.save_and_reraise_exception(): [ 662.046142] env[61923]: ERROR nova.compute.manager [instance: 41b6744e-d548-442c-b6cc-e12bd403f229] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 662.046142] env[61923]: ERROR nova.compute.manager [instance: 41b6744e-d548-442c-b6cc-e12bd403f229] self.force_reraise() [ 662.046142] env[61923]: ERROR nova.compute.manager [instance: 41b6744e-d548-442c-b6cc-e12bd403f229] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 662.046142] env[61923]: ERROR nova.compute.manager [instance: 41b6744e-d548-442c-b6cc-e12bd403f229] raise self.value [ 662.046142] env[61923]: ERROR nova.compute.manager [instance: 41b6744e-d548-442c-b6cc-e12bd403f229] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 662.046142] env[61923]: ERROR nova.compute.manager [instance: 41b6744e-d548-442c-b6cc-e12bd403f229] updated_port = self._update_port( [ 662.046142] env[61923]: ERROR nova.compute.manager [instance: 41b6744e-d548-442c-b6cc-e12bd403f229] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 662.046142] env[61923]: ERROR nova.compute.manager [instance: 41b6744e-d548-442c-b6cc-e12bd403f229] _ensure_no_port_binding_failure(port) [ 662.046142] env[61923]: ERROR nova.compute.manager [instance: 41b6744e-d548-442c-b6cc-e12bd403f229] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 662.046142] env[61923]: ERROR nova.compute.manager [instance: 41b6744e-d548-442c-b6cc-e12bd403f229] raise exception.PortBindingFailed(port_id=port['id']) [ 662.046445] env[61923]: ERROR nova.compute.manager [instance: 41b6744e-d548-442c-b6cc-e12bd403f229] nova.exception.PortBindingFailed: Binding failed for port 32a4c72c-9bfd-4f5f-a3e0-6a0ca4c9b6d2, please check neutron logs for more information. [ 662.046445] env[61923]: ERROR nova.compute.manager [instance: 41b6744e-d548-442c-b6cc-e12bd403f229] [ 662.046445] env[61923]: INFO nova.compute.manager [None req-859ffdff-aaee-46b4-bdd6-0a15cce9792f tempest-ServerRescueNegativeTestJSON-1435265126 tempest-ServerRescueNegativeTestJSON-1435265126-project-member] [instance: 41b6744e-d548-442c-b6cc-e12bd403f229] Terminating instance [ 662.046445] env[61923]: DEBUG oslo_concurrency.lockutils [None req-859ffdff-aaee-46b4-bdd6-0a15cce9792f tempest-ServerRescueNegativeTestJSON-1435265126 tempest-ServerRescueNegativeTestJSON-1435265126-project-member] Acquiring lock "refresh_cache-41b6744e-d548-442c-b6cc-e12bd403f229" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 662.143702] env[61923]: DEBUG nova.compute.resource_tracker [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Instance 590907a7-b6cc-48fc-a4bf-e4a2e48b05b8 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61923) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 662.284346] env[61923]: DEBUG nova.network.neutron [req-221a06af-8597-44f5-ae75-4b84bf29b02f req-fa80152e-c305-4db6-93ef-657598365baf service nova] [instance: 41b6744e-d548-442c-b6cc-e12bd403f229] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 662.390494] env[61923]: DEBUG nova.network.neutron [req-221a06af-8597-44f5-ae75-4b84bf29b02f req-fa80152e-c305-4db6-93ef-657598365baf service nova] [instance: 41b6744e-d548-442c-b6cc-e12bd403f229] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 662.647108] env[61923]: DEBUG nova.compute.resource_tracker [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Instance 60ffc8b0-4a7b-4e2a-8774-d16e502d8a06 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61923) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 662.685506] env[61923]: DEBUG oslo_concurrency.lockutils [None req-347b83bf-b4ed-4904-9bd3-d6021183ae1c tempest-ServerShowV254Test-921228709 tempest-ServerShowV254Test-921228709-project-member] Acquiring lock "ecedb77f-4a2a-4047-964f-80796954d847" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 662.685758] env[61923]: DEBUG oslo_concurrency.lockutils [None req-347b83bf-b4ed-4904-9bd3-d6021183ae1c tempest-ServerShowV254Test-921228709 tempest-ServerShowV254Test-921228709-project-member] Lock "ecedb77f-4a2a-4047-964f-80796954d847" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 662.895984] env[61923]: DEBUG oslo_concurrency.lockutils [req-221a06af-8597-44f5-ae75-4b84bf29b02f req-fa80152e-c305-4db6-93ef-657598365baf service nova] Releasing lock "refresh_cache-41b6744e-d548-442c-b6cc-e12bd403f229" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 662.895984] env[61923]: DEBUG oslo_concurrency.lockutils [None req-859ffdff-aaee-46b4-bdd6-0a15cce9792f tempest-ServerRescueNegativeTestJSON-1435265126 tempest-ServerRescueNegativeTestJSON-1435265126-project-member] Acquired lock "refresh_cache-41b6744e-d548-442c-b6cc-e12bd403f229" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 662.895984] env[61923]: DEBUG nova.network.neutron [None req-859ffdff-aaee-46b4-bdd6-0a15cce9792f tempest-ServerRescueNegativeTestJSON-1435265126 tempest-ServerRescueNegativeTestJSON-1435265126-project-member] [instance: 41b6744e-d548-442c-b6cc-e12bd403f229] Building network info cache for instance {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 663.151332] env[61923]: DEBUG nova.compute.resource_tracker [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Instance 4f5348ec-f0b6-45e2-be04-31dbf1d49a3a has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61923) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 663.427752] env[61923]: DEBUG nova.network.neutron [None req-859ffdff-aaee-46b4-bdd6-0a15cce9792f tempest-ServerRescueNegativeTestJSON-1435265126 tempest-ServerRescueNegativeTestJSON-1435265126-project-member] [instance: 41b6744e-d548-442c-b6cc-e12bd403f229] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 663.541203] env[61923]: DEBUG nova.network.neutron [None req-859ffdff-aaee-46b4-bdd6-0a15cce9792f tempest-ServerRescueNegativeTestJSON-1435265126 tempest-ServerRescueNegativeTestJSON-1435265126-project-member] [instance: 41b6744e-d548-442c-b6cc-e12bd403f229] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 663.653600] env[61923]: DEBUG nova.compute.resource_tracker [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Instance fb548893-2df5-4629-bf3a-3b0f807f6357 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61923) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 663.780577] env[61923]: DEBUG nova.compute.manager [req-fef25364-f178-41da-a662-7c21220ab9d7 req-9f9834a2-1aec-4eaa-9f77-5105b929a715 service nova] [instance: 41b6744e-d548-442c-b6cc-e12bd403f229] Received event network-vif-deleted-32a4c72c-9bfd-4f5f-a3e0-6a0ca4c9b6d2 {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 663.907633] env[61923]: DEBUG oslo_concurrency.lockutils [None req-d3d04fc8-2bde-4c80-b2d6-dd8057edd757 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] Acquiring lock "b497f3f3-001f-4e5e-b6b7-8e2006116856" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 663.907876] env[61923]: DEBUG oslo_concurrency.lockutils [None req-d3d04fc8-2bde-4c80-b2d6-dd8057edd757 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] Lock "b497f3f3-001f-4e5e-b6b7-8e2006116856" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 664.046542] env[61923]: DEBUG oslo_concurrency.lockutils [None req-859ffdff-aaee-46b4-bdd6-0a15cce9792f tempest-ServerRescueNegativeTestJSON-1435265126 tempest-ServerRescueNegativeTestJSON-1435265126-project-member] Releasing lock "refresh_cache-41b6744e-d548-442c-b6cc-e12bd403f229" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 664.046542] env[61923]: DEBUG nova.compute.manager [None req-859ffdff-aaee-46b4-bdd6-0a15cce9792f tempest-ServerRescueNegativeTestJSON-1435265126 tempest-ServerRescueNegativeTestJSON-1435265126-project-member] [instance: 41b6744e-d548-442c-b6cc-e12bd403f229] Start destroying the instance on the hypervisor. {{(pid=61923) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 664.046542] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-859ffdff-aaee-46b4-bdd6-0a15cce9792f tempest-ServerRescueNegativeTestJSON-1435265126 tempest-ServerRescueNegativeTestJSON-1435265126-project-member] [instance: 41b6744e-d548-442c-b6cc-e12bd403f229] Destroying instance {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 664.047599] env[61923]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-72319d4e-6c6b-4099-815f-4d3dde9ea666 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 664.055782] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1ca8d02-e08c-4a71-bb23-206eed2e81cb {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 664.078914] env[61923]: WARNING nova.virt.vmwareapi.vmops [None req-859ffdff-aaee-46b4-bdd6-0a15cce9792f tempest-ServerRescueNegativeTestJSON-1435265126 tempest-ServerRescueNegativeTestJSON-1435265126-project-member] [instance: 41b6744e-d548-442c-b6cc-e12bd403f229] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 41b6744e-d548-442c-b6cc-e12bd403f229 could not be found. [ 664.079144] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-859ffdff-aaee-46b4-bdd6-0a15cce9792f tempest-ServerRescueNegativeTestJSON-1435265126 tempest-ServerRescueNegativeTestJSON-1435265126-project-member] [instance: 41b6744e-d548-442c-b6cc-e12bd403f229] Instance destroyed {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 664.079324] env[61923]: INFO nova.compute.manager [None req-859ffdff-aaee-46b4-bdd6-0a15cce9792f tempest-ServerRescueNegativeTestJSON-1435265126 tempest-ServerRescueNegativeTestJSON-1435265126-project-member] [instance: 41b6744e-d548-442c-b6cc-e12bd403f229] Took 0.03 seconds to destroy the instance on the hypervisor. [ 664.079552] env[61923]: DEBUG oslo.service.loopingcall [None req-859ffdff-aaee-46b4-bdd6-0a15cce9792f tempest-ServerRescueNegativeTestJSON-1435265126 tempest-ServerRescueNegativeTestJSON-1435265126-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61923) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 664.079812] env[61923]: DEBUG nova.compute.manager [-] [instance: 41b6744e-d548-442c-b6cc-e12bd403f229] Deallocating network for instance {{(pid=61923) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 664.079930] env[61923]: DEBUG nova.network.neutron [-] [instance: 41b6744e-d548-442c-b6cc-e12bd403f229] deallocate_for_instance() {{(pid=61923) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 664.094203] env[61923]: DEBUG nova.network.neutron [-] [instance: 41b6744e-d548-442c-b6cc-e12bd403f229] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 664.156123] env[61923]: DEBUG nova.compute.resource_tracker [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Instance 62144317-0354-4751-8f2b-4eae588c9da7 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61923) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 664.596524] env[61923]: DEBUG nova.network.neutron [-] [instance: 41b6744e-d548-442c-b6cc-e12bd403f229] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 664.658989] env[61923]: DEBUG nova.compute.resource_tracker [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Instance 25b62db6-6e64-4e38-bfd3-82a2e9a4883a has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61923) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 665.099551] env[61923]: INFO nova.compute.manager [-] [instance: 41b6744e-d548-442c-b6cc-e12bd403f229] Took 1.02 seconds to deallocate network for instance. [ 665.101891] env[61923]: DEBUG nova.compute.claims [None req-859ffdff-aaee-46b4-bdd6-0a15cce9792f tempest-ServerRescueNegativeTestJSON-1435265126 tempest-ServerRescueNegativeTestJSON-1435265126-project-member] [instance: 41b6744e-d548-442c-b6cc-e12bd403f229] Aborting claim: {{(pid=61923) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 665.102082] env[61923]: DEBUG oslo_concurrency.lockutils [None req-859ffdff-aaee-46b4-bdd6-0a15cce9792f tempest-ServerRescueNegativeTestJSON-1435265126 tempest-ServerRescueNegativeTestJSON-1435265126-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 665.162082] env[61923]: DEBUG nova.compute.resource_tracker [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Instance 2939d1e5-07fe-48b3-bc87-5f67fa1a7ae4 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61923) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 665.666652] env[61923]: DEBUG nova.compute.resource_tracker [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Instance 74b32299-983f-41f7-b3f4-3a12815a83e6 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61923) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 666.169881] env[61923]: DEBUG nova.compute.resource_tracker [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Instance 7c0341f9-7401-42ee-9512-afd832c76940 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61923) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 666.673178] env[61923]: DEBUG nova.compute.resource_tracker [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Instance f7e1d517-3e39-4381-83be-0152368b8120 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61923) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 667.176429] env[61923]: DEBUG nova.compute.resource_tracker [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Instance 0eab2200-2528-4a89-821a-cdd6a3c99297 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61923) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 667.679194] env[61923]: DEBUG nova.compute.resource_tracker [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Instance 22d1e8ed-75c5-42e9-ae55-12d59e565914 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61923) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 668.182900] env[61923]: DEBUG nova.compute.resource_tracker [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Instance 15eaaa7c-77be-4b8b-89d0-0b2b11a0c49e has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61923) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 668.686617] env[61923]: DEBUG nova.compute.resource_tracker [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Instance 62d8a170-e7ad-4d14-90d8-6f6ce32b7cf8 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61923) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 669.190810] env[61923]: DEBUG nova.compute.resource_tracker [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Instance b6712878-3270-43f4-9518-d41eb77124b7 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61923) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 669.695193] env[61923]: DEBUG nova.compute.resource_tracker [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Instance b0bc08c1-38d5-45ee-b521-1cffda41c77c has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61923) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 670.198289] env[61923]: DEBUG nova.compute.resource_tracker [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Instance 655b01ae-76a5-4c09-9245-e629fec8fd4b has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61923) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 670.198557] env[61923]: DEBUG nova.compute.resource_tracker [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Total usable vcpus: 48, total allocated vcpus: 5 {{(pid=61923) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 670.198687] env[61923]: DEBUG nova.compute.resource_tracker [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1472MB phys_disk=200GB used_disk=5GB total_vcpus=48 used_vcpus=5 pci_stats=[] {{(pid=61923) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 670.478574] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8d3adc7-ad97-46e4-b470-fd8013803401 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 670.486278] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18bbff63-aafc-4c11-83bf-73b62cc51a86 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 670.515991] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-052cccac-7cdd-44c0-9b7a-4fa3434a4222 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 670.523078] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-929b98b1-dc35-4b8a-b965-416b71701da4 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 670.535668] env[61923]: DEBUG nova.compute.provider_tree [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 671.039284] env[61923]: DEBUG nova.scheduler.client.report [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 671.546293] env[61923]: DEBUG nova.compute.resource_tracker [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61923) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 671.546551] env[61923]: DEBUG oslo_concurrency.lockutils [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 11.961s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 671.546826] env[61923]: DEBUG oslo_concurrency.lockutils [None req-76010439-9fb6-4296-9b5a-107842db76db tempest-ServerExternalEventsTest-1479796458 tempest-ServerExternalEventsTest-1479796458-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 26.015s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 672.324341] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea67b261-24d0-438f-83c8-fcc15e1da719 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 672.331905] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d54d7101-d472-4db0-a7ce-3a3792ae4018 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 672.360321] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10078d71-35a3-48f3-8095-ea1465fcd29c {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 672.367078] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-131aa3e1-110d-472d-9a71-da5093df38ba {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 672.380617] env[61923]: DEBUG nova.compute.provider_tree [None req-76010439-9fb6-4296-9b5a-107842db76db tempest-ServerExternalEventsTest-1479796458 tempest-ServerExternalEventsTest-1479796458-project-member] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 672.884195] env[61923]: DEBUG nova.scheduler.client.report [None req-76010439-9fb6-4296-9b5a-107842db76db tempest-ServerExternalEventsTest-1479796458 tempest-ServerExternalEventsTest-1479796458-project-member] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 673.390023] env[61923]: DEBUG oslo_concurrency.lockutils [None req-76010439-9fb6-4296-9b5a-107842db76db tempest-ServerExternalEventsTest-1479796458 tempest-ServerExternalEventsTest-1479796458-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.843s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 673.390447] env[61923]: ERROR nova.compute.manager [None req-76010439-9fb6-4296-9b5a-107842db76db tempest-ServerExternalEventsTest-1479796458 tempest-ServerExternalEventsTest-1479796458-project-member] [instance: c0f87aa5-447b-4920-8251-354d89a14fbb] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port c67bdeb7-ce64-47ec-9e91-62b8f48de195, please check neutron logs for more information. [ 673.390447] env[61923]: ERROR nova.compute.manager [instance: c0f87aa5-447b-4920-8251-354d89a14fbb] Traceback (most recent call last): [ 673.390447] env[61923]: ERROR nova.compute.manager [instance: c0f87aa5-447b-4920-8251-354d89a14fbb] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 673.390447] env[61923]: ERROR nova.compute.manager [instance: c0f87aa5-447b-4920-8251-354d89a14fbb] self.driver.spawn(context, instance, image_meta, [ 673.390447] env[61923]: ERROR nova.compute.manager [instance: c0f87aa5-447b-4920-8251-354d89a14fbb] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 673.390447] env[61923]: ERROR nova.compute.manager [instance: c0f87aa5-447b-4920-8251-354d89a14fbb] self._vmops.spawn(context, instance, image_meta, injected_files, [ 673.390447] env[61923]: ERROR nova.compute.manager [instance: c0f87aa5-447b-4920-8251-354d89a14fbb] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 673.390447] env[61923]: ERROR nova.compute.manager [instance: c0f87aa5-447b-4920-8251-354d89a14fbb] vm_ref = self.build_virtual_machine(instance, [ 673.390447] env[61923]: ERROR nova.compute.manager [instance: c0f87aa5-447b-4920-8251-354d89a14fbb] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 673.390447] env[61923]: ERROR nova.compute.manager [instance: c0f87aa5-447b-4920-8251-354d89a14fbb] vif_infos = vmwarevif.get_vif_info(self._session, [ 673.390447] env[61923]: ERROR nova.compute.manager [instance: c0f87aa5-447b-4920-8251-354d89a14fbb] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 673.390765] env[61923]: ERROR nova.compute.manager [instance: c0f87aa5-447b-4920-8251-354d89a14fbb] for vif in network_info: [ 673.390765] env[61923]: ERROR nova.compute.manager [instance: c0f87aa5-447b-4920-8251-354d89a14fbb] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 673.390765] env[61923]: ERROR nova.compute.manager [instance: c0f87aa5-447b-4920-8251-354d89a14fbb] return self._sync_wrapper(fn, *args, **kwargs) [ 673.390765] env[61923]: ERROR nova.compute.manager [instance: c0f87aa5-447b-4920-8251-354d89a14fbb] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 673.390765] env[61923]: ERROR nova.compute.manager [instance: c0f87aa5-447b-4920-8251-354d89a14fbb] self.wait() [ 673.390765] env[61923]: ERROR nova.compute.manager [instance: c0f87aa5-447b-4920-8251-354d89a14fbb] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 673.390765] env[61923]: ERROR nova.compute.manager [instance: c0f87aa5-447b-4920-8251-354d89a14fbb] self[:] = self._gt.wait() [ 673.390765] env[61923]: ERROR nova.compute.manager [instance: c0f87aa5-447b-4920-8251-354d89a14fbb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 673.390765] env[61923]: ERROR nova.compute.manager [instance: c0f87aa5-447b-4920-8251-354d89a14fbb] return self._exit_event.wait() [ 673.390765] env[61923]: ERROR nova.compute.manager [instance: c0f87aa5-447b-4920-8251-354d89a14fbb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 673.390765] env[61923]: ERROR nova.compute.manager [instance: c0f87aa5-447b-4920-8251-354d89a14fbb] result = hub.switch() [ 673.390765] env[61923]: ERROR nova.compute.manager [instance: c0f87aa5-447b-4920-8251-354d89a14fbb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 673.390765] env[61923]: ERROR nova.compute.manager [instance: c0f87aa5-447b-4920-8251-354d89a14fbb] return self.greenlet.switch() [ 673.391120] env[61923]: ERROR nova.compute.manager [instance: c0f87aa5-447b-4920-8251-354d89a14fbb] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 673.391120] env[61923]: ERROR nova.compute.manager [instance: c0f87aa5-447b-4920-8251-354d89a14fbb] result = function(*args, **kwargs) [ 673.391120] env[61923]: ERROR nova.compute.manager [instance: c0f87aa5-447b-4920-8251-354d89a14fbb] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 673.391120] env[61923]: ERROR nova.compute.manager [instance: c0f87aa5-447b-4920-8251-354d89a14fbb] return func(*args, **kwargs) [ 673.391120] env[61923]: ERROR nova.compute.manager [instance: c0f87aa5-447b-4920-8251-354d89a14fbb] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 673.391120] env[61923]: ERROR nova.compute.manager [instance: c0f87aa5-447b-4920-8251-354d89a14fbb] raise e [ 673.391120] env[61923]: ERROR nova.compute.manager [instance: c0f87aa5-447b-4920-8251-354d89a14fbb] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 673.391120] env[61923]: ERROR nova.compute.manager [instance: c0f87aa5-447b-4920-8251-354d89a14fbb] nwinfo = self.network_api.allocate_for_instance( [ 673.391120] env[61923]: ERROR nova.compute.manager [instance: c0f87aa5-447b-4920-8251-354d89a14fbb] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 673.391120] env[61923]: ERROR nova.compute.manager [instance: c0f87aa5-447b-4920-8251-354d89a14fbb] created_port_ids = self._update_ports_for_instance( [ 673.391120] env[61923]: ERROR nova.compute.manager [instance: c0f87aa5-447b-4920-8251-354d89a14fbb] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 673.391120] env[61923]: ERROR nova.compute.manager [instance: c0f87aa5-447b-4920-8251-354d89a14fbb] with excutils.save_and_reraise_exception(): [ 673.391120] env[61923]: ERROR nova.compute.manager [instance: c0f87aa5-447b-4920-8251-354d89a14fbb] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 673.391425] env[61923]: ERROR nova.compute.manager [instance: c0f87aa5-447b-4920-8251-354d89a14fbb] self.force_reraise() [ 673.391425] env[61923]: ERROR nova.compute.manager [instance: c0f87aa5-447b-4920-8251-354d89a14fbb] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 673.391425] env[61923]: ERROR nova.compute.manager [instance: c0f87aa5-447b-4920-8251-354d89a14fbb] raise self.value [ 673.391425] env[61923]: ERROR nova.compute.manager [instance: c0f87aa5-447b-4920-8251-354d89a14fbb] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 673.391425] env[61923]: ERROR nova.compute.manager [instance: c0f87aa5-447b-4920-8251-354d89a14fbb] updated_port = self._update_port( [ 673.391425] env[61923]: ERROR nova.compute.manager [instance: c0f87aa5-447b-4920-8251-354d89a14fbb] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 673.391425] env[61923]: ERROR nova.compute.manager [instance: c0f87aa5-447b-4920-8251-354d89a14fbb] _ensure_no_port_binding_failure(port) [ 673.391425] env[61923]: ERROR nova.compute.manager [instance: c0f87aa5-447b-4920-8251-354d89a14fbb] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 673.391425] env[61923]: ERROR nova.compute.manager [instance: c0f87aa5-447b-4920-8251-354d89a14fbb] raise exception.PortBindingFailed(port_id=port['id']) [ 673.391425] env[61923]: ERROR nova.compute.manager [instance: c0f87aa5-447b-4920-8251-354d89a14fbb] nova.exception.PortBindingFailed: Binding failed for port c67bdeb7-ce64-47ec-9e91-62b8f48de195, please check neutron logs for more information. [ 673.391425] env[61923]: ERROR nova.compute.manager [instance: c0f87aa5-447b-4920-8251-354d89a14fbb] [ 673.391818] env[61923]: DEBUG nova.compute.utils [None req-76010439-9fb6-4296-9b5a-107842db76db tempest-ServerExternalEventsTest-1479796458 tempest-ServerExternalEventsTest-1479796458-project-member] [instance: c0f87aa5-447b-4920-8251-354d89a14fbb] Binding failed for port c67bdeb7-ce64-47ec-9e91-62b8f48de195, please check neutron logs for more information. {{(pid=61923) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 673.392313] env[61923]: DEBUG oslo_concurrency.lockutils [None req-85ac5bbf-a6d0-481b-b0bf-1428265b44a9 tempest-AttachInterfacesV270Test-411696101 tempest-AttachInterfacesV270Test-411696101-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 26.351s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 673.395176] env[61923]: DEBUG nova.compute.manager [None req-76010439-9fb6-4296-9b5a-107842db76db tempest-ServerExternalEventsTest-1479796458 tempest-ServerExternalEventsTest-1479796458-project-member] [instance: c0f87aa5-447b-4920-8251-354d89a14fbb] Build of instance c0f87aa5-447b-4920-8251-354d89a14fbb was re-scheduled: Binding failed for port c67bdeb7-ce64-47ec-9e91-62b8f48de195, please check neutron logs for more information. {{(pid=61923) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 673.395801] env[61923]: DEBUG nova.compute.manager [None req-76010439-9fb6-4296-9b5a-107842db76db tempest-ServerExternalEventsTest-1479796458 tempest-ServerExternalEventsTest-1479796458-project-member] [instance: c0f87aa5-447b-4920-8251-354d89a14fbb] Unplugging VIFs for instance {{(pid=61923) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 673.396149] env[61923]: DEBUG oslo_concurrency.lockutils [None req-76010439-9fb6-4296-9b5a-107842db76db tempest-ServerExternalEventsTest-1479796458 tempest-ServerExternalEventsTest-1479796458-project-member] Acquiring lock "refresh_cache-c0f87aa5-447b-4920-8251-354d89a14fbb" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 673.396391] env[61923]: DEBUG oslo_concurrency.lockutils [None req-76010439-9fb6-4296-9b5a-107842db76db tempest-ServerExternalEventsTest-1479796458 tempest-ServerExternalEventsTest-1479796458-project-member] Acquired lock "refresh_cache-c0f87aa5-447b-4920-8251-354d89a14fbb" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 673.396651] env[61923]: DEBUG nova.network.neutron [None req-76010439-9fb6-4296-9b5a-107842db76db tempest-ServerExternalEventsTest-1479796458 tempest-ServerExternalEventsTest-1479796458-project-member] [instance: c0f87aa5-447b-4920-8251-354d89a14fbb] Building network info cache for instance {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 673.923526] env[61923]: DEBUG nova.network.neutron [None req-76010439-9fb6-4296-9b5a-107842db76db tempest-ServerExternalEventsTest-1479796458 tempest-ServerExternalEventsTest-1479796458-project-member] [instance: c0f87aa5-447b-4920-8251-354d89a14fbb] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 674.007427] env[61923]: DEBUG nova.network.neutron [None req-76010439-9fb6-4296-9b5a-107842db76db tempest-ServerExternalEventsTest-1479796458 tempest-ServerExternalEventsTest-1479796458-project-member] [instance: c0f87aa5-447b-4920-8251-354d89a14fbb] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 674.233035] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95c5fe55-ea30-49c8-874b-f2478b1eb92e {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.240502] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d1409c9-fbe8-41f6-a052-1dac9ce9dc3d {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.270714] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb851314-63c9-4a06-a870-c85fc62746f1 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.278007] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2992dfc2-4b17-4c2a-a0c7-cb211653f04b {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.292527] env[61923]: DEBUG nova.compute.provider_tree [None req-85ac5bbf-a6d0-481b-b0bf-1428265b44a9 tempest-AttachInterfacesV270Test-411696101 tempest-AttachInterfacesV270Test-411696101-project-member] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 674.514050] env[61923]: DEBUG oslo_concurrency.lockutils [None req-76010439-9fb6-4296-9b5a-107842db76db tempest-ServerExternalEventsTest-1479796458 tempest-ServerExternalEventsTest-1479796458-project-member] Releasing lock "refresh_cache-c0f87aa5-447b-4920-8251-354d89a14fbb" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 674.514050] env[61923]: DEBUG nova.compute.manager [None req-76010439-9fb6-4296-9b5a-107842db76db tempest-ServerExternalEventsTest-1479796458 tempest-ServerExternalEventsTest-1479796458-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61923) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 674.514050] env[61923]: DEBUG nova.compute.manager [None req-76010439-9fb6-4296-9b5a-107842db76db tempest-ServerExternalEventsTest-1479796458 tempest-ServerExternalEventsTest-1479796458-project-member] [instance: c0f87aa5-447b-4920-8251-354d89a14fbb] Deallocating network for instance {{(pid=61923) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 674.514050] env[61923]: DEBUG nova.network.neutron [None req-76010439-9fb6-4296-9b5a-107842db76db tempest-ServerExternalEventsTest-1479796458 tempest-ServerExternalEventsTest-1479796458-project-member] [instance: c0f87aa5-447b-4920-8251-354d89a14fbb] deallocate_for_instance() {{(pid=61923) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 674.527742] env[61923]: DEBUG nova.network.neutron [None req-76010439-9fb6-4296-9b5a-107842db76db tempest-ServerExternalEventsTest-1479796458 tempest-ServerExternalEventsTest-1479796458-project-member] [instance: c0f87aa5-447b-4920-8251-354d89a14fbb] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 674.795560] env[61923]: DEBUG nova.scheduler.client.report [None req-85ac5bbf-a6d0-481b-b0bf-1428265b44a9 tempest-AttachInterfacesV270Test-411696101 tempest-AttachInterfacesV270Test-411696101-project-member] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 675.030729] env[61923]: DEBUG nova.network.neutron [None req-76010439-9fb6-4296-9b5a-107842db76db tempest-ServerExternalEventsTest-1479796458 tempest-ServerExternalEventsTest-1479796458-project-member] [instance: c0f87aa5-447b-4920-8251-354d89a14fbb] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 675.300871] env[61923]: DEBUG oslo_concurrency.lockutils [None req-85ac5bbf-a6d0-481b-b0bf-1428265b44a9 tempest-AttachInterfacesV270Test-411696101 tempest-AttachInterfacesV270Test-411696101-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.908s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 675.301136] env[61923]: ERROR nova.compute.manager [None req-85ac5bbf-a6d0-481b-b0bf-1428265b44a9 tempest-AttachInterfacesV270Test-411696101 tempest-AttachInterfacesV270Test-411696101-project-member] [instance: 525f37d5-2262-40c8-a339-fc262c53ce6e] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 0aa60830-cb1d-45ad-8800-f2352dd43766, please check neutron logs for more information. [ 675.301136] env[61923]: ERROR nova.compute.manager [instance: 525f37d5-2262-40c8-a339-fc262c53ce6e] Traceback (most recent call last): [ 675.301136] env[61923]: ERROR nova.compute.manager [instance: 525f37d5-2262-40c8-a339-fc262c53ce6e] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 675.301136] env[61923]: ERROR nova.compute.manager [instance: 525f37d5-2262-40c8-a339-fc262c53ce6e] self.driver.spawn(context, instance, image_meta, [ 675.301136] env[61923]: ERROR nova.compute.manager [instance: 525f37d5-2262-40c8-a339-fc262c53ce6e] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 675.301136] env[61923]: ERROR nova.compute.manager [instance: 525f37d5-2262-40c8-a339-fc262c53ce6e] self._vmops.spawn(context, instance, image_meta, injected_files, [ 675.301136] env[61923]: ERROR nova.compute.manager [instance: 525f37d5-2262-40c8-a339-fc262c53ce6e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 675.301136] env[61923]: ERROR nova.compute.manager [instance: 525f37d5-2262-40c8-a339-fc262c53ce6e] vm_ref = self.build_virtual_machine(instance, [ 675.301136] env[61923]: ERROR nova.compute.manager [instance: 525f37d5-2262-40c8-a339-fc262c53ce6e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 675.301136] env[61923]: ERROR nova.compute.manager [instance: 525f37d5-2262-40c8-a339-fc262c53ce6e] vif_infos = vmwarevif.get_vif_info(self._session, [ 675.301136] env[61923]: ERROR nova.compute.manager [instance: 525f37d5-2262-40c8-a339-fc262c53ce6e] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 675.301446] env[61923]: ERROR nova.compute.manager [instance: 525f37d5-2262-40c8-a339-fc262c53ce6e] for vif in network_info: [ 675.301446] env[61923]: ERROR nova.compute.manager [instance: 525f37d5-2262-40c8-a339-fc262c53ce6e] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 675.301446] env[61923]: ERROR nova.compute.manager [instance: 525f37d5-2262-40c8-a339-fc262c53ce6e] return self._sync_wrapper(fn, *args, **kwargs) [ 675.301446] env[61923]: ERROR nova.compute.manager [instance: 525f37d5-2262-40c8-a339-fc262c53ce6e] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 675.301446] env[61923]: ERROR nova.compute.manager [instance: 525f37d5-2262-40c8-a339-fc262c53ce6e] self.wait() [ 675.301446] env[61923]: ERROR nova.compute.manager [instance: 525f37d5-2262-40c8-a339-fc262c53ce6e] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 675.301446] env[61923]: ERROR nova.compute.manager [instance: 525f37d5-2262-40c8-a339-fc262c53ce6e] self[:] = self._gt.wait() [ 675.301446] env[61923]: ERROR nova.compute.manager [instance: 525f37d5-2262-40c8-a339-fc262c53ce6e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 675.301446] env[61923]: ERROR nova.compute.manager [instance: 525f37d5-2262-40c8-a339-fc262c53ce6e] return self._exit_event.wait() [ 675.301446] env[61923]: ERROR nova.compute.manager [instance: 525f37d5-2262-40c8-a339-fc262c53ce6e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 675.301446] env[61923]: ERROR nova.compute.manager [instance: 525f37d5-2262-40c8-a339-fc262c53ce6e] result = hub.switch() [ 675.301446] env[61923]: ERROR nova.compute.manager [instance: 525f37d5-2262-40c8-a339-fc262c53ce6e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 675.301446] env[61923]: ERROR nova.compute.manager [instance: 525f37d5-2262-40c8-a339-fc262c53ce6e] return self.greenlet.switch() [ 675.301795] env[61923]: ERROR nova.compute.manager [instance: 525f37d5-2262-40c8-a339-fc262c53ce6e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 675.301795] env[61923]: ERROR nova.compute.manager [instance: 525f37d5-2262-40c8-a339-fc262c53ce6e] result = function(*args, **kwargs) [ 675.301795] env[61923]: ERROR nova.compute.manager [instance: 525f37d5-2262-40c8-a339-fc262c53ce6e] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 675.301795] env[61923]: ERROR nova.compute.manager [instance: 525f37d5-2262-40c8-a339-fc262c53ce6e] return func(*args, **kwargs) [ 675.301795] env[61923]: ERROR nova.compute.manager [instance: 525f37d5-2262-40c8-a339-fc262c53ce6e] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 675.301795] env[61923]: ERROR nova.compute.manager [instance: 525f37d5-2262-40c8-a339-fc262c53ce6e] raise e [ 675.301795] env[61923]: ERROR nova.compute.manager [instance: 525f37d5-2262-40c8-a339-fc262c53ce6e] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 675.301795] env[61923]: ERROR nova.compute.manager [instance: 525f37d5-2262-40c8-a339-fc262c53ce6e] nwinfo = self.network_api.allocate_for_instance( [ 675.301795] env[61923]: ERROR nova.compute.manager [instance: 525f37d5-2262-40c8-a339-fc262c53ce6e] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 675.301795] env[61923]: ERROR nova.compute.manager [instance: 525f37d5-2262-40c8-a339-fc262c53ce6e] created_port_ids = self._update_ports_for_instance( [ 675.301795] env[61923]: ERROR nova.compute.manager [instance: 525f37d5-2262-40c8-a339-fc262c53ce6e] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 675.301795] env[61923]: ERROR nova.compute.manager [instance: 525f37d5-2262-40c8-a339-fc262c53ce6e] with excutils.save_and_reraise_exception(): [ 675.301795] env[61923]: ERROR nova.compute.manager [instance: 525f37d5-2262-40c8-a339-fc262c53ce6e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 675.302144] env[61923]: ERROR nova.compute.manager [instance: 525f37d5-2262-40c8-a339-fc262c53ce6e] self.force_reraise() [ 675.302144] env[61923]: ERROR nova.compute.manager [instance: 525f37d5-2262-40c8-a339-fc262c53ce6e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 675.302144] env[61923]: ERROR nova.compute.manager [instance: 525f37d5-2262-40c8-a339-fc262c53ce6e] raise self.value [ 675.302144] env[61923]: ERROR nova.compute.manager [instance: 525f37d5-2262-40c8-a339-fc262c53ce6e] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 675.302144] env[61923]: ERROR nova.compute.manager [instance: 525f37d5-2262-40c8-a339-fc262c53ce6e] updated_port = self._update_port( [ 675.302144] env[61923]: ERROR nova.compute.manager [instance: 525f37d5-2262-40c8-a339-fc262c53ce6e] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 675.302144] env[61923]: ERROR nova.compute.manager [instance: 525f37d5-2262-40c8-a339-fc262c53ce6e] _ensure_no_port_binding_failure(port) [ 675.302144] env[61923]: ERROR nova.compute.manager [instance: 525f37d5-2262-40c8-a339-fc262c53ce6e] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 675.302144] env[61923]: ERROR nova.compute.manager [instance: 525f37d5-2262-40c8-a339-fc262c53ce6e] raise exception.PortBindingFailed(port_id=port['id']) [ 675.302144] env[61923]: ERROR nova.compute.manager [instance: 525f37d5-2262-40c8-a339-fc262c53ce6e] nova.exception.PortBindingFailed: Binding failed for port 0aa60830-cb1d-45ad-8800-f2352dd43766, please check neutron logs for more information. [ 675.302144] env[61923]: ERROR nova.compute.manager [instance: 525f37d5-2262-40c8-a339-fc262c53ce6e] [ 675.302411] env[61923]: DEBUG nova.compute.utils [None req-85ac5bbf-a6d0-481b-b0bf-1428265b44a9 tempest-AttachInterfacesV270Test-411696101 tempest-AttachInterfacesV270Test-411696101-project-member] [instance: 525f37d5-2262-40c8-a339-fc262c53ce6e] Binding failed for port 0aa60830-cb1d-45ad-8800-f2352dd43766, please check neutron logs for more information. {{(pid=61923) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 675.303051] env[61923]: DEBUG oslo_concurrency.lockutils [None req-4bd5c12b-d52a-4c54-93f7-87894f09013a tempest-ServersTestJSON-483293076 tempest-ServersTestJSON-483293076-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 26.031s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 675.306100] env[61923]: DEBUG nova.compute.manager [None req-85ac5bbf-a6d0-481b-b0bf-1428265b44a9 tempest-AttachInterfacesV270Test-411696101 tempest-AttachInterfacesV270Test-411696101-project-member] [instance: 525f37d5-2262-40c8-a339-fc262c53ce6e] Build of instance 525f37d5-2262-40c8-a339-fc262c53ce6e was re-scheduled: Binding failed for port 0aa60830-cb1d-45ad-8800-f2352dd43766, please check neutron logs for more information. {{(pid=61923) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 675.306540] env[61923]: DEBUG nova.compute.manager [None req-85ac5bbf-a6d0-481b-b0bf-1428265b44a9 tempest-AttachInterfacesV270Test-411696101 tempest-AttachInterfacesV270Test-411696101-project-member] [instance: 525f37d5-2262-40c8-a339-fc262c53ce6e] Unplugging VIFs for instance {{(pid=61923) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 675.306780] env[61923]: DEBUG oslo_concurrency.lockutils [None req-85ac5bbf-a6d0-481b-b0bf-1428265b44a9 tempest-AttachInterfacesV270Test-411696101 tempest-AttachInterfacesV270Test-411696101-project-member] Acquiring lock "refresh_cache-525f37d5-2262-40c8-a339-fc262c53ce6e" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 675.306925] env[61923]: DEBUG oslo_concurrency.lockutils [None req-85ac5bbf-a6d0-481b-b0bf-1428265b44a9 tempest-AttachInterfacesV270Test-411696101 tempest-AttachInterfacesV270Test-411696101-project-member] Acquired lock "refresh_cache-525f37d5-2262-40c8-a339-fc262c53ce6e" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 675.307096] env[61923]: DEBUG nova.network.neutron [None req-85ac5bbf-a6d0-481b-b0bf-1428265b44a9 tempest-AttachInterfacesV270Test-411696101 tempest-AttachInterfacesV270Test-411696101-project-member] [instance: 525f37d5-2262-40c8-a339-fc262c53ce6e] Building network info cache for instance {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 675.533206] env[61923]: INFO nova.compute.manager [None req-76010439-9fb6-4296-9b5a-107842db76db tempest-ServerExternalEventsTest-1479796458 tempest-ServerExternalEventsTest-1479796458-project-member] [instance: c0f87aa5-447b-4920-8251-354d89a14fbb] Took 1.02 seconds to deallocate network for instance. [ 675.832683] env[61923]: DEBUG nova.network.neutron [None req-85ac5bbf-a6d0-481b-b0bf-1428265b44a9 tempest-AttachInterfacesV270Test-411696101 tempest-AttachInterfacesV270Test-411696101-project-member] [instance: 525f37d5-2262-40c8-a339-fc262c53ce6e] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 675.906143] env[61923]: DEBUG nova.network.neutron [None req-85ac5bbf-a6d0-481b-b0bf-1428265b44a9 tempest-AttachInterfacesV270Test-411696101 tempest-AttachInterfacesV270Test-411696101-project-member] [instance: 525f37d5-2262-40c8-a339-fc262c53ce6e] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 676.118314] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a861f8f-4cb9-43a0-88b9-27b6f26b6d04 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.125983] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-997dfaf1-f338-4264-ad47-437bb2f7f833 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.156125] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c589179-4884-491c-a2b0-146a93e7561d {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.162606] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5780e25-8cd1-47dc-8295-149a03ade18a {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.175276] env[61923]: DEBUG nova.compute.provider_tree [None req-4bd5c12b-d52a-4c54-93f7-87894f09013a tempest-ServersTestJSON-483293076 tempest-ServersTestJSON-483293076-project-member] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 676.414708] env[61923]: DEBUG oslo_concurrency.lockutils [None req-85ac5bbf-a6d0-481b-b0bf-1428265b44a9 tempest-AttachInterfacesV270Test-411696101 tempest-AttachInterfacesV270Test-411696101-project-member] Releasing lock "refresh_cache-525f37d5-2262-40c8-a339-fc262c53ce6e" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 676.414708] env[61923]: DEBUG nova.compute.manager [None req-85ac5bbf-a6d0-481b-b0bf-1428265b44a9 tempest-AttachInterfacesV270Test-411696101 tempest-AttachInterfacesV270Test-411696101-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61923) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 676.414708] env[61923]: DEBUG nova.compute.manager [None req-85ac5bbf-a6d0-481b-b0bf-1428265b44a9 tempest-AttachInterfacesV270Test-411696101 tempest-AttachInterfacesV270Test-411696101-project-member] [instance: 525f37d5-2262-40c8-a339-fc262c53ce6e] Deallocating network for instance {{(pid=61923) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 676.414708] env[61923]: DEBUG nova.network.neutron [None req-85ac5bbf-a6d0-481b-b0bf-1428265b44a9 tempest-AttachInterfacesV270Test-411696101 tempest-AttachInterfacesV270Test-411696101-project-member] [instance: 525f37d5-2262-40c8-a339-fc262c53ce6e] deallocate_for_instance() {{(pid=61923) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 676.435650] env[61923]: DEBUG nova.network.neutron [None req-85ac5bbf-a6d0-481b-b0bf-1428265b44a9 tempest-AttachInterfacesV270Test-411696101 tempest-AttachInterfacesV270Test-411696101-project-member] [instance: 525f37d5-2262-40c8-a339-fc262c53ce6e] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 676.573157] env[61923]: INFO nova.scheduler.client.report [None req-76010439-9fb6-4296-9b5a-107842db76db tempest-ServerExternalEventsTest-1479796458 tempest-ServerExternalEventsTest-1479796458-project-member] Deleted allocations for instance c0f87aa5-447b-4920-8251-354d89a14fbb [ 676.682209] env[61923]: DEBUG nova.scheduler.client.report [None req-4bd5c12b-d52a-4c54-93f7-87894f09013a tempest-ServersTestJSON-483293076 tempest-ServersTestJSON-483293076-project-member] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 676.938175] env[61923]: DEBUG nova.network.neutron [None req-85ac5bbf-a6d0-481b-b0bf-1428265b44a9 tempest-AttachInterfacesV270Test-411696101 tempest-AttachInterfacesV270Test-411696101-project-member] [instance: 525f37d5-2262-40c8-a339-fc262c53ce6e] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 677.080152] env[61923]: DEBUG oslo_concurrency.lockutils [None req-76010439-9fb6-4296-9b5a-107842db76db tempest-ServerExternalEventsTest-1479796458 tempest-ServerExternalEventsTest-1479796458-project-member] Lock "c0f87aa5-447b-4920-8251-354d89a14fbb" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 116.050s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 677.187405] env[61923]: DEBUG oslo_concurrency.lockutils [None req-4bd5c12b-d52a-4c54-93f7-87894f09013a tempest-ServersTestJSON-483293076 tempest-ServersTestJSON-483293076-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.884s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 677.188135] env[61923]: ERROR nova.compute.manager [None req-4bd5c12b-d52a-4c54-93f7-87894f09013a tempest-ServersTestJSON-483293076 tempest-ServersTestJSON-483293076-project-member] [instance: 55b60fc2-6484-4fad-aa40-3e73e020841a] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 445a37d4-005a-49e8-afb5-816754610ddc, please check neutron logs for more information. [ 677.188135] env[61923]: ERROR nova.compute.manager [instance: 55b60fc2-6484-4fad-aa40-3e73e020841a] Traceback (most recent call last): [ 677.188135] env[61923]: ERROR nova.compute.manager [instance: 55b60fc2-6484-4fad-aa40-3e73e020841a] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 677.188135] env[61923]: ERROR nova.compute.manager [instance: 55b60fc2-6484-4fad-aa40-3e73e020841a] self.driver.spawn(context, instance, image_meta, [ 677.188135] env[61923]: ERROR nova.compute.manager [instance: 55b60fc2-6484-4fad-aa40-3e73e020841a] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 677.188135] env[61923]: ERROR nova.compute.manager [instance: 55b60fc2-6484-4fad-aa40-3e73e020841a] self._vmops.spawn(context, instance, image_meta, injected_files, [ 677.188135] env[61923]: ERROR nova.compute.manager [instance: 55b60fc2-6484-4fad-aa40-3e73e020841a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 677.188135] env[61923]: ERROR nova.compute.manager [instance: 55b60fc2-6484-4fad-aa40-3e73e020841a] vm_ref = self.build_virtual_machine(instance, [ 677.188135] env[61923]: ERROR nova.compute.manager [instance: 55b60fc2-6484-4fad-aa40-3e73e020841a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 677.188135] env[61923]: ERROR nova.compute.manager [instance: 55b60fc2-6484-4fad-aa40-3e73e020841a] vif_infos = vmwarevif.get_vif_info(self._session, [ 677.188135] env[61923]: ERROR nova.compute.manager [instance: 55b60fc2-6484-4fad-aa40-3e73e020841a] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 677.188524] env[61923]: ERROR nova.compute.manager [instance: 55b60fc2-6484-4fad-aa40-3e73e020841a] for vif in network_info: [ 677.188524] env[61923]: ERROR nova.compute.manager [instance: 55b60fc2-6484-4fad-aa40-3e73e020841a] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 677.188524] env[61923]: ERROR nova.compute.manager [instance: 55b60fc2-6484-4fad-aa40-3e73e020841a] return self._sync_wrapper(fn, *args, **kwargs) [ 677.188524] env[61923]: ERROR nova.compute.manager [instance: 55b60fc2-6484-4fad-aa40-3e73e020841a] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 677.188524] env[61923]: ERROR nova.compute.manager [instance: 55b60fc2-6484-4fad-aa40-3e73e020841a] self.wait() [ 677.188524] env[61923]: ERROR nova.compute.manager [instance: 55b60fc2-6484-4fad-aa40-3e73e020841a] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 677.188524] env[61923]: ERROR nova.compute.manager [instance: 55b60fc2-6484-4fad-aa40-3e73e020841a] self[:] = self._gt.wait() [ 677.188524] env[61923]: ERROR nova.compute.manager [instance: 55b60fc2-6484-4fad-aa40-3e73e020841a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 677.188524] env[61923]: ERROR nova.compute.manager [instance: 55b60fc2-6484-4fad-aa40-3e73e020841a] return self._exit_event.wait() [ 677.188524] env[61923]: ERROR nova.compute.manager [instance: 55b60fc2-6484-4fad-aa40-3e73e020841a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 677.188524] env[61923]: ERROR nova.compute.manager [instance: 55b60fc2-6484-4fad-aa40-3e73e020841a] result = hub.switch() [ 677.188524] env[61923]: ERROR nova.compute.manager [instance: 55b60fc2-6484-4fad-aa40-3e73e020841a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 677.188524] env[61923]: ERROR nova.compute.manager [instance: 55b60fc2-6484-4fad-aa40-3e73e020841a] return self.greenlet.switch() [ 677.188860] env[61923]: ERROR nova.compute.manager [instance: 55b60fc2-6484-4fad-aa40-3e73e020841a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 677.188860] env[61923]: ERROR nova.compute.manager [instance: 55b60fc2-6484-4fad-aa40-3e73e020841a] result = function(*args, **kwargs) [ 677.188860] env[61923]: ERROR nova.compute.manager [instance: 55b60fc2-6484-4fad-aa40-3e73e020841a] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 677.188860] env[61923]: ERROR nova.compute.manager [instance: 55b60fc2-6484-4fad-aa40-3e73e020841a] return func(*args, **kwargs) [ 677.188860] env[61923]: ERROR nova.compute.manager [instance: 55b60fc2-6484-4fad-aa40-3e73e020841a] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 677.188860] env[61923]: ERROR nova.compute.manager [instance: 55b60fc2-6484-4fad-aa40-3e73e020841a] raise e [ 677.188860] env[61923]: ERROR nova.compute.manager [instance: 55b60fc2-6484-4fad-aa40-3e73e020841a] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 677.188860] env[61923]: ERROR nova.compute.manager [instance: 55b60fc2-6484-4fad-aa40-3e73e020841a] nwinfo = self.network_api.allocate_for_instance( [ 677.188860] env[61923]: ERROR nova.compute.manager [instance: 55b60fc2-6484-4fad-aa40-3e73e020841a] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 677.188860] env[61923]: ERROR nova.compute.manager [instance: 55b60fc2-6484-4fad-aa40-3e73e020841a] created_port_ids = self._update_ports_for_instance( [ 677.188860] env[61923]: ERROR nova.compute.manager [instance: 55b60fc2-6484-4fad-aa40-3e73e020841a] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 677.188860] env[61923]: ERROR nova.compute.manager [instance: 55b60fc2-6484-4fad-aa40-3e73e020841a] with excutils.save_and_reraise_exception(): [ 677.188860] env[61923]: ERROR nova.compute.manager [instance: 55b60fc2-6484-4fad-aa40-3e73e020841a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 677.189192] env[61923]: ERROR nova.compute.manager [instance: 55b60fc2-6484-4fad-aa40-3e73e020841a] self.force_reraise() [ 677.189192] env[61923]: ERROR nova.compute.manager [instance: 55b60fc2-6484-4fad-aa40-3e73e020841a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 677.189192] env[61923]: ERROR nova.compute.manager [instance: 55b60fc2-6484-4fad-aa40-3e73e020841a] raise self.value [ 677.189192] env[61923]: ERROR nova.compute.manager [instance: 55b60fc2-6484-4fad-aa40-3e73e020841a] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 677.189192] env[61923]: ERROR nova.compute.manager [instance: 55b60fc2-6484-4fad-aa40-3e73e020841a] updated_port = self._update_port( [ 677.189192] env[61923]: ERROR nova.compute.manager [instance: 55b60fc2-6484-4fad-aa40-3e73e020841a] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 677.189192] env[61923]: ERROR nova.compute.manager [instance: 55b60fc2-6484-4fad-aa40-3e73e020841a] _ensure_no_port_binding_failure(port) [ 677.189192] env[61923]: ERROR nova.compute.manager [instance: 55b60fc2-6484-4fad-aa40-3e73e020841a] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 677.189192] env[61923]: ERROR nova.compute.manager [instance: 55b60fc2-6484-4fad-aa40-3e73e020841a] raise exception.PortBindingFailed(port_id=port['id']) [ 677.189192] env[61923]: ERROR nova.compute.manager [instance: 55b60fc2-6484-4fad-aa40-3e73e020841a] nova.exception.PortBindingFailed: Binding failed for port 445a37d4-005a-49e8-afb5-816754610ddc, please check neutron logs for more information. [ 677.189192] env[61923]: ERROR nova.compute.manager [instance: 55b60fc2-6484-4fad-aa40-3e73e020841a] [ 677.189529] env[61923]: DEBUG nova.compute.utils [None req-4bd5c12b-d52a-4c54-93f7-87894f09013a tempest-ServersTestJSON-483293076 tempest-ServersTestJSON-483293076-project-member] [instance: 55b60fc2-6484-4fad-aa40-3e73e020841a] Binding failed for port 445a37d4-005a-49e8-afb5-816754610ddc, please check neutron logs for more information. {{(pid=61923) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 677.190172] env[61923]: DEBUG oslo_concurrency.lockutils [None req-1f24f04c-cf4b-4940-9447-eacc11fa5e66 tempest-ServersTestFqdnHostnames-345291565 tempest-ServersTestFqdnHostnames-345291565-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 25.569s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 677.192506] env[61923]: INFO nova.compute.claims [None req-1f24f04c-cf4b-4940-9447-eacc11fa5e66 tempest-ServersTestFqdnHostnames-345291565 tempest-ServersTestFqdnHostnames-345291565-project-member] [instance: c94e8e46-7697-426c-ae2f-aece493fa8f4] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 677.195312] env[61923]: DEBUG nova.compute.manager [None req-4bd5c12b-d52a-4c54-93f7-87894f09013a tempest-ServersTestJSON-483293076 tempest-ServersTestJSON-483293076-project-member] [instance: 55b60fc2-6484-4fad-aa40-3e73e020841a] Build of instance 55b60fc2-6484-4fad-aa40-3e73e020841a was re-scheduled: Binding failed for port 445a37d4-005a-49e8-afb5-816754610ddc, please check neutron logs for more information. {{(pid=61923) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 677.195810] env[61923]: DEBUG nova.compute.manager [None req-4bd5c12b-d52a-4c54-93f7-87894f09013a tempest-ServersTestJSON-483293076 tempest-ServersTestJSON-483293076-project-member] [instance: 55b60fc2-6484-4fad-aa40-3e73e020841a] Unplugging VIFs for instance {{(pid=61923) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 677.196056] env[61923]: DEBUG oslo_concurrency.lockutils [None req-4bd5c12b-d52a-4c54-93f7-87894f09013a tempest-ServersTestJSON-483293076 tempest-ServersTestJSON-483293076-project-member] Acquiring lock "refresh_cache-55b60fc2-6484-4fad-aa40-3e73e020841a" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 677.196230] env[61923]: DEBUG oslo_concurrency.lockutils [None req-4bd5c12b-d52a-4c54-93f7-87894f09013a tempest-ServersTestJSON-483293076 tempest-ServersTestJSON-483293076-project-member] Acquired lock "refresh_cache-55b60fc2-6484-4fad-aa40-3e73e020841a" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 677.196406] env[61923]: DEBUG nova.network.neutron [None req-4bd5c12b-d52a-4c54-93f7-87894f09013a tempest-ServersTestJSON-483293076 tempest-ServersTestJSON-483293076-project-member] [instance: 55b60fc2-6484-4fad-aa40-3e73e020841a] Building network info cache for instance {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 677.441615] env[61923]: INFO nova.compute.manager [None req-85ac5bbf-a6d0-481b-b0bf-1428265b44a9 tempest-AttachInterfacesV270Test-411696101 tempest-AttachInterfacesV270Test-411696101-project-member] [instance: 525f37d5-2262-40c8-a339-fc262c53ce6e] Took 1.03 seconds to deallocate network for instance. [ 677.583468] env[61923]: DEBUG nova.compute.manager [None req-df4eb7d8-5b64-4db4-b042-683b16e0b991 tempest-ServersTestBootFromVolume-2104901628 tempest-ServersTestBootFromVolume-2104901628-project-member] [instance: fb548893-2df5-4629-bf3a-3b0f807f6357] Starting instance... {{(pid=61923) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 677.718220] env[61923]: DEBUG nova.network.neutron [None req-4bd5c12b-d52a-4c54-93f7-87894f09013a tempest-ServersTestJSON-483293076 tempest-ServersTestJSON-483293076-project-member] [instance: 55b60fc2-6484-4fad-aa40-3e73e020841a] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 677.842350] env[61923]: DEBUG nova.network.neutron [None req-4bd5c12b-d52a-4c54-93f7-87894f09013a tempest-ServersTestJSON-483293076 tempest-ServersTestJSON-483293076-project-member] [instance: 55b60fc2-6484-4fad-aa40-3e73e020841a] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 678.106600] env[61923]: DEBUG oslo_concurrency.lockutils [None req-df4eb7d8-5b64-4db4-b042-683b16e0b991 tempest-ServersTestBootFromVolume-2104901628 tempest-ServersTestBootFromVolume-2104901628-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 678.346092] env[61923]: DEBUG oslo_concurrency.lockutils [None req-4bd5c12b-d52a-4c54-93f7-87894f09013a tempest-ServersTestJSON-483293076 tempest-ServersTestJSON-483293076-project-member] Releasing lock "refresh_cache-55b60fc2-6484-4fad-aa40-3e73e020841a" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 678.346342] env[61923]: DEBUG nova.compute.manager [None req-4bd5c12b-d52a-4c54-93f7-87894f09013a tempest-ServersTestJSON-483293076 tempest-ServersTestJSON-483293076-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61923) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 678.346521] env[61923]: DEBUG nova.compute.manager [None req-4bd5c12b-d52a-4c54-93f7-87894f09013a tempest-ServersTestJSON-483293076 tempest-ServersTestJSON-483293076-project-member] [instance: 55b60fc2-6484-4fad-aa40-3e73e020841a] Deallocating network for instance {{(pid=61923) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 678.346684] env[61923]: DEBUG nova.network.neutron [None req-4bd5c12b-d52a-4c54-93f7-87894f09013a tempest-ServersTestJSON-483293076 tempest-ServersTestJSON-483293076-project-member] [instance: 55b60fc2-6484-4fad-aa40-3e73e020841a] deallocate_for_instance() {{(pid=61923) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 678.369636] env[61923]: DEBUG nova.network.neutron [None req-4bd5c12b-d52a-4c54-93f7-87894f09013a tempest-ServersTestJSON-483293076 tempest-ServersTestJSON-483293076-project-member] [instance: 55b60fc2-6484-4fad-aa40-3e73e020841a] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 678.473082] env[61923]: INFO nova.scheduler.client.report [None req-85ac5bbf-a6d0-481b-b0bf-1428265b44a9 tempest-AttachInterfacesV270Test-411696101 tempest-AttachInterfacesV270Test-411696101-project-member] Deleted allocations for instance 525f37d5-2262-40c8-a339-fc262c53ce6e [ 678.526033] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1775de10-65ad-4bfb-83ee-b9348ab17993 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.535395] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9c2e4de-d279-48cd-8a61-7b92ce6efc90 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.568535] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d150eda4-89d5-44d6-8b0b-065566b895ec {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.576063] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea08e53b-a6e2-4a8e-be86-10d56fd3b738 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.590371] env[61923]: DEBUG nova.compute.provider_tree [None req-1f24f04c-cf4b-4940-9447-eacc11fa5e66 tempest-ServersTestFqdnHostnames-345291565 tempest-ServersTestFqdnHostnames-345291565-project-member] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 678.872095] env[61923]: DEBUG nova.network.neutron [None req-4bd5c12b-d52a-4c54-93f7-87894f09013a tempest-ServersTestJSON-483293076 tempest-ServersTestJSON-483293076-project-member] [instance: 55b60fc2-6484-4fad-aa40-3e73e020841a] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 678.985971] env[61923]: DEBUG oslo_concurrency.lockutils [None req-85ac5bbf-a6d0-481b-b0bf-1428265b44a9 tempest-AttachInterfacesV270Test-411696101 tempest-AttachInterfacesV270Test-411696101-project-member] Lock "525f37d5-2262-40c8-a339-fc262c53ce6e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 108.582s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 679.095584] env[61923]: DEBUG nova.scheduler.client.report [None req-1f24f04c-cf4b-4940-9447-eacc11fa5e66 tempest-ServersTestFqdnHostnames-345291565 tempest-ServersTestFqdnHostnames-345291565-project-member] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 679.375276] env[61923]: INFO nova.compute.manager [None req-4bd5c12b-d52a-4c54-93f7-87894f09013a tempest-ServersTestJSON-483293076 tempest-ServersTestJSON-483293076-project-member] [instance: 55b60fc2-6484-4fad-aa40-3e73e020841a] Took 1.03 seconds to deallocate network for instance. [ 679.488540] env[61923]: DEBUG nova.compute.manager [None req-44a76586-0f76-448f-af41-a5d846bfda81 tempest-ServerActionsV293TestJSON-849444864 tempest-ServerActionsV293TestJSON-849444864-project-member] [instance: 62144317-0354-4751-8f2b-4eae588c9da7] Starting instance... {{(pid=61923) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 679.600957] env[61923]: DEBUG oslo_concurrency.lockutils [None req-1f24f04c-cf4b-4940-9447-eacc11fa5e66 tempest-ServersTestFqdnHostnames-345291565 tempest-ServersTestFqdnHostnames-345291565-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.411s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 679.601499] env[61923]: DEBUG nova.compute.manager [None req-1f24f04c-cf4b-4940-9447-eacc11fa5e66 tempest-ServersTestFqdnHostnames-345291565 tempest-ServersTestFqdnHostnames-345291565-project-member] [instance: c94e8e46-7697-426c-ae2f-aece493fa8f4] Start building networks asynchronously for instance. {{(pid=61923) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 679.604585] env[61923]: DEBUG oslo_concurrency.lockutils [None req-aadda1bb-f770-4742-8c34-cd25e3f7e7cb tempest-InstanceActionsNegativeTestJSON-1700332362 tempest-InstanceActionsNegativeTestJSON-1700332362-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 26.167s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 679.608081] env[61923]: INFO nova.compute.claims [None req-aadda1bb-f770-4742-8c34-cd25e3f7e7cb tempest-InstanceActionsNegativeTestJSON-1700332362 tempest-InstanceActionsNegativeTestJSON-1700332362-project-member] [instance: 64584976-b3f3-4da5-a76b-a05cf2ed6aa4] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 680.012721] env[61923]: DEBUG oslo_concurrency.lockutils [None req-44a76586-0f76-448f-af41-a5d846bfda81 tempest-ServerActionsV293TestJSON-849444864 tempest-ServerActionsV293TestJSON-849444864-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 680.109947] env[61923]: DEBUG nova.compute.utils [None req-1f24f04c-cf4b-4940-9447-eacc11fa5e66 tempest-ServersTestFqdnHostnames-345291565 tempest-ServersTestFqdnHostnames-345291565-project-member] Using /dev/sd instead of None {{(pid=61923) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 680.111512] env[61923]: DEBUG nova.compute.manager [None req-1f24f04c-cf4b-4940-9447-eacc11fa5e66 tempest-ServersTestFqdnHostnames-345291565 tempest-ServersTestFqdnHostnames-345291565-project-member] [instance: c94e8e46-7697-426c-ae2f-aece493fa8f4] Allocating IP information in the background. {{(pid=61923) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 680.111681] env[61923]: DEBUG nova.network.neutron [None req-1f24f04c-cf4b-4940-9447-eacc11fa5e66 tempest-ServersTestFqdnHostnames-345291565 tempest-ServersTestFqdnHostnames-345291565-project-member] [instance: c94e8e46-7697-426c-ae2f-aece493fa8f4] allocate_for_instance() {{(pid=61923) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 680.163267] env[61923]: DEBUG nova.policy [None req-1f24f04c-cf4b-4940-9447-eacc11fa5e66 tempest-ServersTestFqdnHostnames-345291565 tempest-ServersTestFqdnHostnames-345291565-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c09f78f4aff34413b7e9c569ef62a242', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '9e83b468193841dbbd74b035bf7692ae', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61923) authorize /opt/stack/nova/nova/policy.py:201}} [ 680.411013] env[61923]: INFO nova.scheduler.client.report [None req-4bd5c12b-d52a-4c54-93f7-87894f09013a tempest-ServersTestJSON-483293076 tempest-ServersTestJSON-483293076-project-member] Deleted allocations for instance 55b60fc2-6484-4fad-aa40-3e73e020841a [ 680.623743] env[61923]: DEBUG nova.compute.manager [None req-1f24f04c-cf4b-4940-9447-eacc11fa5e66 tempest-ServersTestFqdnHostnames-345291565 tempest-ServersTestFqdnHostnames-345291565-project-member] [instance: c94e8e46-7697-426c-ae2f-aece493fa8f4] Start building block device mappings for instance. {{(pid=61923) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 680.743484] env[61923]: DEBUG nova.network.neutron [None req-1f24f04c-cf4b-4940-9447-eacc11fa5e66 tempest-ServersTestFqdnHostnames-345291565 tempest-ServersTestFqdnHostnames-345291565-project-member] [instance: c94e8e46-7697-426c-ae2f-aece493fa8f4] Successfully created port: 70370c2b-9eb9-48b7-8acc-8a7eafc1d41d {{(pid=61923) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 680.919269] env[61923]: DEBUG oslo_concurrency.lockutils [None req-4bd5c12b-d52a-4c54-93f7-87894f09013a tempest-ServersTestJSON-483293076 tempest-ServersTestJSON-483293076-project-member] Lock "55b60fc2-6484-4fad-aa40-3e73e020841a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 107.932s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 681.015106] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86257009-de11-4768-bacd-b288fc94e8f5 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.024640] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a2e1a89-5d8a-4d46-8bd1-6aa98decdf5a {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.057760] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0c3edf2-858d-4b60-9b17-f58701e77458 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.065842] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97e5818a-0676-4b7e-82d6-80ba157b1c20 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.080239] env[61923]: DEBUG nova.compute.provider_tree [None req-aadda1bb-f770-4742-8c34-cd25e3f7e7cb tempest-InstanceActionsNegativeTestJSON-1700332362 tempest-InstanceActionsNegativeTestJSON-1700332362-project-member] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 681.424673] env[61923]: DEBUG nova.compute.manager [None req-c09a27e9-8e33-411f-b17b-7cd8d66a175c tempest-ServerMetadataNegativeTestJSON-1976245157 tempest-ServerMetadataNegativeTestJSON-1976245157-project-member] [instance: 25b62db6-6e64-4e38-bfd3-82a2e9a4883a] Starting instance... {{(pid=61923) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 681.583884] env[61923]: DEBUG nova.scheduler.client.report [None req-aadda1bb-f770-4742-8c34-cd25e3f7e7cb tempest-InstanceActionsNegativeTestJSON-1700332362 tempest-InstanceActionsNegativeTestJSON-1700332362-project-member] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 681.636631] env[61923]: DEBUG nova.compute.manager [None req-1f24f04c-cf4b-4940-9447-eacc11fa5e66 tempest-ServersTestFqdnHostnames-345291565 tempest-ServersTestFqdnHostnames-345291565-project-member] [instance: c94e8e46-7697-426c-ae2f-aece493fa8f4] Start spawning the instance on the hypervisor. {{(pid=61923) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 681.678816] env[61923]: DEBUG nova.virt.hardware [None req-1f24f04c-cf4b-4940-9447-eacc11fa5e66 tempest-ServersTestFqdnHostnames-345291565 tempest-ServersTestFqdnHostnames-345291565-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-29T20:07:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-29T20:07:35Z,direct_url=,disk_format='vmdk',id=0f153f63-ae0a-45b1-b7f5-7f9b673c947f,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='4e7b5e3b3c3443c8bd5c70f8a15739f5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-29T20:07:36Z,virtual_size=,visibility=), allow threads: False {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 681.679117] env[61923]: DEBUG nova.virt.hardware [None req-1f24f04c-cf4b-4940-9447-eacc11fa5e66 tempest-ServersTestFqdnHostnames-345291565 tempest-ServersTestFqdnHostnames-345291565-project-member] Flavor limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 681.682124] env[61923]: DEBUG nova.virt.hardware [None req-1f24f04c-cf4b-4940-9447-eacc11fa5e66 tempest-ServersTestFqdnHostnames-345291565 tempest-ServersTestFqdnHostnames-345291565-project-member] Image limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 681.682434] env[61923]: DEBUG nova.virt.hardware [None req-1f24f04c-cf4b-4940-9447-eacc11fa5e66 tempest-ServersTestFqdnHostnames-345291565 tempest-ServersTestFqdnHostnames-345291565-project-member] Flavor pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 681.682655] env[61923]: DEBUG nova.virt.hardware [None req-1f24f04c-cf4b-4940-9447-eacc11fa5e66 tempest-ServersTestFqdnHostnames-345291565 tempest-ServersTestFqdnHostnames-345291565-project-member] Image pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 681.682742] env[61923]: DEBUG nova.virt.hardware [None req-1f24f04c-cf4b-4940-9447-eacc11fa5e66 tempest-ServersTestFqdnHostnames-345291565 tempest-ServersTestFqdnHostnames-345291565-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 681.682963] env[61923]: DEBUG nova.virt.hardware [None req-1f24f04c-cf4b-4940-9447-eacc11fa5e66 tempest-ServersTestFqdnHostnames-345291565 tempest-ServersTestFqdnHostnames-345291565-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 681.683184] env[61923]: DEBUG nova.virt.hardware [None req-1f24f04c-cf4b-4940-9447-eacc11fa5e66 tempest-ServersTestFqdnHostnames-345291565 tempest-ServersTestFqdnHostnames-345291565-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 681.683373] env[61923]: DEBUG nova.virt.hardware [None req-1f24f04c-cf4b-4940-9447-eacc11fa5e66 tempest-ServersTestFqdnHostnames-345291565 tempest-ServersTestFqdnHostnames-345291565-project-member] Got 1 possible topologies {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 681.683544] env[61923]: DEBUG nova.virt.hardware [None req-1f24f04c-cf4b-4940-9447-eacc11fa5e66 tempest-ServersTestFqdnHostnames-345291565 tempest-ServersTestFqdnHostnames-345291565-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 681.684119] env[61923]: DEBUG nova.virt.hardware [None req-1f24f04c-cf4b-4940-9447-eacc11fa5e66 tempest-ServersTestFqdnHostnames-345291565 tempest-ServersTestFqdnHostnames-345291565-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 681.684673] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04372c40-c62a-4806-bd78-f3f68c550c3d {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.697767] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef6a5c33-38f8-4bd2-8657-795225fa17c7 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 681.963707] env[61923]: DEBUG oslo_concurrency.lockutils [None req-c09a27e9-8e33-411f-b17b-7cd8d66a175c tempest-ServerMetadataNegativeTestJSON-1976245157 tempest-ServerMetadataNegativeTestJSON-1976245157-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 682.126802] env[61923]: DEBUG oslo_concurrency.lockutils [None req-aadda1bb-f770-4742-8c34-cd25e3f7e7cb tempest-InstanceActionsNegativeTestJSON-1700332362 tempest-InstanceActionsNegativeTestJSON-1700332362-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.522s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 682.126943] env[61923]: DEBUG nova.compute.manager [None req-aadda1bb-f770-4742-8c34-cd25e3f7e7cb tempest-InstanceActionsNegativeTestJSON-1700332362 tempest-InstanceActionsNegativeTestJSON-1700332362-project-member] [instance: 64584976-b3f3-4da5-a76b-a05cf2ed6aa4] Start building networks asynchronously for instance. {{(pid=61923) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 682.131555] env[61923]: DEBUG nova.compute.manager [req-c90cfa8d-0181-49be-b838-bca4c8e3c7d6 req-6e4acef3-c9db-434a-ba9c-6bc0ebf66375 service nova] [instance: c94e8e46-7697-426c-ae2f-aece493fa8f4] Received event network-changed-70370c2b-9eb9-48b7-8acc-8a7eafc1d41d {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 682.131555] env[61923]: DEBUG nova.compute.manager [req-c90cfa8d-0181-49be-b838-bca4c8e3c7d6 req-6e4acef3-c9db-434a-ba9c-6bc0ebf66375 service nova] [instance: c94e8e46-7697-426c-ae2f-aece493fa8f4] Refreshing instance network info cache due to event network-changed-70370c2b-9eb9-48b7-8acc-8a7eafc1d41d. {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 682.131555] env[61923]: DEBUG oslo_concurrency.lockutils [req-c90cfa8d-0181-49be-b838-bca4c8e3c7d6 req-6e4acef3-c9db-434a-ba9c-6bc0ebf66375 service nova] Acquiring lock "refresh_cache-c94e8e46-7697-426c-ae2f-aece493fa8f4" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 682.131555] env[61923]: DEBUG oslo_concurrency.lockutils [req-c90cfa8d-0181-49be-b838-bca4c8e3c7d6 req-6e4acef3-c9db-434a-ba9c-6bc0ebf66375 service nova] Acquired lock "refresh_cache-c94e8e46-7697-426c-ae2f-aece493fa8f4" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 682.131555] env[61923]: DEBUG nova.network.neutron [req-c90cfa8d-0181-49be-b838-bca4c8e3c7d6 req-6e4acef3-c9db-434a-ba9c-6bc0ebf66375 service nova] [instance: c94e8e46-7697-426c-ae2f-aece493fa8f4] Refreshing network info cache for port 70370c2b-9eb9-48b7-8acc-8a7eafc1d41d {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 682.132240] env[61923]: DEBUG oslo_concurrency.lockutils [None req-56950b6c-0b87-48bf-a575-f699fcd1a5b0 tempest-FloatingIPsAssociationTestJSON-1125124803 tempest-FloatingIPsAssociationTestJSON-1125124803-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 25.420s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 682.383630] env[61923]: ERROR nova.compute.manager [None req-1f24f04c-cf4b-4940-9447-eacc11fa5e66 tempest-ServersTestFqdnHostnames-345291565 tempest-ServersTestFqdnHostnames-345291565-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 70370c2b-9eb9-48b7-8acc-8a7eafc1d41d, please check neutron logs for more information. [ 682.383630] env[61923]: ERROR nova.compute.manager Traceback (most recent call last): [ 682.383630] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 682.383630] env[61923]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 682.383630] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 682.383630] env[61923]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 682.383630] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 682.383630] env[61923]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 682.383630] env[61923]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 682.383630] env[61923]: ERROR nova.compute.manager self.force_reraise() [ 682.383630] env[61923]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 682.383630] env[61923]: ERROR nova.compute.manager raise self.value [ 682.383630] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 682.383630] env[61923]: ERROR nova.compute.manager updated_port = self._update_port( [ 682.383630] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 682.383630] env[61923]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 682.384236] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 682.384236] env[61923]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 682.384236] env[61923]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 70370c2b-9eb9-48b7-8acc-8a7eafc1d41d, please check neutron logs for more information. [ 682.384236] env[61923]: ERROR nova.compute.manager [ 682.384236] env[61923]: Traceback (most recent call last): [ 682.384236] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 682.384236] env[61923]: listener.cb(fileno) [ 682.384236] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 682.384236] env[61923]: result = function(*args, **kwargs) [ 682.384236] env[61923]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 682.384236] env[61923]: return func(*args, **kwargs) [ 682.384236] env[61923]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 682.384236] env[61923]: raise e [ 682.384236] env[61923]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 682.384236] env[61923]: nwinfo = self.network_api.allocate_for_instance( [ 682.384236] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 682.384236] env[61923]: created_port_ids = self._update_ports_for_instance( [ 682.384236] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 682.384236] env[61923]: with excutils.save_and_reraise_exception(): [ 682.384236] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 682.384236] env[61923]: self.force_reraise() [ 682.384236] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 682.384236] env[61923]: raise self.value [ 682.384236] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 682.384236] env[61923]: updated_port = self._update_port( [ 682.384236] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 682.384236] env[61923]: _ensure_no_port_binding_failure(port) [ 682.384236] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 682.384236] env[61923]: raise exception.PortBindingFailed(port_id=port['id']) [ 682.385065] env[61923]: nova.exception.PortBindingFailed: Binding failed for port 70370c2b-9eb9-48b7-8acc-8a7eafc1d41d, please check neutron logs for more information. [ 682.385065] env[61923]: Removing descriptor: 14 [ 682.385065] env[61923]: ERROR nova.compute.manager [None req-1f24f04c-cf4b-4940-9447-eacc11fa5e66 tempest-ServersTestFqdnHostnames-345291565 tempest-ServersTestFqdnHostnames-345291565-project-member] [instance: c94e8e46-7697-426c-ae2f-aece493fa8f4] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 70370c2b-9eb9-48b7-8acc-8a7eafc1d41d, please check neutron logs for more information. [ 682.385065] env[61923]: ERROR nova.compute.manager [instance: c94e8e46-7697-426c-ae2f-aece493fa8f4] Traceback (most recent call last): [ 682.385065] env[61923]: ERROR nova.compute.manager [instance: c94e8e46-7697-426c-ae2f-aece493fa8f4] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 682.385065] env[61923]: ERROR nova.compute.manager [instance: c94e8e46-7697-426c-ae2f-aece493fa8f4] yield resources [ 682.385065] env[61923]: ERROR nova.compute.manager [instance: c94e8e46-7697-426c-ae2f-aece493fa8f4] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 682.385065] env[61923]: ERROR nova.compute.manager [instance: c94e8e46-7697-426c-ae2f-aece493fa8f4] self.driver.spawn(context, instance, image_meta, [ 682.385065] env[61923]: ERROR nova.compute.manager [instance: c94e8e46-7697-426c-ae2f-aece493fa8f4] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 682.385065] env[61923]: ERROR nova.compute.manager [instance: c94e8e46-7697-426c-ae2f-aece493fa8f4] self._vmops.spawn(context, instance, image_meta, injected_files, [ 682.385065] env[61923]: ERROR nova.compute.manager [instance: c94e8e46-7697-426c-ae2f-aece493fa8f4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 682.385065] env[61923]: ERROR nova.compute.manager [instance: c94e8e46-7697-426c-ae2f-aece493fa8f4] vm_ref = self.build_virtual_machine(instance, [ 682.385546] env[61923]: ERROR nova.compute.manager [instance: c94e8e46-7697-426c-ae2f-aece493fa8f4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 682.385546] env[61923]: ERROR nova.compute.manager [instance: c94e8e46-7697-426c-ae2f-aece493fa8f4] vif_infos = vmwarevif.get_vif_info(self._session, [ 682.385546] env[61923]: ERROR nova.compute.manager [instance: c94e8e46-7697-426c-ae2f-aece493fa8f4] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 682.385546] env[61923]: ERROR nova.compute.manager [instance: c94e8e46-7697-426c-ae2f-aece493fa8f4] for vif in network_info: [ 682.385546] env[61923]: ERROR nova.compute.manager [instance: c94e8e46-7697-426c-ae2f-aece493fa8f4] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 682.385546] env[61923]: ERROR nova.compute.manager [instance: c94e8e46-7697-426c-ae2f-aece493fa8f4] return self._sync_wrapper(fn, *args, **kwargs) [ 682.385546] env[61923]: ERROR nova.compute.manager [instance: c94e8e46-7697-426c-ae2f-aece493fa8f4] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 682.385546] env[61923]: ERROR nova.compute.manager [instance: c94e8e46-7697-426c-ae2f-aece493fa8f4] self.wait() [ 682.385546] env[61923]: ERROR nova.compute.manager [instance: c94e8e46-7697-426c-ae2f-aece493fa8f4] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 682.385546] env[61923]: ERROR nova.compute.manager [instance: c94e8e46-7697-426c-ae2f-aece493fa8f4] self[:] = self._gt.wait() [ 682.385546] env[61923]: ERROR nova.compute.manager [instance: c94e8e46-7697-426c-ae2f-aece493fa8f4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 682.385546] env[61923]: ERROR nova.compute.manager [instance: c94e8e46-7697-426c-ae2f-aece493fa8f4] return self._exit_event.wait() [ 682.385546] env[61923]: ERROR nova.compute.manager [instance: c94e8e46-7697-426c-ae2f-aece493fa8f4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 682.385971] env[61923]: ERROR nova.compute.manager [instance: c94e8e46-7697-426c-ae2f-aece493fa8f4] result = hub.switch() [ 682.385971] env[61923]: ERROR nova.compute.manager [instance: c94e8e46-7697-426c-ae2f-aece493fa8f4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 682.385971] env[61923]: ERROR nova.compute.manager [instance: c94e8e46-7697-426c-ae2f-aece493fa8f4] return self.greenlet.switch() [ 682.385971] env[61923]: ERROR nova.compute.manager [instance: c94e8e46-7697-426c-ae2f-aece493fa8f4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 682.385971] env[61923]: ERROR nova.compute.manager [instance: c94e8e46-7697-426c-ae2f-aece493fa8f4] result = function(*args, **kwargs) [ 682.385971] env[61923]: ERROR nova.compute.manager [instance: c94e8e46-7697-426c-ae2f-aece493fa8f4] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 682.385971] env[61923]: ERROR nova.compute.manager [instance: c94e8e46-7697-426c-ae2f-aece493fa8f4] return func(*args, **kwargs) [ 682.385971] env[61923]: ERROR nova.compute.manager [instance: c94e8e46-7697-426c-ae2f-aece493fa8f4] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 682.385971] env[61923]: ERROR nova.compute.manager [instance: c94e8e46-7697-426c-ae2f-aece493fa8f4] raise e [ 682.385971] env[61923]: ERROR nova.compute.manager [instance: c94e8e46-7697-426c-ae2f-aece493fa8f4] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 682.385971] env[61923]: ERROR nova.compute.manager [instance: c94e8e46-7697-426c-ae2f-aece493fa8f4] nwinfo = self.network_api.allocate_for_instance( [ 682.385971] env[61923]: ERROR nova.compute.manager [instance: c94e8e46-7697-426c-ae2f-aece493fa8f4] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 682.385971] env[61923]: ERROR nova.compute.manager [instance: c94e8e46-7697-426c-ae2f-aece493fa8f4] created_port_ids = self._update_ports_for_instance( [ 682.386351] env[61923]: ERROR nova.compute.manager [instance: c94e8e46-7697-426c-ae2f-aece493fa8f4] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 682.386351] env[61923]: ERROR nova.compute.manager [instance: c94e8e46-7697-426c-ae2f-aece493fa8f4] with excutils.save_and_reraise_exception(): [ 682.386351] env[61923]: ERROR nova.compute.manager [instance: c94e8e46-7697-426c-ae2f-aece493fa8f4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 682.386351] env[61923]: ERROR nova.compute.manager [instance: c94e8e46-7697-426c-ae2f-aece493fa8f4] self.force_reraise() [ 682.386351] env[61923]: ERROR nova.compute.manager [instance: c94e8e46-7697-426c-ae2f-aece493fa8f4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 682.386351] env[61923]: ERROR nova.compute.manager [instance: c94e8e46-7697-426c-ae2f-aece493fa8f4] raise self.value [ 682.386351] env[61923]: ERROR nova.compute.manager [instance: c94e8e46-7697-426c-ae2f-aece493fa8f4] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 682.386351] env[61923]: ERROR nova.compute.manager [instance: c94e8e46-7697-426c-ae2f-aece493fa8f4] updated_port = self._update_port( [ 682.386351] env[61923]: ERROR nova.compute.manager [instance: c94e8e46-7697-426c-ae2f-aece493fa8f4] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 682.386351] env[61923]: ERROR nova.compute.manager [instance: c94e8e46-7697-426c-ae2f-aece493fa8f4] _ensure_no_port_binding_failure(port) [ 682.386351] env[61923]: ERROR nova.compute.manager [instance: c94e8e46-7697-426c-ae2f-aece493fa8f4] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 682.386351] env[61923]: ERROR nova.compute.manager [instance: c94e8e46-7697-426c-ae2f-aece493fa8f4] raise exception.PortBindingFailed(port_id=port['id']) [ 682.388715] env[61923]: ERROR nova.compute.manager [instance: c94e8e46-7697-426c-ae2f-aece493fa8f4] nova.exception.PortBindingFailed: Binding failed for port 70370c2b-9eb9-48b7-8acc-8a7eafc1d41d, please check neutron logs for more information. [ 682.388715] env[61923]: ERROR nova.compute.manager [instance: c94e8e46-7697-426c-ae2f-aece493fa8f4] [ 682.388715] env[61923]: INFO nova.compute.manager [None req-1f24f04c-cf4b-4940-9447-eacc11fa5e66 tempest-ServersTestFqdnHostnames-345291565 tempest-ServersTestFqdnHostnames-345291565-project-member] [instance: c94e8e46-7697-426c-ae2f-aece493fa8f4] Terminating instance [ 682.389425] env[61923]: DEBUG oslo_concurrency.lockutils [None req-1f24f04c-cf4b-4940-9447-eacc11fa5e66 tempest-ServersTestFqdnHostnames-345291565 tempest-ServersTestFqdnHostnames-345291565-project-member] Acquiring lock "refresh_cache-c94e8e46-7697-426c-ae2f-aece493fa8f4" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 682.637150] env[61923]: DEBUG nova.compute.utils [None req-aadda1bb-f770-4742-8c34-cd25e3f7e7cb tempest-InstanceActionsNegativeTestJSON-1700332362 tempest-InstanceActionsNegativeTestJSON-1700332362-project-member] Using /dev/sd instead of None {{(pid=61923) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 682.637150] env[61923]: DEBUG nova.compute.manager [None req-aadda1bb-f770-4742-8c34-cd25e3f7e7cb tempest-InstanceActionsNegativeTestJSON-1700332362 tempest-InstanceActionsNegativeTestJSON-1700332362-project-member] [instance: 64584976-b3f3-4da5-a76b-a05cf2ed6aa4] Allocating IP information in the background. {{(pid=61923) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 682.637150] env[61923]: DEBUG nova.network.neutron [None req-aadda1bb-f770-4742-8c34-cd25e3f7e7cb tempest-InstanceActionsNegativeTestJSON-1700332362 tempest-InstanceActionsNegativeTestJSON-1700332362-project-member] [instance: 64584976-b3f3-4da5-a76b-a05cf2ed6aa4] allocate_for_instance() {{(pid=61923) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 682.684517] env[61923]: DEBUG nova.network.neutron [req-c90cfa8d-0181-49be-b838-bca4c8e3c7d6 req-6e4acef3-c9db-434a-ba9c-6bc0ebf66375 service nova] [instance: c94e8e46-7697-426c-ae2f-aece493fa8f4] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 682.757287] env[61923]: DEBUG nova.policy [None req-aadda1bb-f770-4742-8c34-cd25e3f7e7cb tempest-InstanceActionsNegativeTestJSON-1700332362 tempest-InstanceActionsNegativeTestJSON-1700332362-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6436ca01567648ca946eae20d23e5252', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '6c8ba1f1da4a4849bfa0fb209183ac44', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61923) authorize /opt/stack/nova/nova/policy.py:201}} [ 682.945192] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b82a89d-47e6-4516-9ef5-7e086a6ddb39 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 682.948169] env[61923]: DEBUG nova.network.neutron [req-c90cfa8d-0181-49be-b838-bca4c8e3c7d6 req-6e4acef3-c9db-434a-ba9c-6bc0ebf66375 service nova] [instance: c94e8e46-7697-426c-ae2f-aece493fa8f4] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 682.954578] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b6b8066-c837-4b2f-961f-803b1b52e83e {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 682.989063] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7d0c796-64c4-4158-bd43-a4956c63a182 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 682.998346] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c82d8ac-fa9e-46bb-8f22-412d47b6e020 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 683.013373] env[61923]: DEBUG nova.compute.provider_tree [None req-56950b6c-0b87-48bf-a575-f699fcd1a5b0 tempest-FloatingIPsAssociationTestJSON-1125124803 tempest-FloatingIPsAssociationTestJSON-1125124803-project-member] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 683.149278] env[61923]: DEBUG nova.compute.manager [None req-aadda1bb-f770-4742-8c34-cd25e3f7e7cb tempest-InstanceActionsNegativeTestJSON-1700332362 tempest-InstanceActionsNegativeTestJSON-1700332362-project-member] [instance: 64584976-b3f3-4da5-a76b-a05cf2ed6aa4] Start building block device mappings for instance. {{(pid=61923) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 683.453698] env[61923]: DEBUG oslo_concurrency.lockutils [req-c90cfa8d-0181-49be-b838-bca4c8e3c7d6 req-6e4acef3-c9db-434a-ba9c-6bc0ebf66375 service nova] Releasing lock "refresh_cache-c94e8e46-7697-426c-ae2f-aece493fa8f4" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 683.454495] env[61923]: DEBUG oslo_concurrency.lockutils [None req-1f24f04c-cf4b-4940-9447-eacc11fa5e66 tempest-ServersTestFqdnHostnames-345291565 tempest-ServersTestFqdnHostnames-345291565-project-member] Acquired lock "refresh_cache-c94e8e46-7697-426c-ae2f-aece493fa8f4" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 683.454676] env[61923]: DEBUG nova.network.neutron [None req-1f24f04c-cf4b-4940-9447-eacc11fa5e66 tempest-ServersTestFqdnHostnames-345291565 tempest-ServersTestFqdnHostnames-345291565-project-member] [instance: c94e8e46-7697-426c-ae2f-aece493fa8f4] Building network info cache for instance {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 683.520140] env[61923]: DEBUG nova.scheduler.client.report [None req-56950b6c-0b87-48bf-a575-f699fcd1a5b0 tempest-FloatingIPsAssociationTestJSON-1125124803 tempest-FloatingIPsAssociationTestJSON-1125124803-project-member] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 683.932295] env[61923]: DEBUG nova.network.neutron [None req-aadda1bb-f770-4742-8c34-cd25e3f7e7cb tempest-InstanceActionsNegativeTestJSON-1700332362 tempest-InstanceActionsNegativeTestJSON-1700332362-project-member] [instance: 64584976-b3f3-4da5-a76b-a05cf2ed6aa4] Successfully created port: 13363b75-27bb-4922-b5da-febd2dc08e25 {{(pid=61923) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 683.991870] env[61923]: DEBUG nova.network.neutron [None req-1f24f04c-cf4b-4940-9447-eacc11fa5e66 tempest-ServersTestFqdnHostnames-345291565 tempest-ServersTestFqdnHostnames-345291565-project-member] [instance: c94e8e46-7697-426c-ae2f-aece493fa8f4] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 684.026134] env[61923]: DEBUG oslo_concurrency.lockutils [None req-56950b6c-0b87-48bf-a575-f699fcd1a5b0 tempest-FloatingIPsAssociationTestJSON-1125124803 tempest-FloatingIPsAssociationTestJSON-1125124803-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.891s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 684.026134] env[61923]: ERROR nova.compute.manager [None req-56950b6c-0b87-48bf-a575-f699fcd1a5b0 tempest-FloatingIPsAssociationTestJSON-1125124803 tempest-FloatingIPsAssociationTestJSON-1125124803-project-member] [instance: aeca6a94-478d-435a-9d93-59936f26654e] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 61e9a5e4-10ed-40b2-a961-2d5cbae80f02, please check neutron logs for more information. [ 684.026134] env[61923]: ERROR nova.compute.manager [instance: aeca6a94-478d-435a-9d93-59936f26654e] Traceback (most recent call last): [ 684.026134] env[61923]: ERROR nova.compute.manager [instance: aeca6a94-478d-435a-9d93-59936f26654e] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 684.026134] env[61923]: ERROR nova.compute.manager [instance: aeca6a94-478d-435a-9d93-59936f26654e] self.driver.spawn(context, instance, image_meta, [ 684.026134] env[61923]: ERROR nova.compute.manager [instance: aeca6a94-478d-435a-9d93-59936f26654e] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 684.026134] env[61923]: ERROR nova.compute.manager [instance: aeca6a94-478d-435a-9d93-59936f26654e] self._vmops.spawn(context, instance, image_meta, injected_files, [ 684.026134] env[61923]: ERROR nova.compute.manager [instance: aeca6a94-478d-435a-9d93-59936f26654e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 684.026134] env[61923]: ERROR nova.compute.manager [instance: aeca6a94-478d-435a-9d93-59936f26654e] vm_ref = self.build_virtual_machine(instance, [ 684.026392] env[61923]: ERROR nova.compute.manager [instance: aeca6a94-478d-435a-9d93-59936f26654e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 684.026392] env[61923]: ERROR nova.compute.manager [instance: aeca6a94-478d-435a-9d93-59936f26654e] vif_infos = vmwarevif.get_vif_info(self._session, [ 684.026392] env[61923]: ERROR nova.compute.manager [instance: aeca6a94-478d-435a-9d93-59936f26654e] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 684.026392] env[61923]: ERROR nova.compute.manager [instance: aeca6a94-478d-435a-9d93-59936f26654e] for vif in network_info: [ 684.026392] env[61923]: ERROR nova.compute.manager [instance: aeca6a94-478d-435a-9d93-59936f26654e] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 684.026392] env[61923]: ERROR nova.compute.manager [instance: aeca6a94-478d-435a-9d93-59936f26654e] return self._sync_wrapper(fn, *args, **kwargs) [ 684.026392] env[61923]: ERROR nova.compute.manager [instance: aeca6a94-478d-435a-9d93-59936f26654e] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 684.026392] env[61923]: ERROR nova.compute.manager [instance: aeca6a94-478d-435a-9d93-59936f26654e] self.wait() [ 684.026392] env[61923]: ERROR nova.compute.manager [instance: aeca6a94-478d-435a-9d93-59936f26654e] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 684.026392] env[61923]: ERROR nova.compute.manager [instance: aeca6a94-478d-435a-9d93-59936f26654e] self[:] = self._gt.wait() [ 684.026392] env[61923]: ERROR nova.compute.manager [instance: aeca6a94-478d-435a-9d93-59936f26654e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 684.026392] env[61923]: ERROR nova.compute.manager [instance: aeca6a94-478d-435a-9d93-59936f26654e] return self._exit_event.wait() [ 684.026392] env[61923]: ERROR nova.compute.manager [instance: aeca6a94-478d-435a-9d93-59936f26654e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 684.026774] env[61923]: ERROR nova.compute.manager [instance: aeca6a94-478d-435a-9d93-59936f26654e] result = hub.switch() [ 684.026774] env[61923]: ERROR nova.compute.manager [instance: aeca6a94-478d-435a-9d93-59936f26654e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 684.026774] env[61923]: ERROR nova.compute.manager [instance: aeca6a94-478d-435a-9d93-59936f26654e] return self.greenlet.switch() [ 684.026774] env[61923]: ERROR nova.compute.manager [instance: aeca6a94-478d-435a-9d93-59936f26654e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 684.026774] env[61923]: ERROR nova.compute.manager [instance: aeca6a94-478d-435a-9d93-59936f26654e] result = function(*args, **kwargs) [ 684.026774] env[61923]: ERROR nova.compute.manager [instance: aeca6a94-478d-435a-9d93-59936f26654e] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 684.026774] env[61923]: ERROR nova.compute.manager [instance: aeca6a94-478d-435a-9d93-59936f26654e] return func(*args, **kwargs) [ 684.026774] env[61923]: ERROR nova.compute.manager [instance: aeca6a94-478d-435a-9d93-59936f26654e] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 684.026774] env[61923]: ERROR nova.compute.manager [instance: aeca6a94-478d-435a-9d93-59936f26654e] raise e [ 684.026774] env[61923]: ERROR nova.compute.manager [instance: aeca6a94-478d-435a-9d93-59936f26654e] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 684.026774] env[61923]: ERROR nova.compute.manager [instance: aeca6a94-478d-435a-9d93-59936f26654e] nwinfo = self.network_api.allocate_for_instance( [ 684.026774] env[61923]: ERROR nova.compute.manager [instance: aeca6a94-478d-435a-9d93-59936f26654e] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 684.026774] env[61923]: ERROR nova.compute.manager [instance: aeca6a94-478d-435a-9d93-59936f26654e] created_port_ids = self._update_ports_for_instance( [ 684.027087] env[61923]: ERROR nova.compute.manager [instance: aeca6a94-478d-435a-9d93-59936f26654e] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 684.027087] env[61923]: ERROR nova.compute.manager [instance: aeca6a94-478d-435a-9d93-59936f26654e] with excutils.save_and_reraise_exception(): [ 684.027087] env[61923]: ERROR nova.compute.manager [instance: aeca6a94-478d-435a-9d93-59936f26654e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 684.027087] env[61923]: ERROR nova.compute.manager [instance: aeca6a94-478d-435a-9d93-59936f26654e] self.force_reraise() [ 684.027087] env[61923]: ERROR nova.compute.manager [instance: aeca6a94-478d-435a-9d93-59936f26654e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 684.027087] env[61923]: ERROR nova.compute.manager [instance: aeca6a94-478d-435a-9d93-59936f26654e] raise self.value [ 684.027087] env[61923]: ERROR nova.compute.manager [instance: aeca6a94-478d-435a-9d93-59936f26654e] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 684.027087] env[61923]: ERROR nova.compute.manager [instance: aeca6a94-478d-435a-9d93-59936f26654e] updated_port = self._update_port( [ 684.027087] env[61923]: ERROR nova.compute.manager [instance: aeca6a94-478d-435a-9d93-59936f26654e] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 684.027087] env[61923]: ERROR nova.compute.manager [instance: aeca6a94-478d-435a-9d93-59936f26654e] _ensure_no_port_binding_failure(port) [ 684.027087] env[61923]: ERROR nova.compute.manager [instance: aeca6a94-478d-435a-9d93-59936f26654e] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 684.027087] env[61923]: ERROR nova.compute.manager [instance: aeca6a94-478d-435a-9d93-59936f26654e] raise exception.PortBindingFailed(port_id=port['id']) [ 684.027372] env[61923]: ERROR nova.compute.manager [instance: aeca6a94-478d-435a-9d93-59936f26654e] nova.exception.PortBindingFailed: Binding failed for port 61e9a5e4-10ed-40b2-a961-2d5cbae80f02, please check neutron logs for more information. [ 684.027372] env[61923]: ERROR nova.compute.manager [instance: aeca6a94-478d-435a-9d93-59936f26654e] [ 684.027372] env[61923]: DEBUG nova.compute.utils [None req-56950b6c-0b87-48bf-a575-f699fcd1a5b0 tempest-FloatingIPsAssociationTestJSON-1125124803 tempest-FloatingIPsAssociationTestJSON-1125124803-project-member] [instance: aeca6a94-478d-435a-9d93-59936f26654e] Binding failed for port 61e9a5e4-10ed-40b2-a961-2d5cbae80f02, please check neutron logs for more information. {{(pid=61923) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 684.028335] env[61923]: DEBUG nova.compute.manager [None req-56950b6c-0b87-48bf-a575-f699fcd1a5b0 tempest-FloatingIPsAssociationTestJSON-1125124803 tempest-FloatingIPsAssociationTestJSON-1125124803-project-member] [instance: aeca6a94-478d-435a-9d93-59936f26654e] Build of instance aeca6a94-478d-435a-9d93-59936f26654e was re-scheduled: Binding failed for port 61e9a5e4-10ed-40b2-a961-2d5cbae80f02, please check neutron logs for more information. {{(pid=61923) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 684.031121] env[61923]: DEBUG nova.compute.manager [None req-56950b6c-0b87-48bf-a575-f699fcd1a5b0 tempest-FloatingIPsAssociationTestJSON-1125124803 tempest-FloatingIPsAssociationTestJSON-1125124803-project-member] [instance: aeca6a94-478d-435a-9d93-59936f26654e] Unplugging VIFs for instance {{(pid=61923) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 684.031121] env[61923]: DEBUG oslo_concurrency.lockutils [None req-56950b6c-0b87-48bf-a575-f699fcd1a5b0 tempest-FloatingIPsAssociationTestJSON-1125124803 tempest-FloatingIPsAssociationTestJSON-1125124803-project-member] Acquiring lock "refresh_cache-aeca6a94-478d-435a-9d93-59936f26654e" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 684.031121] env[61923]: DEBUG oslo_concurrency.lockutils [None req-56950b6c-0b87-48bf-a575-f699fcd1a5b0 tempest-FloatingIPsAssociationTestJSON-1125124803 tempest-FloatingIPsAssociationTestJSON-1125124803-project-member] Acquired lock "refresh_cache-aeca6a94-478d-435a-9d93-59936f26654e" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 684.031121] env[61923]: DEBUG nova.network.neutron [None req-56950b6c-0b87-48bf-a575-f699fcd1a5b0 tempest-FloatingIPsAssociationTestJSON-1125124803 tempest-FloatingIPsAssociationTestJSON-1125124803-project-member] [instance: aeca6a94-478d-435a-9d93-59936f26654e] Building network info cache for instance {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 684.031266] env[61923]: DEBUG oslo_concurrency.lockutils [None req-5f08bfba-63c5-4161-be1b-a884c0ae71af tempest-VolumesAdminNegativeTest-1422120048 tempest-VolumesAdminNegativeTest-1422120048-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 26.083s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 684.032490] env[61923]: INFO nova.compute.claims [None req-5f08bfba-63c5-4161-be1b-a884c0ae71af tempest-VolumesAdminNegativeTest-1422120048 tempest-VolumesAdminNegativeTest-1422120048-project-member] [instance: 590907a7-b6cc-48fc-a4bf-e4a2e48b05b8] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 684.169405] env[61923]: DEBUG nova.compute.manager [req-cf59c476-3760-4edd-bd06-2419bb44c626 req-ee54db4a-47cd-4136-b749-5328ee584e9d service nova] [instance: c94e8e46-7697-426c-ae2f-aece493fa8f4] Received event network-vif-deleted-70370c2b-9eb9-48b7-8acc-8a7eafc1d41d {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 684.174907] env[61923]: DEBUG nova.compute.manager [None req-aadda1bb-f770-4742-8c34-cd25e3f7e7cb tempest-InstanceActionsNegativeTestJSON-1700332362 tempest-InstanceActionsNegativeTestJSON-1700332362-project-member] [instance: 64584976-b3f3-4da5-a76b-a05cf2ed6aa4] Start spawning the instance on the hypervisor. {{(pid=61923) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 684.209711] env[61923]: DEBUG nova.virt.hardware [None req-aadda1bb-f770-4742-8c34-cd25e3f7e7cb tempest-InstanceActionsNegativeTestJSON-1700332362 tempest-InstanceActionsNegativeTestJSON-1700332362-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-29T20:07:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-29T20:07:35Z,direct_url=,disk_format='vmdk',id=0f153f63-ae0a-45b1-b7f5-7f9b673c947f,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='4e7b5e3b3c3443c8bd5c70f8a15739f5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-29T20:07:36Z,virtual_size=,visibility=), allow threads: False {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 684.210346] env[61923]: DEBUG nova.virt.hardware [None req-aadda1bb-f770-4742-8c34-cd25e3f7e7cb tempest-InstanceActionsNegativeTestJSON-1700332362 tempest-InstanceActionsNegativeTestJSON-1700332362-project-member] Flavor limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 684.210346] env[61923]: DEBUG nova.virt.hardware [None req-aadda1bb-f770-4742-8c34-cd25e3f7e7cb tempest-InstanceActionsNegativeTestJSON-1700332362 tempest-InstanceActionsNegativeTestJSON-1700332362-project-member] Image limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 684.210346] env[61923]: DEBUG nova.virt.hardware [None req-aadda1bb-f770-4742-8c34-cd25e3f7e7cb tempest-InstanceActionsNegativeTestJSON-1700332362 tempest-InstanceActionsNegativeTestJSON-1700332362-project-member] Flavor pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 684.210601] env[61923]: DEBUG nova.virt.hardware [None req-aadda1bb-f770-4742-8c34-cd25e3f7e7cb tempest-InstanceActionsNegativeTestJSON-1700332362 tempest-InstanceActionsNegativeTestJSON-1700332362-project-member] Image pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 684.210601] env[61923]: DEBUG nova.virt.hardware [None req-aadda1bb-f770-4742-8c34-cd25e3f7e7cb tempest-InstanceActionsNegativeTestJSON-1700332362 tempest-InstanceActionsNegativeTestJSON-1700332362-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 684.210740] env[61923]: DEBUG nova.virt.hardware [None req-aadda1bb-f770-4742-8c34-cd25e3f7e7cb tempest-InstanceActionsNegativeTestJSON-1700332362 tempest-InstanceActionsNegativeTestJSON-1700332362-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 684.210880] env[61923]: DEBUG nova.virt.hardware [None req-aadda1bb-f770-4742-8c34-cd25e3f7e7cb tempest-InstanceActionsNegativeTestJSON-1700332362 tempest-InstanceActionsNegativeTestJSON-1700332362-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 684.212565] env[61923]: DEBUG nova.virt.hardware [None req-aadda1bb-f770-4742-8c34-cd25e3f7e7cb tempest-InstanceActionsNegativeTestJSON-1700332362 tempest-InstanceActionsNegativeTestJSON-1700332362-project-member] Got 1 possible topologies {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 684.212868] env[61923]: DEBUG nova.virt.hardware [None req-aadda1bb-f770-4742-8c34-cd25e3f7e7cb tempest-InstanceActionsNegativeTestJSON-1700332362 tempest-InstanceActionsNegativeTestJSON-1700332362-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 684.213020] env[61923]: DEBUG nova.virt.hardware [None req-aadda1bb-f770-4742-8c34-cd25e3f7e7cb tempest-InstanceActionsNegativeTestJSON-1700332362 tempest-InstanceActionsNegativeTestJSON-1700332362-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 684.215669] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-daa14d1a-5ab8-44cb-bea6-6d76d6dce012 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.224384] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1607b07e-0e97-446b-b246-0a25b94032b4 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.302410] env[61923]: DEBUG nova.network.neutron [None req-1f24f04c-cf4b-4940-9447-eacc11fa5e66 tempest-ServersTestFqdnHostnames-345291565 tempest-ServersTestFqdnHostnames-345291565-project-member] [instance: c94e8e46-7697-426c-ae2f-aece493fa8f4] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 684.586498] env[61923]: DEBUG nova.network.neutron [None req-56950b6c-0b87-48bf-a575-f699fcd1a5b0 tempest-FloatingIPsAssociationTestJSON-1125124803 tempest-FloatingIPsAssociationTestJSON-1125124803-project-member] [instance: aeca6a94-478d-435a-9d93-59936f26654e] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 684.806148] env[61923]: DEBUG oslo_concurrency.lockutils [None req-1f24f04c-cf4b-4940-9447-eacc11fa5e66 tempest-ServersTestFqdnHostnames-345291565 tempest-ServersTestFqdnHostnames-345291565-project-member] Releasing lock "refresh_cache-c94e8e46-7697-426c-ae2f-aece493fa8f4" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 684.806682] env[61923]: DEBUG nova.compute.manager [None req-1f24f04c-cf4b-4940-9447-eacc11fa5e66 tempest-ServersTestFqdnHostnames-345291565 tempest-ServersTestFqdnHostnames-345291565-project-member] [instance: c94e8e46-7697-426c-ae2f-aece493fa8f4] Start destroying the instance on the hypervisor. {{(pid=61923) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 684.806803] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-1f24f04c-cf4b-4940-9447-eacc11fa5e66 tempest-ServersTestFqdnHostnames-345291565 tempest-ServersTestFqdnHostnames-345291565-project-member] [instance: c94e8e46-7697-426c-ae2f-aece493fa8f4] Destroying instance {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 684.807151] env[61923]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-29a86239-30ba-4844-a0de-d1a4e0941990 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.817398] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-845d41d1-221d-4e4a-b268-bca3d57b960a {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.842074] env[61923]: DEBUG nova.network.neutron [None req-56950b6c-0b87-48bf-a575-f699fcd1a5b0 tempest-FloatingIPsAssociationTestJSON-1125124803 tempest-FloatingIPsAssociationTestJSON-1125124803-project-member] [instance: aeca6a94-478d-435a-9d93-59936f26654e] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 684.853803] env[61923]: WARNING nova.virt.vmwareapi.vmops [None req-1f24f04c-cf4b-4940-9447-eacc11fa5e66 tempest-ServersTestFqdnHostnames-345291565 tempest-ServersTestFqdnHostnames-345291565-project-member] [instance: c94e8e46-7697-426c-ae2f-aece493fa8f4] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance c94e8e46-7697-426c-ae2f-aece493fa8f4 could not be found. [ 684.854051] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-1f24f04c-cf4b-4940-9447-eacc11fa5e66 tempest-ServersTestFqdnHostnames-345291565 tempest-ServersTestFqdnHostnames-345291565-project-member] [instance: c94e8e46-7697-426c-ae2f-aece493fa8f4] Instance destroyed {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 684.854413] env[61923]: INFO nova.compute.manager [None req-1f24f04c-cf4b-4940-9447-eacc11fa5e66 tempest-ServersTestFqdnHostnames-345291565 tempest-ServersTestFqdnHostnames-345291565-project-member] [instance: c94e8e46-7697-426c-ae2f-aece493fa8f4] Took 0.05 seconds to destroy the instance on the hypervisor. [ 684.854544] env[61923]: DEBUG oslo.service.loopingcall [None req-1f24f04c-cf4b-4940-9447-eacc11fa5e66 tempest-ServersTestFqdnHostnames-345291565 tempest-ServersTestFqdnHostnames-345291565-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61923) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 684.854824] env[61923]: DEBUG nova.compute.manager [-] [instance: c94e8e46-7697-426c-ae2f-aece493fa8f4] Deallocating network for instance {{(pid=61923) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 684.854924] env[61923]: DEBUG nova.network.neutron [-] [instance: c94e8e46-7697-426c-ae2f-aece493fa8f4] deallocate_for_instance() {{(pid=61923) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 684.888026] env[61923]: DEBUG nova.network.neutron [-] [instance: c94e8e46-7697-426c-ae2f-aece493fa8f4] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 685.346250] env[61923]: DEBUG oslo_concurrency.lockutils [None req-56950b6c-0b87-48bf-a575-f699fcd1a5b0 tempest-FloatingIPsAssociationTestJSON-1125124803 tempest-FloatingIPsAssociationTestJSON-1125124803-project-member] Releasing lock "refresh_cache-aeca6a94-478d-435a-9d93-59936f26654e" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 685.346250] env[61923]: DEBUG nova.compute.manager [None req-56950b6c-0b87-48bf-a575-f699fcd1a5b0 tempest-FloatingIPsAssociationTestJSON-1125124803 tempest-FloatingIPsAssociationTestJSON-1125124803-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61923) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 685.346250] env[61923]: DEBUG nova.compute.manager [None req-56950b6c-0b87-48bf-a575-f699fcd1a5b0 tempest-FloatingIPsAssociationTestJSON-1125124803 tempest-FloatingIPsAssociationTestJSON-1125124803-project-member] [instance: aeca6a94-478d-435a-9d93-59936f26654e] Deallocating network for instance {{(pid=61923) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 685.346250] env[61923]: DEBUG nova.network.neutron [None req-56950b6c-0b87-48bf-a575-f699fcd1a5b0 tempest-FloatingIPsAssociationTestJSON-1125124803 tempest-FloatingIPsAssociationTestJSON-1125124803-project-member] [instance: aeca6a94-478d-435a-9d93-59936f26654e] deallocate_for_instance() {{(pid=61923) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 685.375337] env[61923]: DEBUG nova.network.neutron [None req-56950b6c-0b87-48bf-a575-f699fcd1a5b0 tempest-FloatingIPsAssociationTestJSON-1125124803 tempest-FloatingIPsAssociationTestJSON-1125124803-project-member] [instance: aeca6a94-478d-435a-9d93-59936f26654e] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 685.390817] env[61923]: DEBUG nova.network.neutron [-] [instance: c94e8e46-7697-426c-ae2f-aece493fa8f4] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 685.447744] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63f6e156-674d-45d5-abda-9a105d735975 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.455938] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c64bbf55-690d-44da-b6da-68e8f42598ae {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.490577] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7d8f5d0-3cf7-461d-bf47-0e7b03906bf2 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.498361] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7c5d45b-bc80-427a-95b4-313a5bf1567a {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 685.516837] env[61923]: DEBUG nova.compute.provider_tree [None req-5f08bfba-63c5-4161-be1b-a884c0ae71af tempest-VolumesAdminNegativeTest-1422120048 tempest-VolumesAdminNegativeTest-1422120048-project-member] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 685.695820] env[61923]: DEBUG oslo_concurrency.lockutils [None req-1d78b71c-3449-4286-a17a-dc5396f589fe tempest-ServersAdminTestJSON-678434436 tempest-ServersAdminTestJSON-678434436-project-member] Acquiring lock "faff83a1-3532-4396-ac3f-600f86653749" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 685.696349] env[61923]: DEBUG oslo_concurrency.lockutils [None req-1d78b71c-3449-4286-a17a-dc5396f589fe tempest-ServersAdminTestJSON-678434436 tempest-ServersAdminTestJSON-678434436-project-member] Lock "faff83a1-3532-4396-ac3f-600f86653749" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 685.882408] env[61923]: DEBUG nova.network.neutron [None req-56950b6c-0b87-48bf-a575-f699fcd1a5b0 tempest-FloatingIPsAssociationTestJSON-1125124803 tempest-FloatingIPsAssociationTestJSON-1125124803-project-member] [instance: aeca6a94-478d-435a-9d93-59936f26654e] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 685.894735] env[61923]: INFO nova.compute.manager [-] [instance: c94e8e46-7697-426c-ae2f-aece493fa8f4] Took 1.04 seconds to deallocate network for instance. [ 685.896807] env[61923]: DEBUG nova.compute.claims [None req-1f24f04c-cf4b-4940-9447-eacc11fa5e66 tempest-ServersTestFqdnHostnames-345291565 tempest-ServersTestFqdnHostnames-345291565-project-member] [instance: c94e8e46-7697-426c-ae2f-aece493fa8f4] Aborting claim: {{(pid=61923) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 685.897268] env[61923]: DEBUG oslo_concurrency.lockutils [None req-1f24f04c-cf4b-4940-9447-eacc11fa5e66 tempest-ServersTestFqdnHostnames-345291565 tempest-ServersTestFqdnHostnames-345291565-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 686.021489] env[61923]: DEBUG nova.scheduler.client.report [None req-5f08bfba-63c5-4161-be1b-a884c0ae71af tempest-VolumesAdminNegativeTest-1422120048 tempest-VolumesAdminNegativeTest-1422120048-project-member] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 686.209347] env[61923]: DEBUG nova.compute.manager [req-7e66dda9-b0f1-4419-b7e2-38e72b89c5b7 req-a87a0c65-c387-4181-810c-dbae678dfc26 service nova] [instance: 64584976-b3f3-4da5-a76b-a05cf2ed6aa4] Received event network-changed-13363b75-27bb-4922-b5da-febd2dc08e25 {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 686.209857] env[61923]: DEBUG nova.compute.manager [req-7e66dda9-b0f1-4419-b7e2-38e72b89c5b7 req-a87a0c65-c387-4181-810c-dbae678dfc26 service nova] [instance: 64584976-b3f3-4da5-a76b-a05cf2ed6aa4] Refreshing instance network info cache due to event network-changed-13363b75-27bb-4922-b5da-febd2dc08e25. {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 686.210189] env[61923]: DEBUG oslo_concurrency.lockutils [req-7e66dda9-b0f1-4419-b7e2-38e72b89c5b7 req-a87a0c65-c387-4181-810c-dbae678dfc26 service nova] Acquiring lock "refresh_cache-64584976-b3f3-4da5-a76b-a05cf2ed6aa4" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 686.210392] env[61923]: DEBUG oslo_concurrency.lockutils [req-7e66dda9-b0f1-4419-b7e2-38e72b89c5b7 req-a87a0c65-c387-4181-810c-dbae678dfc26 service nova] Acquired lock "refresh_cache-64584976-b3f3-4da5-a76b-a05cf2ed6aa4" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 686.210657] env[61923]: DEBUG nova.network.neutron [req-7e66dda9-b0f1-4419-b7e2-38e72b89c5b7 req-a87a0c65-c387-4181-810c-dbae678dfc26 service nova] [instance: 64584976-b3f3-4da5-a76b-a05cf2ed6aa4] Refreshing network info cache for port 13363b75-27bb-4922-b5da-febd2dc08e25 {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 686.385978] env[61923]: INFO nova.compute.manager [None req-56950b6c-0b87-48bf-a575-f699fcd1a5b0 tempest-FloatingIPsAssociationTestJSON-1125124803 tempest-FloatingIPsAssociationTestJSON-1125124803-project-member] [instance: aeca6a94-478d-435a-9d93-59936f26654e] Took 1.04 seconds to deallocate network for instance. [ 686.417091] env[61923]: ERROR nova.compute.manager [None req-aadda1bb-f770-4742-8c34-cd25e3f7e7cb tempest-InstanceActionsNegativeTestJSON-1700332362 tempest-InstanceActionsNegativeTestJSON-1700332362-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 13363b75-27bb-4922-b5da-febd2dc08e25, please check neutron logs for more information. [ 686.417091] env[61923]: ERROR nova.compute.manager Traceback (most recent call last): [ 686.417091] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 686.417091] env[61923]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 686.417091] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 686.417091] env[61923]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 686.417091] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 686.417091] env[61923]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 686.417091] env[61923]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 686.417091] env[61923]: ERROR nova.compute.manager self.force_reraise() [ 686.417091] env[61923]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 686.417091] env[61923]: ERROR nova.compute.manager raise self.value [ 686.417091] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 686.417091] env[61923]: ERROR nova.compute.manager updated_port = self._update_port( [ 686.417091] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 686.417091] env[61923]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 686.417580] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 686.417580] env[61923]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 686.417580] env[61923]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 13363b75-27bb-4922-b5da-febd2dc08e25, please check neutron logs for more information. [ 686.417580] env[61923]: ERROR nova.compute.manager [ 686.417580] env[61923]: Traceback (most recent call last): [ 686.417580] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 686.417580] env[61923]: listener.cb(fileno) [ 686.417580] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 686.417580] env[61923]: result = function(*args, **kwargs) [ 686.417580] env[61923]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 686.417580] env[61923]: return func(*args, **kwargs) [ 686.417580] env[61923]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 686.417580] env[61923]: raise e [ 686.417580] env[61923]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 686.417580] env[61923]: nwinfo = self.network_api.allocate_for_instance( [ 686.417580] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 686.417580] env[61923]: created_port_ids = self._update_ports_for_instance( [ 686.417580] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 686.417580] env[61923]: with excutils.save_and_reraise_exception(): [ 686.417580] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 686.417580] env[61923]: self.force_reraise() [ 686.417580] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 686.417580] env[61923]: raise self.value [ 686.417580] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 686.417580] env[61923]: updated_port = self._update_port( [ 686.417580] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 686.417580] env[61923]: _ensure_no_port_binding_failure(port) [ 686.417580] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 686.417580] env[61923]: raise exception.PortBindingFailed(port_id=port['id']) [ 686.418312] env[61923]: nova.exception.PortBindingFailed: Binding failed for port 13363b75-27bb-4922-b5da-febd2dc08e25, please check neutron logs for more information. [ 686.418312] env[61923]: Removing descriptor: 14 [ 686.418312] env[61923]: ERROR nova.compute.manager [None req-aadda1bb-f770-4742-8c34-cd25e3f7e7cb tempest-InstanceActionsNegativeTestJSON-1700332362 tempest-InstanceActionsNegativeTestJSON-1700332362-project-member] [instance: 64584976-b3f3-4da5-a76b-a05cf2ed6aa4] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 13363b75-27bb-4922-b5da-febd2dc08e25, please check neutron logs for more information. [ 686.418312] env[61923]: ERROR nova.compute.manager [instance: 64584976-b3f3-4da5-a76b-a05cf2ed6aa4] Traceback (most recent call last): [ 686.418312] env[61923]: ERROR nova.compute.manager [instance: 64584976-b3f3-4da5-a76b-a05cf2ed6aa4] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 686.418312] env[61923]: ERROR nova.compute.manager [instance: 64584976-b3f3-4da5-a76b-a05cf2ed6aa4] yield resources [ 686.418312] env[61923]: ERROR nova.compute.manager [instance: 64584976-b3f3-4da5-a76b-a05cf2ed6aa4] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 686.418312] env[61923]: ERROR nova.compute.manager [instance: 64584976-b3f3-4da5-a76b-a05cf2ed6aa4] self.driver.spawn(context, instance, image_meta, [ 686.418312] env[61923]: ERROR nova.compute.manager [instance: 64584976-b3f3-4da5-a76b-a05cf2ed6aa4] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 686.418312] env[61923]: ERROR nova.compute.manager [instance: 64584976-b3f3-4da5-a76b-a05cf2ed6aa4] self._vmops.spawn(context, instance, image_meta, injected_files, [ 686.418312] env[61923]: ERROR nova.compute.manager [instance: 64584976-b3f3-4da5-a76b-a05cf2ed6aa4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 686.418312] env[61923]: ERROR nova.compute.manager [instance: 64584976-b3f3-4da5-a76b-a05cf2ed6aa4] vm_ref = self.build_virtual_machine(instance, [ 686.418618] env[61923]: ERROR nova.compute.manager [instance: 64584976-b3f3-4da5-a76b-a05cf2ed6aa4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 686.418618] env[61923]: ERROR nova.compute.manager [instance: 64584976-b3f3-4da5-a76b-a05cf2ed6aa4] vif_infos = vmwarevif.get_vif_info(self._session, [ 686.418618] env[61923]: ERROR nova.compute.manager [instance: 64584976-b3f3-4da5-a76b-a05cf2ed6aa4] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 686.418618] env[61923]: ERROR nova.compute.manager [instance: 64584976-b3f3-4da5-a76b-a05cf2ed6aa4] for vif in network_info: [ 686.418618] env[61923]: ERROR nova.compute.manager [instance: 64584976-b3f3-4da5-a76b-a05cf2ed6aa4] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 686.418618] env[61923]: ERROR nova.compute.manager [instance: 64584976-b3f3-4da5-a76b-a05cf2ed6aa4] return self._sync_wrapper(fn, *args, **kwargs) [ 686.418618] env[61923]: ERROR nova.compute.manager [instance: 64584976-b3f3-4da5-a76b-a05cf2ed6aa4] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 686.418618] env[61923]: ERROR nova.compute.manager [instance: 64584976-b3f3-4da5-a76b-a05cf2ed6aa4] self.wait() [ 686.418618] env[61923]: ERROR nova.compute.manager [instance: 64584976-b3f3-4da5-a76b-a05cf2ed6aa4] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 686.418618] env[61923]: ERROR nova.compute.manager [instance: 64584976-b3f3-4da5-a76b-a05cf2ed6aa4] self[:] = self._gt.wait() [ 686.418618] env[61923]: ERROR nova.compute.manager [instance: 64584976-b3f3-4da5-a76b-a05cf2ed6aa4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 686.418618] env[61923]: ERROR nova.compute.manager [instance: 64584976-b3f3-4da5-a76b-a05cf2ed6aa4] return self._exit_event.wait() [ 686.418618] env[61923]: ERROR nova.compute.manager [instance: 64584976-b3f3-4da5-a76b-a05cf2ed6aa4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 686.419014] env[61923]: ERROR nova.compute.manager [instance: 64584976-b3f3-4da5-a76b-a05cf2ed6aa4] result = hub.switch() [ 686.419014] env[61923]: ERROR nova.compute.manager [instance: 64584976-b3f3-4da5-a76b-a05cf2ed6aa4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 686.419014] env[61923]: ERROR nova.compute.manager [instance: 64584976-b3f3-4da5-a76b-a05cf2ed6aa4] return self.greenlet.switch() [ 686.419014] env[61923]: ERROR nova.compute.manager [instance: 64584976-b3f3-4da5-a76b-a05cf2ed6aa4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 686.419014] env[61923]: ERROR nova.compute.manager [instance: 64584976-b3f3-4da5-a76b-a05cf2ed6aa4] result = function(*args, **kwargs) [ 686.419014] env[61923]: ERROR nova.compute.manager [instance: 64584976-b3f3-4da5-a76b-a05cf2ed6aa4] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 686.419014] env[61923]: ERROR nova.compute.manager [instance: 64584976-b3f3-4da5-a76b-a05cf2ed6aa4] return func(*args, **kwargs) [ 686.419014] env[61923]: ERROR nova.compute.manager [instance: 64584976-b3f3-4da5-a76b-a05cf2ed6aa4] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 686.419014] env[61923]: ERROR nova.compute.manager [instance: 64584976-b3f3-4da5-a76b-a05cf2ed6aa4] raise e [ 686.419014] env[61923]: ERROR nova.compute.manager [instance: 64584976-b3f3-4da5-a76b-a05cf2ed6aa4] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 686.419014] env[61923]: ERROR nova.compute.manager [instance: 64584976-b3f3-4da5-a76b-a05cf2ed6aa4] nwinfo = self.network_api.allocate_for_instance( [ 686.419014] env[61923]: ERROR nova.compute.manager [instance: 64584976-b3f3-4da5-a76b-a05cf2ed6aa4] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 686.419014] env[61923]: ERROR nova.compute.manager [instance: 64584976-b3f3-4da5-a76b-a05cf2ed6aa4] created_port_ids = self._update_ports_for_instance( [ 686.419327] env[61923]: ERROR nova.compute.manager [instance: 64584976-b3f3-4da5-a76b-a05cf2ed6aa4] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 686.419327] env[61923]: ERROR nova.compute.manager [instance: 64584976-b3f3-4da5-a76b-a05cf2ed6aa4] with excutils.save_and_reraise_exception(): [ 686.419327] env[61923]: ERROR nova.compute.manager [instance: 64584976-b3f3-4da5-a76b-a05cf2ed6aa4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 686.419327] env[61923]: ERROR nova.compute.manager [instance: 64584976-b3f3-4da5-a76b-a05cf2ed6aa4] self.force_reraise() [ 686.419327] env[61923]: ERROR nova.compute.manager [instance: 64584976-b3f3-4da5-a76b-a05cf2ed6aa4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 686.419327] env[61923]: ERROR nova.compute.manager [instance: 64584976-b3f3-4da5-a76b-a05cf2ed6aa4] raise self.value [ 686.419327] env[61923]: ERROR nova.compute.manager [instance: 64584976-b3f3-4da5-a76b-a05cf2ed6aa4] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 686.419327] env[61923]: ERROR nova.compute.manager [instance: 64584976-b3f3-4da5-a76b-a05cf2ed6aa4] updated_port = self._update_port( [ 686.419327] env[61923]: ERROR nova.compute.manager [instance: 64584976-b3f3-4da5-a76b-a05cf2ed6aa4] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 686.419327] env[61923]: ERROR nova.compute.manager [instance: 64584976-b3f3-4da5-a76b-a05cf2ed6aa4] _ensure_no_port_binding_failure(port) [ 686.419327] env[61923]: ERROR nova.compute.manager [instance: 64584976-b3f3-4da5-a76b-a05cf2ed6aa4] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 686.419327] env[61923]: ERROR nova.compute.manager [instance: 64584976-b3f3-4da5-a76b-a05cf2ed6aa4] raise exception.PortBindingFailed(port_id=port['id']) [ 686.419609] env[61923]: ERROR nova.compute.manager [instance: 64584976-b3f3-4da5-a76b-a05cf2ed6aa4] nova.exception.PortBindingFailed: Binding failed for port 13363b75-27bb-4922-b5da-febd2dc08e25, please check neutron logs for more information. [ 686.419609] env[61923]: ERROR nova.compute.manager [instance: 64584976-b3f3-4da5-a76b-a05cf2ed6aa4] [ 686.419609] env[61923]: INFO nova.compute.manager [None req-aadda1bb-f770-4742-8c34-cd25e3f7e7cb tempest-InstanceActionsNegativeTestJSON-1700332362 tempest-InstanceActionsNegativeTestJSON-1700332362-project-member] [instance: 64584976-b3f3-4da5-a76b-a05cf2ed6aa4] Terminating instance [ 686.420443] env[61923]: DEBUG oslo_concurrency.lockutils [None req-aadda1bb-f770-4742-8c34-cd25e3f7e7cb tempest-InstanceActionsNegativeTestJSON-1700332362 tempest-InstanceActionsNegativeTestJSON-1700332362-project-member] Acquiring lock "refresh_cache-64584976-b3f3-4da5-a76b-a05cf2ed6aa4" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 686.532553] env[61923]: DEBUG oslo_concurrency.lockutils [None req-5f08bfba-63c5-4161-be1b-a884c0ae71af tempest-VolumesAdminNegativeTest-1422120048 tempest-VolumesAdminNegativeTest-1422120048-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.502s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 686.533226] env[61923]: DEBUG nova.compute.manager [None req-5f08bfba-63c5-4161-be1b-a884c0ae71af tempest-VolumesAdminNegativeTest-1422120048 tempest-VolumesAdminNegativeTest-1422120048-project-member] [instance: 590907a7-b6cc-48fc-a4bf-e4a2e48b05b8] Start building networks asynchronously for instance. {{(pid=61923) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 686.539868] env[61923]: DEBUG oslo_concurrency.lockutils [None req-08382130-d70f-42e8-a5b5-93e86f5fa49a tempest-AttachVolumeShelveTestJSON-1252849694 tempest-AttachVolumeShelveTestJSON-1252849694-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 26.578s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 686.540572] env[61923]: INFO nova.compute.claims [None req-08382130-d70f-42e8-a5b5-93e86f5fa49a tempest-AttachVolumeShelveTestJSON-1252849694 tempest-AttachVolumeShelveTestJSON-1252849694-project-member] [instance: 60ffc8b0-4a7b-4e2a-8774-d16e502d8a06] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 686.622160] env[61923]: DEBUG oslo_concurrency.lockutils [None req-6e6344ac-0cd2-410e-863e-3289b0c1fe3f tempest-ServersAdminTestJSON-678434436 tempest-ServersAdminTestJSON-678434436-project-member] Acquiring lock "3f298e29-69ff-46ac-8018-ceb3990b9848" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 686.622318] env[61923]: DEBUG oslo_concurrency.lockutils [None req-6e6344ac-0cd2-410e-863e-3289b0c1fe3f tempest-ServersAdminTestJSON-678434436 tempest-ServersAdminTestJSON-678434436-project-member] Lock "3f298e29-69ff-46ac-8018-ceb3990b9848" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 686.735691] env[61923]: DEBUG nova.network.neutron [req-7e66dda9-b0f1-4419-b7e2-38e72b89c5b7 req-a87a0c65-c387-4181-810c-dbae678dfc26 service nova] [instance: 64584976-b3f3-4da5-a76b-a05cf2ed6aa4] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 686.870152] env[61923]: DEBUG nova.network.neutron [req-7e66dda9-b0f1-4419-b7e2-38e72b89c5b7 req-a87a0c65-c387-4181-810c-dbae678dfc26 service nova] [instance: 64584976-b3f3-4da5-a76b-a05cf2ed6aa4] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 687.042446] env[61923]: DEBUG nova.compute.utils [None req-5f08bfba-63c5-4161-be1b-a884c0ae71af tempest-VolumesAdminNegativeTest-1422120048 tempest-VolumesAdminNegativeTest-1422120048-project-member] Using /dev/sd instead of None {{(pid=61923) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 687.042446] env[61923]: DEBUG nova.compute.manager [None req-5f08bfba-63c5-4161-be1b-a884c0ae71af tempest-VolumesAdminNegativeTest-1422120048 tempest-VolumesAdminNegativeTest-1422120048-project-member] [instance: 590907a7-b6cc-48fc-a4bf-e4a2e48b05b8] Allocating IP information in the background. {{(pid=61923) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 687.042446] env[61923]: DEBUG nova.network.neutron [None req-5f08bfba-63c5-4161-be1b-a884c0ae71af tempest-VolumesAdminNegativeTest-1422120048 tempest-VolumesAdminNegativeTest-1422120048-project-member] [instance: 590907a7-b6cc-48fc-a4bf-e4a2e48b05b8] allocate_for_instance() {{(pid=61923) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 687.155566] env[61923]: DEBUG nova.policy [None req-5f08bfba-63c5-4161-be1b-a884c0ae71af tempest-VolumesAdminNegativeTest-1422120048 tempest-VolumesAdminNegativeTest-1422120048-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '98dcbc29b88741d5b51f311bb39f8a7a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '97570799e5f54fdeb7e8dfd2f50cc8a6', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61923) authorize /opt/stack/nova/nova/policy.py:201}} [ 687.375539] env[61923]: DEBUG oslo_concurrency.lockutils [req-7e66dda9-b0f1-4419-b7e2-38e72b89c5b7 req-a87a0c65-c387-4181-810c-dbae678dfc26 service nova] Releasing lock "refresh_cache-64584976-b3f3-4da5-a76b-a05cf2ed6aa4" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 687.376056] env[61923]: DEBUG oslo_concurrency.lockutils [None req-aadda1bb-f770-4742-8c34-cd25e3f7e7cb tempest-InstanceActionsNegativeTestJSON-1700332362 tempest-InstanceActionsNegativeTestJSON-1700332362-project-member] Acquired lock "refresh_cache-64584976-b3f3-4da5-a76b-a05cf2ed6aa4" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 687.376259] env[61923]: DEBUG nova.network.neutron [None req-aadda1bb-f770-4742-8c34-cd25e3f7e7cb tempest-InstanceActionsNegativeTestJSON-1700332362 tempest-InstanceActionsNegativeTestJSON-1700332362-project-member] [instance: 64584976-b3f3-4da5-a76b-a05cf2ed6aa4] Building network info cache for instance {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 687.422975] env[61923]: INFO nova.scheduler.client.report [None req-56950b6c-0b87-48bf-a575-f699fcd1a5b0 tempest-FloatingIPsAssociationTestJSON-1125124803 tempest-FloatingIPsAssociationTestJSON-1125124803-project-member] Deleted allocations for instance aeca6a94-478d-435a-9d93-59936f26654e [ 687.547396] env[61923]: DEBUG nova.compute.manager [None req-5f08bfba-63c5-4161-be1b-a884c0ae71af tempest-VolumesAdminNegativeTest-1422120048 tempest-VolumesAdminNegativeTest-1422120048-project-member] [instance: 590907a7-b6cc-48fc-a4bf-e4a2e48b05b8] Start building block device mappings for instance. {{(pid=61923) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 687.605029] env[61923]: DEBUG nova.network.neutron [None req-5f08bfba-63c5-4161-be1b-a884c0ae71af tempest-VolumesAdminNegativeTest-1422120048 tempest-VolumesAdminNegativeTest-1422120048-project-member] [instance: 590907a7-b6cc-48fc-a4bf-e4a2e48b05b8] Successfully created port: 989fba39-a679-413f-8cca-904ac748fcbe {{(pid=61923) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 687.662700] env[61923]: DEBUG oslo_concurrency.lockutils [None req-37409b66-7d98-4871-81c4-a83dd0296328 tempest-ServerAddressesNegativeTestJSON-675891034 tempest-ServerAddressesNegativeTestJSON-675891034-project-member] Acquiring lock "6fe76341-d578-4cb3-9dce-c025f0d798b8" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 687.664043] env[61923]: DEBUG oslo_concurrency.lockutils [None req-37409b66-7d98-4871-81c4-a83dd0296328 tempest-ServerAddressesNegativeTestJSON-675891034 tempest-ServerAddressesNegativeTestJSON-675891034-project-member] Lock "6fe76341-d578-4cb3-9dce-c025f0d798b8" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 687.914387] env[61923]: DEBUG nova.network.neutron [None req-aadda1bb-f770-4742-8c34-cd25e3f7e7cb tempest-InstanceActionsNegativeTestJSON-1700332362 tempest-InstanceActionsNegativeTestJSON-1700332362-project-member] [instance: 64584976-b3f3-4da5-a76b-a05cf2ed6aa4] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 687.938285] env[61923]: DEBUG oslo_concurrency.lockutils [None req-56950b6c-0b87-48bf-a575-f699fcd1a5b0 tempest-FloatingIPsAssociationTestJSON-1125124803 tempest-FloatingIPsAssociationTestJSON-1125124803-project-member] Lock "aeca6a94-478d-435a-9d93-59936f26654e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 114.131s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 687.962039] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4210f513-84dc-4f20-84d6-05b5c677c4fb {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.971626] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ae53a95-b8d2-47b4-b190-b74c251bc4d7 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 688.006355] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6853ae8b-9471-4a12-a1c7-7d4f06e03687 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 688.013906] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc5390fd-6ba2-44c4-8655-72af6b3e5674 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 688.027077] env[61923]: DEBUG nova.compute.provider_tree [None req-08382130-d70f-42e8-a5b5-93e86f5fa49a tempest-AttachVolumeShelveTestJSON-1252849694 tempest-AttachVolumeShelveTestJSON-1252849694-project-member] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 688.117301] env[61923]: DEBUG nova.network.neutron [None req-aadda1bb-f770-4742-8c34-cd25e3f7e7cb tempest-InstanceActionsNegativeTestJSON-1700332362 tempest-InstanceActionsNegativeTestJSON-1700332362-project-member] [instance: 64584976-b3f3-4da5-a76b-a05cf2ed6aa4] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 688.334483] env[61923]: DEBUG nova.compute.manager [req-d635215e-1cee-4b52-9516-106c6dc4b654 req-7173879e-5272-4c7d-b7b8-284dc28e879a service nova] [instance: 64584976-b3f3-4da5-a76b-a05cf2ed6aa4] Received event network-vif-deleted-13363b75-27bb-4922-b5da-febd2dc08e25 {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 688.440938] env[61923]: DEBUG nova.compute.manager [None req-24801cc4-3974-42da-978f-a22cb8f70af0 tempest-SecurityGroupsTestJSON-509490211 tempest-SecurityGroupsTestJSON-509490211-project-member] [instance: 2939d1e5-07fe-48b3-bc87-5f67fa1a7ae4] Starting instance... {{(pid=61923) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 688.532663] env[61923]: DEBUG nova.scheduler.client.report [None req-08382130-d70f-42e8-a5b5-93e86f5fa49a tempest-AttachVolumeShelveTestJSON-1252849694 tempest-AttachVolumeShelveTestJSON-1252849694-project-member] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 688.557058] env[61923]: DEBUG nova.compute.manager [None req-5f08bfba-63c5-4161-be1b-a884c0ae71af tempest-VolumesAdminNegativeTest-1422120048 tempest-VolumesAdminNegativeTest-1422120048-project-member] [instance: 590907a7-b6cc-48fc-a4bf-e4a2e48b05b8] Start spawning the instance on the hypervisor. {{(pid=61923) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 688.584511] env[61923]: DEBUG nova.virt.hardware [None req-5f08bfba-63c5-4161-be1b-a884c0ae71af tempest-VolumesAdminNegativeTest-1422120048 tempest-VolumesAdminNegativeTest-1422120048-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-29T20:07:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-29T20:07:35Z,direct_url=,disk_format='vmdk',id=0f153f63-ae0a-45b1-b7f5-7f9b673c947f,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='4e7b5e3b3c3443c8bd5c70f8a15739f5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-29T20:07:36Z,virtual_size=,visibility=), allow threads: False {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 688.584853] env[61923]: DEBUG nova.virt.hardware [None req-5f08bfba-63c5-4161-be1b-a884c0ae71af tempest-VolumesAdminNegativeTest-1422120048 tempest-VolumesAdminNegativeTest-1422120048-project-member] Flavor limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 688.584961] env[61923]: DEBUG nova.virt.hardware [None req-5f08bfba-63c5-4161-be1b-a884c0ae71af tempest-VolumesAdminNegativeTest-1422120048 tempest-VolumesAdminNegativeTest-1422120048-project-member] Image limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 688.585152] env[61923]: DEBUG nova.virt.hardware [None req-5f08bfba-63c5-4161-be1b-a884c0ae71af tempest-VolumesAdminNegativeTest-1422120048 tempest-VolumesAdminNegativeTest-1422120048-project-member] Flavor pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 688.585312] env[61923]: DEBUG nova.virt.hardware [None req-5f08bfba-63c5-4161-be1b-a884c0ae71af tempest-VolumesAdminNegativeTest-1422120048 tempest-VolumesAdminNegativeTest-1422120048-project-member] Image pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 688.585457] env[61923]: DEBUG nova.virt.hardware [None req-5f08bfba-63c5-4161-be1b-a884c0ae71af tempest-VolumesAdminNegativeTest-1422120048 tempest-VolumesAdminNegativeTest-1422120048-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 688.585680] env[61923]: DEBUG nova.virt.hardware [None req-5f08bfba-63c5-4161-be1b-a884c0ae71af tempest-VolumesAdminNegativeTest-1422120048 tempest-VolumesAdminNegativeTest-1422120048-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 688.586600] env[61923]: DEBUG nova.virt.hardware [None req-5f08bfba-63c5-4161-be1b-a884c0ae71af tempest-VolumesAdminNegativeTest-1422120048 tempest-VolumesAdminNegativeTest-1422120048-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 688.586600] env[61923]: DEBUG nova.virt.hardware [None req-5f08bfba-63c5-4161-be1b-a884c0ae71af tempest-VolumesAdminNegativeTest-1422120048 tempest-VolumesAdminNegativeTest-1422120048-project-member] Got 1 possible topologies {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 688.586600] env[61923]: DEBUG nova.virt.hardware [None req-5f08bfba-63c5-4161-be1b-a884c0ae71af tempest-VolumesAdminNegativeTest-1422120048 tempest-VolumesAdminNegativeTest-1422120048-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 688.586600] env[61923]: DEBUG nova.virt.hardware [None req-5f08bfba-63c5-4161-be1b-a884c0ae71af tempest-VolumesAdminNegativeTest-1422120048 tempest-VolumesAdminNegativeTest-1422120048-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 688.588122] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79c756ef-299f-461e-b127-062555201320 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 688.597076] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bebd3a23-c6de-462c-9f1d-ed5a4057ac82 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 688.619990] env[61923]: DEBUG oslo_concurrency.lockutils [None req-aadda1bb-f770-4742-8c34-cd25e3f7e7cb tempest-InstanceActionsNegativeTestJSON-1700332362 tempest-InstanceActionsNegativeTestJSON-1700332362-project-member] Releasing lock "refresh_cache-64584976-b3f3-4da5-a76b-a05cf2ed6aa4" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 688.620420] env[61923]: DEBUG nova.compute.manager [None req-aadda1bb-f770-4742-8c34-cd25e3f7e7cb tempest-InstanceActionsNegativeTestJSON-1700332362 tempest-InstanceActionsNegativeTestJSON-1700332362-project-member] [instance: 64584976-b3f3-4da5-a76b-a05cf2ed6aa4] Start destroying the instance on the hypervisor. {{(pid=61923) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 688.620611] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-aadda1bb-f770-4742-8c34-cd25e3f7e7cb tempest-InstanceActionsNegativeTestJSON-1700332362 tempest-InstanceActionsNegativeTestJSON-1700332362-project-member] [instance: 64584976-b3f3-4da5-a76b-a05cf2ed6aa4] Destroying instance {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 688.620887] env[61923]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-b30261c9-5e64-46c2-8f91-aab970058630 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 688.628673] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3da79870-26f7-4d98-914b-4fec50cba2ef {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 688.651426] env[61923]: WARNING nova.virt.vmwareapi.vmops [None req-aadda1bb-f770-4742-8c34-cd25e3f7e7cb tempest-InstanceActionsNegativeTestJSON-1700332362 tempest-InstanceActionsNegativeTestJSON-1700332362-project-member] [instance: 64584976-b3f3-4da5-a76b-a05cf2ed6aa4] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 64584976-b3f3-4da5-a76b-a05cf2ed6aa4 could not be found. [ 688.651701] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-aadda1bb-f770-4742-8c34-cd25e3f7e7cb tempest-InstanceActionsNegativeTestJSON-1700332362 tempest-InstanceActionsNegativeTestJSON-1700332362-project-member] [instance: 64584976-b3f3-4da5-a76b-a05cf2ed6aa4] Instance destroyed {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 688.651884] env[61923]: INFO nova.compute.manager [None req-aadda1bb-f770-4742-8c34-cd25e3f7e7cb tempest-InstanceActionsNegativeTestJSON-1700332362 tempest-InstanceActionsNegativeTestJSON-1700332362-project-member] [instance: 64584976-b3f3-4da5-a76b-a05cf2ed6aa4] Took 0.03 seconds to destroy the instance on the hypervisor. [ 688.652508] env[61923]: DEBUG oslo.service.loopingcall [None req-aadda1bb-f770-4742-8c34-cd25e3f7e7cb tempest-InstanceActionsNegativeTestJSON-1700332362 tempest-InstanceActionsNegativeTestJSON-1700332362-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61923) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 688.652654] env[61923]: DEBUG nova.compute.manager [-] [instance: 64584976-b3f3-4da5-a76b-a05cf2ed6aa4] Deallocating network for instance {{(pid=61923) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 688.652752] env[61923]: DEBUG nova.network.neutron [-] [instance: 64584976-b3f3-4da5-a76b-a05cf2ed6aa4] deallocate_for_instance() {{(pid=61923) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 688.680118] env[61923]: DEBUG nova.network.neutron [-] [instance: 64584976-b3f3-4da5-a76b-a05cf2ed6aa4] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 688.971994] env[61923]: DEBUG oslo_concurrency.lockutils [None req-24801cc4-3974-42da-978f-a22cb8f70af0 tempest-SecurityGroupsTestJSON-509490211 tempest-SecurityGroupsTestJSON-509490211-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 689.043037] env[61923]: DEBUG oslo_concurrency.lockutils [None req-08382130-d70f-42e8-a5b5-93e86f5fa49a tempest-AttachVolumeShelveTestJSON-1252849694 tempest-AttachVolumeShelveTestJSON-1252849694-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.499s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 689.043037] env[61923]: DEBUG nova.compute.manager [None req-08382130-d70f-42e8-a5b5-93e86f5fa49a tempest-AttachVolumeShelveTestJSON-1252849694 tempest-AttachVolumeShelveTestJSON-1252849694-project-member] [instance: 60ffc8b0-4a7b-4e2a-8774-d16e502d8a06] Start building networks asynchronously for instance. {{(pid=61923) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 689.043037] env[61923]: DEBUG oslo_concurrency.lockutils [None req-6833b601-845b-439d-ad97-e2f9c0927775 tempest-ServerRescueNegativeTestJSON-1435265126 tempest-ServerRescueNegativeTestJSON-1435265126-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 27.207s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 689.044009] env[61923]: INFO nova.compute.claims [None req-6833b601-845b-439d-ad97-e2f9c0927775 tempest-ServerRescueNegativeTestJSON-1435265126 tempest-ServerRescueNegativeTestJSON-1435265126-project-member] [instance: 4f5348ec-f0b6-45e2-be04-31dbf1d49a3a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 689.188440] env[61923]: DEBUG nova.network.neutron [-] [instance: 64584976-b3f3-4da5-a76b-a05cf2ed6aa4] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 689.219328] env[61923]: ERROR nova.compute.manager [None req-5f08bfba-63c5-4161-be1b-a884c0ae71af tempest-VolumesAdminNegativeTest-1422120048 tempest-VolumesAdminNegativeTest-1422120048-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 989fba39-a679-413f-8cca-904ac748fcbe, please check neutron logs for more information. [ 689.219328] env[61923]: ERROR nova.compute.manager Traceback (most recent call last): [ 689.219328] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 689.219328] env[61923]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 689.219328] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 689.219328] env[61923]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 689.219328] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 689.219328] env[61923]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 689.219328] env[61923]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 689.219328] env[61923]: ERROR nova.compute.manager self.force_reraise() [ 689.219328] env[61923]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 689.219328] env[61923]: ERROR nova.compute.manager raise self.value [ 689.219328] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 689.219328] env[61923]: ERROR nova.compute.manager updated_port = self._update_port( [ 689.219328] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 689.219328] env[61923]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 689.220700] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 689.220700] env[61923]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 689.220700] env[61923]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 989fba39-a679-413f-8cca-904ac748fcbe, please check neutron logs for more information. [ 689.220700] env[61923]: ERROR nova.compute.manager [ 689.220700] env[61923]: Traceback (most recent call last): [ 689.220700] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 689.220700] env[61923]: listener.cb(fileno) [ 689.220700] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 689.220700] env[61923]: result = function(*args, **kwargs) [ 689.220700] env[61923]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 689.220700] env[61923]: return func(*args, **kwargs) [ 689.220700] env[61923]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 689.220700] env[61923]: raise e [ 689.220700] env[61923]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 689.220700] env[61923]: nwinfo = self.network_api.allocate_for_instance( [ 689.220700] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 689.220700] env[61923]: created_port_ids = self._update_ports_for_instance( [ 689.220700] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 689.220700] env[61923]: with excutils.save_and_reraise_exception(): [ 689.220700] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 689.220700] env[61923]: self.force_reraise() [ 689.220700] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 689.220700] env[61923]: raise self.value [ 689.220700] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 689.220700] env[61923]: updated_port = self._update_port( [ 689.220700] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 689.220700] env[61923]: _ensure_no_port_binding_failure(port) [ 689.220700] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 689.220700] env[61923]: raise exception.PortBindingFailed(port_id=port['id']) [ 689.222014] env[61923]: nova.exception.PortBindingFailed: Binding failed for port 989fba39-a679-413f-8cca-904ac748fcbe, please check neutron logs for more information. [ 689.222014] env[61923]: Removing descriptor: 14 [ 689.222014] env[61923]: ERROR nova.compute.manager [None req-5f08bfba-63c5-4161-be1b-a884c0ae71af tempest-VolumesAdminNegativeTest-1422120048 tempest-VolumesAdminNegativeTest-1422120048-project-member] [instance: 590907a7-b6cc-48fc-a4bf-e4a2e48b05b8] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 989fba39-a679-413f-8cca-904ac748fcbe, please check neutron logs for more information. [ 689.222014] env[61923]: ERROR nova.compute.manager [instance: 590907a7-b6cc-48fc-a4bf-e4a2e48b05b8] Traceback (most recent call last): [ 689.222014] env[61923]: ERROR nova.compute.manager [instance: 590907a7-b6cc-48fc-a4bf-e4a2e48b05b8] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 689.222014] env[61923]: ERROR nova.compute.manager [instance: 590907a7-b6cc-48fc-a4bf-e4a2e48b05b8] yield resources [ 689.222014] env[61923]: ERROR nova.compute.manager [instance: 590907a7-b6cc-48fc-a4bf-e4a2e48b05b8] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 689.222014] env[61923]: ERROR nova.compute.manager [instance: 590907a7-b6cc-48fc-a4bf-e4a2e48b05b8] self.driver.spawn(context, instance, image_meta, [ 689.222014] env[61923]: ERROR nova.compute.manager [instance: 590907a7-b6cc-48fc-a4bf-e4a2e48b05b8] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 689.222014] env[61923]: ERROR nova.compute.manager [instance: 590907a7-b6cc-48fc-a4bf-e4a2e48b05b8] self._vmops.spawn(context, instance, image_meta, injected_files, [ 689.222014] env[61923]: ERROR nova.compute.manager [instance: 590907a7-b6cc-48fc-a4bf-e4a2e48b05b8] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 689.222014] env[61923]: ERROR nova.compute.manager [instance: 590907a7-b6cc-48fc-a4bf-e4a2e48b05b8] vm_ref = self.build_virtual_machine(instance, [ 689.223540] env[61923]: ERROR nova.compute.manager [instance: 590907a7-b6cc-48fc-a4bf-e4a2e48b05b8] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 689.223540] env[61923]: ERROR nova.compute.manager [instance: 590907a7-b6cc-48fc-a4bf-e4a2e48b05b8] vif_infos = vmwarevif.get_vif_info(self._session, [ 689.223540] env[61923]: ERROR nova.compute.manager [instance: 590907a7-b6cc-48fc-a4bf-e4a2e48b05b8] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 689.223540] env[61923]: ERROR nova.compute.manager [instance: 590907a7-b6cc-48fc-a4bf-e4a2e48b05b8] for vif in network_info: [ 689.223540] env[61923]: ERROR nova.compute.manager [instance: 590907a7-b6cc-48fc-a4bf-e4a2e48b05b8] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 689.223540] env[61923]: ERROR nova.compute.manager [instance: 590907a7-b6cc-48fc-a4bf-e4a2e48b05b8] return self._sync_wrapper(fn, *args, **kwargs) [ 689.223540] env[61923]: ERROR nova.compute.manager [instance: 590907a7-b6cc-48fc-a4bf-e4a2e48b05b8] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 689.223540] env[61923]: ERROR nova.compute.manager [instance: 590907a7-b6cc-48fc-a4bf-e4a2e48b05b8] self.wait() [ 689.223540] env[61923]: ERROR nova.compute.manager [instance: 590907a7-b6cc-48fc-a4bf-e4a2e48b05b8] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 689.223540] env[61923]: ERROR nova.compute.manager [instance: 590907a7-b6cc-48fc-a4bf-e4a2e48b05b8] self[:] = self._gt.wait() [ 689.223540] env[61923]: ERROR nova.compute.manager [instance: 590907a7-b6cc-48fc-a4bf-e4a2e48b05b8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 689.223540] env[61923]: ERROR nova.compute.manager [instance: 590907a7-b6cc-48fc-a4bf-e4a2e48b05b8] return self._exit_event.wait() [ 689.223540] env[61923]: ERROR nova.compute.manager [instance: 590907a7-b6cc-48fc-a4bf-e4a2e48b05b8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 689.224102] env[61923]: ERROR nova.compute.manager [instance: 590907a7-b6cc-48fc-a4bf-e4a2e48b05b8] result = hub.switch() [ 689.224102] env[61923]: ERROR nova.compute.manager [instance: 590907a7-b6cc-48fc-a4bf-e4a2e48b05b8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 689.224102] env[61923]: ERROR nova.compute.manager [instance: 590907a7-b6cc-48fc-a4bf-e4a2e48b05b8] return self.greenlet.switch() [ 689.224102] env[61923]: ERROR nova.compute.manager [instance: 590907a7-b6cc-48fc-a4bf-e4a2e48b05b8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 689.224102] env[61923]: ERROR nova.compute.manager [instance: 590907a7-b6cc-48fc-a4bf-e4a2e48b05b8] result = function(*args, **kwargs) [ 689.224102] env[61923]: ERROR nova.compute.manager [instance: 590907a7-b6cc-48fc-a4bf-e4a2e48b05b8] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 689.224102] env[61923]: ERROR nova.compute.manager [instance: 590907a7-b6cc-48fc-a4bf-e4a2e48b05b8] return func(*args, **kwargs) [ 689.224102] env[61923]: ERROR nova.compute.manager [instance: 590907a7-b6cc-48fc-a4bf-e4a2e48b05b8] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 689.224102] env[61923]: ERROR nova.compute.manager [instance: 590907a7-b6cc-48fc-a4bf-e4a2e48b05b8] raise e [ 689.224102] env[61923]: ERROR nova.compute.manager [instance: 590907a7-b6cc-48fc-a4bf-e4a2e48b05b8] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 689.224102] env[61923]: ERROR nova.compute.manager [instance: 590907a7-b6cc-48fc-a4bf-e4a2e48b05b8] nwinfo = self.network_api.allocate_for_instance( [ 689.224102] env[61923]: ERROR nova.compute.manager [instance: 590907a7-b6cc-48fc-a4bf-e4a2e48b05b8] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 689.224102] env[61923]: ERROR nova.compute.manager [instance: 590907a7-b6cc-48fc-a4bf-e4a2e48b05b8] created_port_ids = self._update_ports_for_instance( [ 689.224652] env[61923]: ERROR nova.compute.manager [instance: 590907a7-b6cc-48fc-a4bf-e4a2e48b05b8] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 689.224652] env[61923]: ERROR nova.compute.manager [instance: 590907a7-b6cc-48fc-a4bf-e4a2e48b05b8] with excutils.save_and_reraise_exception(): [ 689.224652] env[61923]: ERROR nova.compute.manager [instance: 590907a7-b6cc-48fc-a4bf-e4a2e48b05b8] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 689.224652] env[61923]: ERROR nova.compute.manager [instance: 590907a7-b6cc-48fc-a4bf-e4a2e48b05b8] self.force_reraise() [ 689.224652] env[61923]: ERROR nova.compute.manager [instance: 590907a7-b6cc-48fc-a4bf-e4a2e48b05b8] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 689.224652] env[61923]: ERROR nova.compute.manager [instance: 590907a7-b6cc-48fc-a4bf-e4a2e48b05b8] raise self.value [ 689.224652] env[61923]: ERROR nova.compute.manager [instance: 590907a7-b6cc-48fc-a4bf-e4a2e48b05b8] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 689.224652] env[61923]: ERROR nova.compute.manager [instance: 590907a7-b6cc-48fc-a4bf-e4a2e48b05b8] updated_port = self._update_port( [ 689.224652] env[61923]: ERROR nova.compute.manager [instance: 590907a7-b6cc-48fc-a4bf-e4a2e48b05b8] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 689.224652] env[61923]: ERROR nova.compute.manager [instance: 590907a7-b6cc-48fc-a4bf-e4a2e48b05b8] _ensure_no_port_binding_failure(port) [ 689.224652] env[61923]: ERROR nova.compute.manager [instance: 590907a7-b6cc-48fc-a4bf-e4a2e48b05b8] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 689.224652] env[61923]: ERROR nova.compute.manager [instance: 590907a7-b6cc-48fc-a4bf-e4a2e48b05b8] raise exception.PortBindingFailed(port_id=port['id']) [ 689.225187] env[61923]: ERROR nova.compute.manager [instance: 590907a7-b6cc-48fc-a4bf-e4a2e48b05b8] nova.exception.PortBindingFailed: Binding failed for port 989fba39-a679-413f-8cca-904ac748fcbe, please check neutron logs for more information. [ 689.225187] env[61923]: ERROR nova.compute.manager [instance: 590907a7-b6cc-48fc-a4bf-e4a2e48b05b8] [ 689.225187] env[61923]: INFO nova.compute.manager [None req-5f08bfba-63c5-4161-be1b-a884c0ae71af tempest-VolumesAdminNegativeTest-1422120048 tempest-VolumesAdminNegativeTest-1422120048-project-member] [instance: 590907a7-b6cc-48fc-a4bf-e4a2e48b05b8] Terminating instance [ 689.225187] env[61923]: DEBUG oslo_concurrency.lockutils [None req-5f08bfba-63c5-4161-be1b-a884c0ae71af tempest-VolumesAdminNegativeTest-1422120048 tempest-VolumesAdminNegativeTest-1422120048-project-member] Acquiring lock "refresh_cache-590907a7-b6cc-48fc-a4bf-e4a2e48b05b8" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 689.225187] env[61923]: DEBUG oslo_concurrency.lockutils [None req-5f08bfba-63c5-4161-be1b-a884c0ae71af tempest-VolumesAdminNegativeTest-1422120048 tempest-VolumesAdminNegativeTest-1422120048-project-member] Acquired lock "refresh_cache-590907a7-b6cc-48fc-a4bf-e4a2e48b05b8" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 689.225187] env[61923]: DEBUG nova.network.neutron [None req-5f08bfba-63c5-4161-be1b-a884c0ae71af tempest-VolumesAdminNegativeTest-1422120048 tempest-VolumesAdminNegativeTest-1422120048-project-member] [instance: 590907a7-b6cc-48fc-a4bf-e4a2e48b05b8] Building network info cache for instance {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 689.549088] env[61923]: DEBUG nova.compute.utils [None req-08382130-d70f-42e8-a5b5-93e86f5fa49a tempest-AttachVolumeShelveTestJSON-1252849694 tempest-AttachVolumeShelveTestJSON-1252849694-project-member] Using /dev/sd instead of None {{(pid=61923) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 689.554017] env[61923]: DEBUG nova.compute.manager [None req-08382130-d70f-42e8-a5b5-93e86f5fa49a tempest-AttachVolumeShelveTestJSON-1252849694 tempest-AttachVolumeShelveTestJSON-1252849694-project-member] [instance: 60ffc8b0-4a7b-4e2a-8774-d16e502d8a06] Allocating IP information in the background. {{(pid=61923) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 689.554265] env[61923]: DEBUG nova.network.neutron [None req-08382130-d70f-42e8-a5b5-93e86f5fa49a tempest-AttachVolumeShelveTestJSON-1252849694 tempest-AttachVolumeShelveTestJSON-1252849694-project-member] [instance: 60ffc8b0-4a7b-4e2a-8774-d16e502d8a06] allocate_for_instance() {{(pid=61923) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 689.607291] env[61923]: DEBUG oslo_concurrency.lockutils [None req-aed8d24a-43d8-4fbe-ac67-0dd6c37614c4 tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Acquiring lock "4d709a3a-65af-4adf-9ad0-4d830a9518c2" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 689.607518] env[61923]: DEBUG oslo_concurrency.lockutils [None req-aed8d24a-43d8-4fbe-ac67-0dd6c37614c4 tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Lock "4d709a3a-65af-4adf-9ad0-4d830a9518c2" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 689.628821] env[61923]: DEBUG nova.policy [None req-08382130-d70f-42e8-a5b5-93e86f5fa49a tempest-AttachVolumeShelveTestJSON-1252849694 tempest-AttachVolumeShelveTestJSON-1252849694-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'da35e21b822b4ccba98f1a19892fe8f5', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f0dfb7e2632b47f48da6dc458d8f155c', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61923) authorize /opt/stack/nova/nova/policy.py:201}} [ 689.644553] env[61923]: DEBUG oslo_concurrency.lockutils [None req-aed8d24a-43d8-4fbe-ac67-0dd6c37614c4 tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Acquiring lock "fe981195-1e2c-4c8d-8c53-327851f845c7" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 689.644808] env[61923]: DEBUG oslo_concurrency.lockutils [None req-aed8d24a-43d8-4fbe-ac67-0dd6c37614c4 tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Lock "fe981195-1e2c-4c8d-8c53-327851f845c7" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 689.691275] env[61923]: INFO nova.compute.manager [-] [instance: 64584976-b3f3-4da5-a76b-a05cf2ed6aa4] Took 1.04 seconds to deallocate network for instance. [ 689.695206] env[61923]: DEBUG nova.compute.claims [None req-aadda1bb-f770-4742-8c34-cd25e3f7e7cb tempest-InstanceActionsNegativeTestJSON-1700332362 tempest-InstanceActionsNegativeTestJSON-1700332362-project-member] [instance: 64584976-b3f3-4da5-a76b-a05cf2ed6aa4] Aborting claim: {{(pid=61923) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 689.695206] env[61923]: DEBUG oslo_concurrency.lockutils [None req-aadda1bb-f770-4742-8c34-cd25e3f7e7cb tempest-InstanceActionsNegativeTestJSON-1700332362 tempest-InstanceActionsNegativeTestJSON-1700332362-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 689.749567] env[61923]: DEBUG nova.network.neutron [None req-5f08bfba-63c5-4161-be1b-a884c0ae71af tempest-VolumesAdminNegativeTest-1422120048 tempest-VolumesAdminNegativeTest-1422120048-project-member] [instance: 590907a7-b6cc-48fc-a4bf-e4a2e48b05b8] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 689.884362] env[61923]: DEBUG nova.network.neutron [None req-5f08bfba-63c5-4161-be1b-a884c0ae71af tempest-VolumesAdminNegativeTest-1422120048 tempest-VolumesAdminNegativeTest-1422120048-project-member] [instance: 590907a7-b6cc-48fc-a4bf-e4a2e48b05b8] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 690.060019] env[61923]: DEBUG nova.compute.manager [None req-08382130-d70f-42e8-a5b5-93e86f5fa49a tempest-AttachVolumeShelveTestJSON-1252849694 tempest-AttachVolumeShelveTestJSON-1252849694-project-member] [instance: 60ffc8b0-4a7b-4e2a-8774-d16e502d8a06] Start building block device mappings for instance. {{(pid=61923) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 690.243246] env[61923]: DEBUG nova.network.neutron [None req-08382130-d70f-42e8-a5b5-93e86f5fa49a tempest-AttachVolumeShelveTestJSON-1252849694 tempest-AttachVolumeShelveTestJSON-1252849694-project-member] [instance: 60ffc8b0-4a7b-4e2a-8774-d16e502d8a06] Successfully created port: c88b194c-774a-448e-8f31-5a89cfdbc187 {{(pid=61923) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 690.391949] env[61923]: DEBUG oslo_concurrency.lockutils [None req-5f08bfba-63c5-4161-be1b-a884c0ae71af tempest-VolumesAdminNegativeTest-1422120048 tempest-VolumesAdminNegativeTest-1422120048-project-member] Releasing lock "refresh_cache-590907a7-b6cc-48fc-a4bf-e4a2e48b05b8" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 690.391949] env[61923]: DEBUG nova.compute.manager [None req-5f08bfba-63c5-4161-be1b-a884c0ae71af tempest-VolumesAdminNegativeTest-1422120048 tempest-VolumesAdminNegativeTest-1422120048-project-member] [instance: 590907a7-b6cc-48fc-a4bf-e4a2e48b05b8] Start destroying the instance on the hypervisor. {{(pid=61923) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 690.391949] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-5f08bfba-63c5-4161-be1b-a884c0ae71af tempest-VolumesAdminNegativeTest-1422120048 tempest-VolumesAdminNegativeTest-1422120048-project-member] [instance: 590907a7-b6cc-48fc-a4bf-e4a2e48b05b8] Destroying instance {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 690.391949] env[61923]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-79a532d1-f919-4723-909d-9ec05863b42c {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.407020] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-652803b3-6e12-442b-b06c-bf8702d7531f {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.432439] env[61923]: WARNING nova.virt.vmwareapi.vmops [None req-5f08bfba-63c5-4161-be1b-a884c0ae71af tempest-VolumesAdminNegativeTest-1422120048 tempest-VolumesAdminNegativeTest-1422120048-project-member] [instance: 590907a7-b6cc-48fc-a4bf-e4a2e48b05b8] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 590907a7-b6cc-48fc-a4bf-e4a2e48b05b8 could not be found. [ 690.432919] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-5f08bfba-63c5-4161-be1b-a884c0ae71af tempest-VolumesAdminNegativeTest-1422120048 tempest-VolumesAdminNegativeTest-1422120048-project-member] [instance: 590907a7-b6cc-48fc-a4bf-e4a2e48b05b8] Instance destroyed {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 690.433179] env[61923]: INFO nova.compute.manager [None req-5f08bfba-63c5-4161-be1b-a884c0ae71af tempest-VolumesAdminNegativeTest-1422120048 tempest-VolumesAdminNegativeTest-1422120048-project-member] [instance: 590907a7-b6cc-48fc-a4bf-e4a2e48b05b8] Took 0.04 seconds to destroy the instance on the hypervisor. [ 690.433596] env[61923]: DEBUG oslo.service.loopingcall [None req-5f08bfba-63c5-4161-be1b-a884c0ae71af tempest-VolumesAdminNegativeTest-1422120048 tempest-VolumesAdminNegativeTest-1422120048-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61923) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 690.433861] env[61923]: DEBUG nova.compute.manager [-] [instance: 590907a7-b6cc-48fc-a4bf-e4a2e48b05b8] Deallocating network for instance {{(pid=61923) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 690.433967] env[61923]: DEBUG nova.network.neutron [-] [instance: 590907a7-b6cc-48fc-a4bf-e4a2e48b05b8] deallocate_for_instance() {{(pid=61923) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 690.474474] env[61923]: DEBUG nova.network.neutron [-] [instance: 590907a7-b6cc-48fc-a4bf-e4a2e48b05b8] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 690.487108] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a483f39f-e8eb-491b-9799-3791af67ec84 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.495578] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fcdc9ef4-2a10-4229-8902-51481dcec564 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.531411] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7332b99a-9137-4fb8-ae4e-e74c2c7e57a5 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.538744] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5277710b-a9ab-434c-9c43-cbc15d6b140a {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 690.553039] env[61923]: DEBUG nova.compute.provider_tree [None req-6833b601-845b-439d-ad97-e2f9c0927775 tempest-ServerRescueNegativeTestJSON-1435265126 tempest-ServerRescueNegativeTestJSON-1435265126-project-member] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 690.661669] env[61923]: DEBUG nova.compute.manager [req-f0c2159d-f5fd-429f-a1c6-88b1136cea80 req-5a4c5e12-2ca1-47c8-a77f-314a3a3a3163 service nova] [instance: 590907a7-b6cc-48fc-a4bf-e4a2e48b05b8] Received event network-changed-989fba39-a679-413f-8cca-904ac748fcbe {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 690.661669] env[61923]: DEBUG nova.compute.manager [req-f0c2159d-f5fd-429f-a1c6-88b1136cea80 req-5a4c5e12-2ca1-47c8-a77f-314a3a3a3163 service nova] [instance: 590907a7-b6cc-48fc-a4bf-e4a2e48b05b8] Refreshing instance network info cache due to event network-changed-989fba39-a679-413f-8cca-904ac748fcbe. {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 690.661669] env[61923]: DEBUG oslo_concurrency.lockutils [req-f0c2159d-f5fd-429f-a1c6-88b1136cea80 req-5a4c5e12-2ca1-47c8-a77f-314a3a3a3163 service nova] Acquiring lock "refresh_cache-590907a7-b6cc-48fc-a4bf-e4a2e48b05b8" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 690.661669] env[61923]: DEBUG oslo_concurrency.lockutils [req-f0c2159d-f5fd-429f-a1c6-88b1136cea80 req-5a4c5e12-2ca1-47c8-a77f-314a3a3a3163 service nova] Acquired lock "refresh_cache-590907a7-b6cc-48fc-a4bf-e4a2e48b05b8" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 690.661669] env[61923]: DEBUG nova.network.neutron [req-f0c2159d-f5fd-429f-a1c6-88b1136cea80 req-5a4c5e12-2ca1-47c8-a77f-314a3a3a3163 service nova] [instance: 590907a7-b6cc-48fc-a4bf-e4a2e48b05b8] Refreshing network info cache for port 989fba39-a679-413f-8cca-904ac748fcbe {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 690.977711] env[61923]: DEBUG nova.network.neutron [-] [instance: 590907a7-b6cc-48fc-a4bf-e4a2e48b05b8] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 691.059576] env[61923]: DEBUG nova.scheduler.client.report [None req-6833b601-845b-439d-ad97-e2f9c0927775 tempest-ServerRescueNegativeTestJSON-1435265126 tempest-ServerRescueNegativeTestJSON-1435265126-project-member] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 691.075322] env[61923]: DEBUG nova.compute.manager [None req-08382130-d70f-42e8-a5b5-93e86f5fa49a tempest-AttachVolumeShelveTestJSON-1252849694 tempest-AttachVolumeShelveTestJSON-1252849694-project-member] [instance: 60ffc8b0-4a7b-4e2a-8774-d16e502d8a06] Start spawning the instance on the hypervisor. {{(pid=61923) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 691.105523] env[61923]: DEBUG nova.virt.hardware [None req-08382130-d70f-42e8-a5b5-93e86f5fa49a tempest-AttachVolumeShelveTestJSON-1252849694 tempest-AttachVolumeShelveTestJSON-1252849694-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-29T20:07:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-29T20:07:35Z,direct_url=,disk_format='vmdk',id=0f153f63-ae0a-45b1-b7f5-7f9b673c947f,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='4e7b5e3b3c3443c8bd5c70f8a15739f5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-29T20:07:36Z,virtual_size=,visibility=), allow threads: False {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 691.105780] env[61923]: DEBUG nova.virt.hardware [None req-08382130-d70f-42e8-a5b5-93e86f5fa49a tempest-AttachVolumeShelveTestJSON-1252849694 tempest-AttachVolumeShelveTestJSON-1252849694-project-member] Flavor limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 691.106022] env[61923]: DEBUG nova.virt.hardware [None req-08382130-d70f-42e8-a5b5-93e86f5fa49a tempest-AttachVolumeShelveTestJSON-1252849694 tempest-AttachVolumeShelveTestJSON-1252849694-project-member] Image limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 691.106138] env[61923]: DEBUG nova.virt.hardware [None req-08382130-d70f-42e8-a5b5-93e86f5fa49a tempest-AttachVolumeShelveTestJSON-1252849694 tempest-AttachVolumeShelveTestJSON-1252849694-project-member] Flavor pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 691.106423] env[61923]: DEBUG nova.virt.hardware [None req-08382130-d70f-42e8-a5b5-93e86f5fa49a tempest-AttachVolumeShelveTestJSON-1252849694 tempest-AttachVolumeShelveTestJSON-1252849694-project-member] Image pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 691.106423] env[61923]: DEBUG nova.virt.hardware [None req-08382130-d70f-42e8-a5b5-93e86f5fa49a tempest-AttachVolumeShelveTestJSON-1252849694 tempest-AttachVolumeShelveTestJSON-1252849694-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 691.106816] env[61923]: DEBUG nova.virt.hardware [None req-08382130-d70f-42e8-a5b5-93e86f5fa49a tempest-AttachVolumeShelveTestJSON-1252849694 tempest-AttachVolumeShelveTestJSON-1252849694-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 691.107050] env[61923]: DEBUG nova.virt.hardware [None req-08382130-d70f-42e8-a5b5-93e86f5fa49a tempest-AttachVolumeShelveTestJSON-1252849694 tempest-AttachVolumeShelveTestJSON-1252849694-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 691.107225] env[61923]: DEBUG nova.virt.hardware [None req-08382130-d70f-42e8-a5b5-93e86f5fa49a tempest-AttachVolumeShelveTestJSON-1252849694 tempest-AttachVolumeShelveTestJSON-1252849694-project-member] Got 1 possible topologies {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 691.107387] env[61923]: DEBUG nova.virt.hardware [None req-08382130-d70f-42e8-a5b5-93e86f5fa49a tempest-AttachVolumeShelveTestJSON-1252849694 tempest-AttachVolumeShelveTestJSON-1252849694-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 691.107866] env[61923]: DEBUG nova.virt.hardware [None req-08382130-d70f-42e8-a5b5-93e86f5fa49a tempest-AttachVolumeShelveTestJSON-1252849694 tempest-AttachVolumeShelveTestJSON-1252849694-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 691.108940] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3eee8d3b-965d-4053-a0a2-24ec68319ad0 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.122028] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24856418-640a-4c72-9185-ac92b7f8f221 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 691.181414] env[61923]: DEBUG nova.network.neutron [req-f0c2159d-f5fd-429f-a1c6-88b1136cea80 req-5a4c5e12-2ca1-47c8-a77f-314a3a3a3163 service nova] [instance: 590907a7-b6cc-48fc-a4bf-e4a2e48b05b8] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 691.314946] env[61923]: DEBUG nova.network.neutron [req-f0c2159d-f5fd-429f-a1c6-88b1136cea80 req-5a4c5e12-2ca1-47c8-a77f-314a3a3a3163 service nova] [instance: 590907a7-b6cc-48fc-a4bf-e4a2e48b05b8] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 691.480384] env[61923]: INFO nova.compute.manager [-] [instance: 590907a7-b6cc-48fc-a4bf-e4a2e48b05b8] Took 1.05 seconds to deallocate network for instance. [ 691.482586] env[61923]: DEBUG nova.compute.claims [None req-5f08bfba-63c5-4161-be1b-a884c0ae71af tempest-VolumesAdminNegativeTest-1422120048 tempest-VolumesAdminNegativeTest-1422120048-project-member] [instance: 590907a7-b6cc-48fc-a4bf-e4a2e48b05b8] Aborting claim: {{(pid=61923) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 691.482916] env[61923]: DEBUG oslo_concurrency.lockutils [None req-5f08bfba-63c5-4161-be1b-a884c0ae71af tempest-VolumesAdminNegativeTest-1422120048 tempest-VolumesAdminNegativeTest-1422120048-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 691.564933] env[61923]: DEBUG oslo_concurrency.lockutils [None req-6833b601-845b-439d-ad97-e2f9c0927775 tempest-ServerRescueNegativeTestJSON-1435265126 tempest-ServerRescueNegativeTestJSON-1435265126-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.522s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 691.565569] env[61923]: DEBUG nova.compute.manager [None req-6833b601-845b-439d-ad97-e2f9c0927775 tempest-ServerRescueNegativeTestJSON-1435265126 tempest-ServerRescueNegativeTestJSON-1435265126-project-member] [instance: 4f5348ec-f0b6-45e2-be04-31dbf1d49a3a] Start building networks asynchronously for instance. {{(pid=61923) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 691.568143] env[61923]: DEBUG oslo_concurrency.lockutils [None req-859ffdff-aaee-46b4-bdd6-0a15cce9792f tempest-ServerRescueNegativeTestJSON-1435265126 tempest-ServerRescueNegativeTestJSON-1435265126-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 26.466s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 691.817384] env[61923]: DEBUG oslo_concurrency.lockutils [req-f0c2159d-f5fd-429f-a1c6-88b1136cea80 req-5a4c5e12-2ca1-47c8-a77f-314a3a3a3163 service nova] Releasing lock "refresh_cache-590907a7-b6cc-48fc-a4bf-e4a2e48b05b8" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 691.817711] env[61923]: DEBUG nova.compute.manager [req-f0c2159d-f5fd-429f-a1c6-88b1136cea80 req-5a4c5e12-2ca1-47c8-a77f-314a3a3a3163 service nova] [instance: 590907a7-b6cc-48fc-a4bf-e4a2e48b05b8] Received event network-vif-deleted-989fba39-a679-413f-8cca-904ac748fcbe {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 691.901060] env[61923]: ERROR nova.compute.manager [None req-08382130-d70f-42e8-a5b5-93e86f5fa49a tempest-AttachVolumeShelveTestJSON-1252849694 tempest-AttachVolumeShelveTestJSON-1252849694-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port c88b194c-774a-448e-8f31-5a89cfdbc187, please check neutron logs for more information. [ 691.901060] env[61923]: ERROR nova.compute.manager Traceback (most recent call last): [ 691.901060] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 691.901060] env[61923]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 691.901060] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 691.901060] env[61923]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 691.901060] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 691.901060] env[61923]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 691.901060] env[61923]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 691.901060] env[61923]: ERROR nova.compute.manager self.force_reraise() [ 691.901060] env[61923]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 691.901060] env[61923]: ERROR nova.compute.manager raise self.value [ 691.901060] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 691.901060] env[61923]: ERROR nova.compute.manager updated_port = self._update_port( [ 691.901060] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 691.901060] env[61923]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 691.901482] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 691.901482] env[61923]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 691.901482] env[61923]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port c88b194c-774a-448e-8f31-5a89cfdbc187, please check neutron logs for more information. [ 691.901482] env[61923]: ERROR nova.compute.manager [ 691.901482] env[61923]: Traceback (most recent call last): [ 691.901482] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 691.901482] env[61923]: listener.cb(fileno) [ 691.901482] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 691.901482] env[61923]: result = function(*args, **kwargs) [ 691.901482] env[61923]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 691.901482] env[61923]: return func(*args, **kwargs) [ 691.901482] env[61923]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 691.901482] env[61923]: raise e [ 691.901482] env[61923]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 691.901482] env[61923]: nwinfo = self.network_api.allocate_for_instance( [ 691.901482] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 691.901482] env[61923]: created_port_ids = self._update_ports_for_instance( [ 691.901482] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 691.901482] env[61923]: with excutils.save_and_reraise_exception(): [ 691.901482] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 691.901482] env[61923]: self.force_reraise() [ 691.901482] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 691.901482] env[61923]: raise self.value [ 691.901482] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 691.901482] env[61923]: updated_port = self._update_port( [ 691.901482] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 691.901482] env[61923]: _ensure_no_port_binding_failure(port) [ 691.901482] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 691.901482] env[61923]: raise exception.PortBindingFailed(port_id=port['id']) [ 691.902167] env[61923]: nova.exception.PortBindingFailed: Binding failed for port c88b194c-774a-448e-8f31-5a89cfdbc187, please check neutron logs for more information. [ 691.902167] env[61923]: Removing descriptor: 14 [ 691.902167] env[61923]: ERROR nova.compute.manager [None req-08382130-d70f-42e8-a5b5-93e86f5fa49a tempest-AttachVolumeShelveTestJSON-1252849694 tempest-AttachVolumeShelveTestJSON-1252849694-project-member] [instance: 60ffc8b0-4a7b-4e2a-8774-d16e502d8a06] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port c88b194c-774a-448e-8f31-5a89cfdbc187, please check neutron logs for more information. [ 691.902167] env[61923]: ERROR nova.compute.manager [instance: 60ffc8b0-4a7b-4e2a-8774-d16e502d8a06] Traceback (most recent call last): [ 691.902167] env[61923]: ERROR nova.compute.manager [instance: 60ffc8b0-4a7b-4e2a-8774-d16e502d8a06] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 691.902167] env[61923]: ERROR nova.compute.manager [instance: 60ffc8b0-4a7b-4e2a-8774-d16e502d8a06] yield resources [ 691.902167] env[61923]: ERROR nova.compute.manager [instance: 60ffc8b0-4a7b-4e2a-8774-d16e502d8a06] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 691.902167] env[61923]: ERROR nova.compute.manager [instance: 60ffc8b0-4a7b-4e2a-8774-d16e502d8a06] self.driver.spawn(context, instance, image_meta, [ 691.902167] env[61923]: ERROR nova.compute.manager [instance: 60ffc8b0-4a7b-4e2a-8774-d16e502d8a06] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 691.902167] env[61923]: ERROR nova.compute.manager [instance: 60ffc8b0-4a7b-4e2a-8774-d16e502d8a06] self._vmops.spawn(context, instance, image_meta, injected_files, [ 691.902167] env[61923]: ERROR nova.compute.manager [instance: 60ffc8b0-4a7b-4e2a-8774-d16e502d8a06] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 691.902167] env[61923]: ERROR nova.compute.manager [instance: 60ffc8b0-4a7b-4e2a-8774-d16e502d8a06] vm_ref = self.build_virtual_machine(instance, [ 691.902516] env[61923]: ERROR nova.compute.manager [instance: 60ffc8b0-4a7b-4e2a-8774-d16e502d8a06] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 691.902516] env[61923]: ERROR nova.compute.manager [instance: 60ffc8b0-4a7b-4e2a-8774-d16e502d8a06] vif_infos = vmwarevif.get_vif_info(self._session, [ 691.902516] env[61923]: ERROR nova.compute.manager [instance: 60ffc8b0-4a7b-4e2a-8774-d16e502d8a06] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 691.902516] env[61923]: ERROR nova.compute.manager [instance: 60ffc8b0-4a7b-4e2a-8774-d16e502d8a06] for vif in network_info: [ 691.902516] env[61923]: ERROR nova.compute.manager [instance: 60ffc8b0-4a7b-4e2a-8774-d16e502d8a06] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 691.902516] env[61923]: ERROR nova.compute.manager [instance: 60ffc8b0-4a7b-4e2a-8774-d16e502d8a06] return self._sync_wrapper(fn, *args, **kwargs) [ 691.902516] env[61923]: ERROR nova.compute.manager [instance: 60ffc8b0-4a7b-4e2a-8774-d16e502d8a06] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 691.902516] env[61923]: ERROR nova.compute.manager [instance: 60ffc8b0-4a7b-4e2a-8774-d16e502d8a06] self.wait() [ 691.902516] env[61923]: ERROR nova.compute.manager [instance: 60ffc8b0-4a7b-4e2a-8774-d16e502d8a06] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 691.902516] env[61923]: ERROR nova.compute.manager [instance: 60ffc8b0-4a7b-4e2a-8774-d16e502d8a06] self[:] = self._gt.wait() [ 691.902516] env[61923]: ERROR nova.compute.manager [instance: 60ffc8b0-4a7b-4e2a-8774-d16e502d8a06] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 691.902516] env[61923]: ERROR nova.compute.manager [instance: 60ffc8b0-4a7b-4e2a-8774-d16e502d8a06] return self._exit_event.wait() [ 691.902516] env[61923]: ERROR nova.compute.manager [instance: 60ffc8b0-4a7b-4e2a-8774-d16e502d8a06] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 691.902886] env[61923]: ERROR nova.compute.manager [instance: 60ffc8b0-4a7b-4e2a-8774-d16e502d8a06] result = hub.switch() [ 691.902886] env[61923]: ERROR nova.compute.manager [instance: 60ffc8b0-4a7b-4e2a-8774-d16e502d8a06] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 691.902886] env[61923]: ERROR nova.compute.manager [instance: 60ffc8b0-4a7b-4e2a-8774-d16e502d8a06] return self.greenlet.switch() [ 691.902886] env[61923]: ERROR nova.compute.manager [instance: 60ffc8b0-4a7b-4e2a-8774-d16e502d8a06] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 691.902886] env[61923]: ERROR nova.compute.manager [instance: 60ffc8b0-4a7b-4e2a-8774-d16e502d8a06] result = function(*args, **kwargs) [ 691.902886] env[61923]: ERROR nova.compute.manager [instance: 60ffc8b0-4a7b-4e2a-8774-d16e502d8a06] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 691.902886] env[61923]: ERROR nova.compute.manager [instance: 60ffc8b0-4a7b-4e2a-8774-d16e502d8a06] return func(*args, **kwargs) [ 691.902886] env[61923]: ERROR nova.compute.manager [instance: 60ffc8b0-4a7b-4e2a-8774-d16e502d8a06] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 691.902886] env[61923]: ERROR nova.compute.manager [instance: 60ffc8b0-4a7b-4e2a-8774-d16e502d8a06] raise e [ 691.902886] env[61923]: ERROR nova.compute.manager [instance: 60ffc8b0-4a7b-4e2a-8774-d16e502d8a06] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 691.902886] env[61923]: ERROR nova.compute.manager [instance: 60ffc8b0-4a7b-4e2a-8774-d16e502d8a06] nwinfo = self.network_api.allocate_for_instance( [ 691.902886] env[61923]: ERROR nova.compute.manager [instance: 60ffc8b0-4a7b-4e2a-8774-d16e502d8a06] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 691.902886] env[61923]: ERROR nova.compute.manager [instance: 60ffc8b0-4a7b-4e2a-8774-d16e502d8a06] created_port_ids = self._update_ports_for_instance( [ 691.903310] env[61923]: ERROR nova.compute.manager [instance: 60ffc8b0-4a7b-4e2a-8774-d16e502d8a06] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 691.903310] env[61923]: ERROR nova.compute.manager [instance: 60ffc8b0-4a7b-4e2a-8774-d16e502d8a06] with excutils.save_and_reraise_exception(): [ 691.903310] env[61923]: ERROR nova.compute.manager [instance: 60ffc8b0-4a7b-4e2a-8774-d16e502d8a06] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 691.903310] env[61923]: ERROR nova.compute.manager [instance: 60ffc8b0-4a7b-4e2a-8774-d16e502d8a06] self.force_reraise() [ 691.903310] env[61923]: ERROR nova.compute.manager [instance: 60ffc8b0-4a7b-4e2a-8774-d16e502d8a06] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 691.903310] env[61923]: ERROR nova.compute.manager [instance: 60ffc8b0-4a7b-4e2a-8774-d16e502d8a06] raise self.value [ 691.903310] env[61923]: ERROR nova.compute.manager [instance: 60ffc8b0-4a7b-4e2a-8774-d16e502d8a06] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 691.903310] env[61923]: ERROR nova.compute.manager [instance: 60ffc8b0-4a7b-4e2a-8774-d16e502d8a06] updated_port = self._update_port( [ 691.903310] env[61923]: ERROR nova.compute.manager [instance: 60ffc8b0-4a7b-4e2a-8774-d16e502d8a06] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 691.903310] env[61923]: ERROR nova.compute.manager [instance: 60ffc8b0-4a7b-4e2a-8774-d16e502d8a06] _ensure_no_port_binding_failure(port) [ 691.903310] env[61923]: ERROR nova.compute.manager [instance: 60ffc8b0-4a7b-4e2a-8774-d16e502d8a06] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 691.903310] env[61923]: ERROR nova.compute.manager [instance: 60ffc8b0-4a7b-4e2a-8774-d16e502d8a06] raise exception.PortBindingFailed(port_id=port['id']) [ 691.903660] env[61923]: ERROR nova.compute.manager [instance: 60ffc8b0-4a7b-4e2a-8774-d16e502d8a06] nova.exception.PortBindingFailed: Binding failed for port c88b194c-774a-448e-8f31-5a89cfdbc187, please check neutron logs for more information. [ 691.903660] env[61923]: ERROR nova.compute.manager [instance: 60ffc8b0-4a7b-4e2a-8774-d16e502d8a06] [ 691.903660] env[61923]: INFO nova.compute.manager [None req-08382130-d70f-42e8-a5b5-93e86f5fa49a tempest-AttachVolumeShelveTestJSON-1252849694 tempest-AttachVolumeShelveTestJSON-1252849694-project-member] [instance: 60ffc8b0-4a7b-4e2a-8774-d16e502d8a06] Terminating instance [ 691.904425] env[61923]: DEBUG oslo_concurrency.lockutils [None req-08382130-d70f-42e8-a5b5-93e86f5fa49a tempest-AttachVolumeShelveTestJSON-1252849694 tempest-AttachVolumeShelveTestJSON-1252849694-project-member] Acquiring lock "refresh_cache-60ffc8b0-4a7b-4e2a-8774-d16e502d8a06" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 691.904586] env[61923]: DEBUG oslo_concurrency.lockutils [None req-08382130-d70f-42e8-a5b5-93e86f5fa49a tempest-AttachVolumeShelveTestJSON-1252849694 tempest-AttachVolumeShelveTestJSON-1252849694-project-member] Acquired lock "refresh_cache-60ffc8b0-4a7b-4e2a-8774-d16e502d8a06" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 691.904747] env[61923]: DEBUG nova.network.neutron [None req-08382130-d70f-42e8-a5b5-93e86f5fa49a tempest-AttachVolumeShelveTestJSON-1252849694 tempest-AttachVolumeShelveTestJSON-1252849694-project-member] [instance: 60ffc8b0-4a7b-4e2a-8774-d16e502d8a06] Building network info cache for instance {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 692.075170] env[61923]: DEBUG nova.compute.utils [None req-6833b601-845b-439d-ad97-e2f9c0927775 tempest-ServerRescueNegativeTestJSON-1435265126 tempest-ServerRescueNegativeTestJSON-1435265126-project-member] Using /dev/sd instead of None {{(pid=61923) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 692.076779] env[61923]: DEBUG nova.compute.manager [None req-6833b601-845b-439d-ad97-e2f9c0927775 tempest-ServerRescueNegativeTestJSON-1435265126 tempest-ServerRescueNegativeTestJSON-1435265126-project-member] [instance: 4f5348ec-f0b6-45e2-be04-31dbf1d49a3a] Allocating IP information in the background. {{(pid=61923) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 692.076852] env[61923]: DEBUG nova.network.neutron [None req-6833b601-845b-439d-ad97-e2f9c0927775 tempest-ServerRescueNegativeTestJSON-1435265126 tempest-ServerRescueNegativeTestJSON-1435265126-project-member] [instance: 4f5348ec-f0b6-45e2-be04-31dbf1d49a3a] allocate_for_instance() {{(pid=61923) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 692.141602] env[61923]: DEBUG nova.policy [None req-6833b601-845b-439d-ad97-e2f9c0927775 tempest-ServerRescueNegativeTestJSON-1435265126 tempest-ServerRescueNegativeTestJSON-1435265126-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'de8bf84e21494593be329e3dd63da127', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '62d9e2aeadec483e881ea244b439d557', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61923) authorize /opt/stack/nova/nova/policy.py:201}} [ 692.409615] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c1f9b59-9efe-480d-b0f9-9e2c13c42654 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.421910] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68de4261-5ae3-4e0c-be62-04708bb7fd87 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.455827] env[61923]: DEBUG nova.network.neutron [None req-08382130-d70f-42e8-a5b5-93e86f5fa49a tempest-AttachVolumeShelveTestJSON-1252849694 tempest-AttachVolumeShelveTestJSON-1252849694-project-member] [instance: 60ffc8b0-4a7b-4e2a-8774-d16e502d8a06] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 692.458066] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a12fd39-29ed-4851-8290-a7ab441b372d {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.465796] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-363e1d68-daac-48a1-82b9-b729682f95e5 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.479053] env[61923]: DEBUG nova.compute.provider_tree [None req-859ffdff-aaee-46b4-bdd6-0a15cce9792f tempest-ServerRescueNegativeTestJSON-1435265126 tempest-ServerRescueNegativeTestJSON-1435265126-project-member] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 692.505303] env[61923]: DEBUG nova.network.neutron [None req-6833b601-845b-439d-ad97-e2f9c0927775 tempest-ServerRescueNegativeTestJSON-1435265126 tempest-ServerRescueNegativeTestJSON-1435265126-project-member] [instance: 4f5348ec-f0b6-45e2-be04-31dbf1d49a3a] Successfully created port: a616fa04-1fe8-475c-9237-a801d7430538 {{(pid=61923) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 692.579310] env[61923]: DEBUG nova.compute.manager [None req-6833b601-845b-439d-ad97-e2f9c0927775 tempest-ServerRescueNegativeTestJSON-1435265126 tempest-ServerRescueNegativeTestJSON-1435265126-project-member] [instance: 4f5348ec-f0b6-45e2-be04-31dbf1d49a3a] Start building block device mappings for instance. {{(pid=61923) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 692.643033] env[61923]: DEBUG nova.network.neutron [None req-08382130-d70f-42e8-a5b5-93e86f5fa49a tempest-AttachVolumeShelveTestJSON-1252849694 tempest-AttachVolumeShelveTestJSON-1252849694-project-member] [instance: 60ffc8b0-4a7b-4e2a-8774-d16e502d8a06] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 692.699703] env[61923]: DEBUG nova.compute.manager [req-1dcab4c9-cedc-4382-86c9-e41adb92550c req-43b08bb8-5794-4d01-a9ef-589072a5838d service nova] [instance: 60ffc8b0-4a7b-4e2a-8774-d16e502d8a06] Received event network-changed-c88b194c-774a-448e-8f31-5a89cfdbc187 {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 692.700116] env[61923]: DEBUG nova.compute.manager [req-1dcab4c9-cedc-4382-86c9-e41adb92550c req-43b08bb8-5794-4d01-a9ef-589072a5838d service nova] [instance: 60ffc8b0-4a7b-4e2a-8774-d16e502d8a06] Refreshing instance network info cache due to event network-changed-c88b194c-774a-448e-8f31-5a89cfdbc187. {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 692.700240] env[61923]: DEBUG oslo_concurrency.lockutils [req-1dcab4c9-cedc-4382-86c9-e41adb92550c req-43b08bb8-5794-4d01-a9ef-589072a5838d service nova] Acquiring lock "refresh_cache-60ffc8b0-4a7b-4e2a-8774-d16e502d8a06" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 692.984046] env[61923]: DEBUG nova.scheduler.client.report [None req-859ffdff-aaee-46b4-bdd6-0a15cce9792f tempest-ServerRescueNegativeTestJSON-1435265126 tempest-ServerRescueNegativeTestJSON-1435265126-project-member] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 693.145520] env[61923]: DEBUG oslo_concurrency.lockutils [None req-08382130-d70f-42e8-a5b5-93e86f5fa49a tempest-AttachVolumeShelveTestJSON-1252849694 tempest-AttachVolumeShelveTestJSON-1252849694-project-member] Releasing lock "refresh_cache-60ffc8b0-4a7b-4e2a-8774-d16e502d8a06" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 693.145963] env[61923]: DEBUG nova.compute.manager [None req-08382130-d70f-42e8-a5b5-93e86f5fa49a tempest-AttachVolumeShelveTestJSON-1252849694 tempest-AttachVolumeShelveTestJSON-1252849694-project-member] [instance: 60ffc8b0-4a7b-4e2a-8774-d16e502d8a06] Start destroying the instance on the hypervisor. {{(pid=61923) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 693.146176] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-08382130-d70f-42e8-a5b5-93e86f5fa49a tempest-AttachVolumeShelveTestJSON-1252849694 tempest-AttachVolumeShelveTestJSON-1252849694-project-member] [instance: 60ffc8b0-4a7b-4e2a-8774-d16e502d8a06] Destroying instance {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 693.147758] env[61923]: DEBUG oslo_concurrency.lockutils [req-1dcab4c9-cedc-4382-86c9-e41adb92550c req-43b08bb8-5794-4d01-a9ef-589072a5838d service nova] Acquired lock "refresh_cache-60ffc8b0-4a7b-4e2a-8774-d16e502d8a06" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 693.149066] env[61923]: DEBUG nova.network.neutron [req-1dcab4c9-cedc-4382-86c9-e41adb92550c req-43b08bb8-5794-4d01-a9ef-589072a5838d service nova] [instance: 60ffc8b0-4a7b-4e2a-8774-d16e502d8a06] Refreshing network info cache for port c88b194c-774a-448e-8f31-5a89cfdbc187 {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 693.150701] env[61923]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-7792f418-3d61-4bd1-a583-e4edc0871f2b {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.161325] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dbcb7c79-28fe-42f3-9351-03bfd9c056f9 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.185873] env[61923]: WARNING nova.virt.vmwareapi.vmops [None req-08382130-d70f-42e8-a5b5-93e86f5fa49a tempest-AttachVolumeShelveTestJSON-1252849694 tempest-AttachVolumeShelveTestJSON-1252849694-project-member] [instance: 60ffc8b0-4a7b-4e2a-8774-d16e502d8a06] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 60ffc8b0-4a7b-4e2a-8774-d16e502d8a06 could not be found. [ 693.186113] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-08382130-d70f-42e8-a5b5-93e86f5fa49a tempest-AttachVolumeShelveTestJSON-1252849694 tempest-AttachVolumeShelveTestJSON-1252849694-project-member] [instance: 60ffc8b0-4a7b-4e2a-8774-d16e502d8a06] Instance destroyed {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 693.186292] env[61923]: INFO nova.compute.manager [None req-08382130-d70f-42e8-a5b5-93e86f5fa49a tempest-AttachVolumeShelveTestJSON-1252849694 tempest-AttachVolumeShelveTestJSON-1252849694-project-member] [instance: 60ffc8b0-4a7b-4e2a-8774-d16e502d8a06] Took 0.04 seconds to destroy the instance on the hypervisor. [ 693.186524] env[61923]: DEBUG oslo.service.loopingcall [None req-08382130-d70f-42e8-a5b5-93e86f5fa49a tempest-AttachVolumeShelveTestJSON-1252849694 tempest-AttachVolumeShelveTestJSON-1252849694-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61923) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 693.186736] env[61923]: DEBUG nova.compute.manager [-] [instance: 60ffc8b0-4a7b-4e2a-8774-d16e502d8a06] Deallocating network for instance {{(pid=61923) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 693.186824] env[61923]: DEBUG nova.network.neutron [-] [instance: 60ffc8b0-4a7b-4e2a-8774-d16e502d8a06] deallocate_for_instance() {{(pid=61923) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 693.221969] env[61923]: DEBUG nova.network.neutron [-] [instance: 60ffc8b0-4a7b-4e2a-8774-d16e502d8a06] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 693.488288] env[61923]: DEBUG oslo_concurrency.lockutils [None req-859ffdff-aaee-46b4-bdd6-0a15cce9792f tempest-ServerRescueNegativeTestJSON-1435265126 tempest-ServerRescueNegativeTestJSON-1435265126-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.920s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 693.488976] env[61923]: ERROR nova.compute.manager [None req-859ffdff-aaee-46b4-bdd6-0a15cce9792f tempest-ServerRescueNegativeTestJSON-1435265126 tempest-ServerRescueNegativeTestJSON-1435265126-project-member] [instance: 41b6744e-d548-442c-b6cc-e12bd403f229] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 32a4c72c-9bfd-4f5f-a3e0-6a0ca4c9b6d2, please check neutron logs for more information. [ 693.488976] env[61923]: ERROR nova.compute.manager [instance: 41b6744e-d548-442c-b6cc-e12bd403f229] Traceback (most recent call last): [ 693.488976] env[61923]: ERROR nova.compute.manager [instance: 41b6744e-d548-442c-b6cc-e12bd403f229] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 693.488976] env[61923]: ERROR nova.compute.manager [instance: 41b6744e-d548-442c-b6cc-e12bd403f229] self.driver.spawn(context, instance, image_meta, [ 693.488976] env[61923]: ERROR nova.compute.manager [instance: 41b6744e-d548-442c-b6cc-e12bd403f229] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 693.488976] env[61923]: ERROR nova.compute.manager [instance: 41b6744e-d548-442c-b6cc-e12bd403f229] self._vmops.spawn(context, instance, image_meta, injected_files, [ 693.488976] env[61923]: ERROR nova.compute.manager [instance: 41b6744e-d548-442c-b6cc-e12bd403f229] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 693.488976] env[61923]: ERROR nova.compute.manager [instance: 41b6744e-d548-442c-b6cc-e12bd403f229] vm_ref = self.build_virtual_machine(instance, [ 693.488976] env[61923]: ERROR nova.compute.manager [instance: 41b6744e-d548-442c-b6cc-e12bd403f229] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 693.488976] env[61923]: ERROR nova.compute.manager [instance: 41b6744e-d548-442c-b6cc-e12bd403f229] vif_infos = vmwarevif.get_vif_info(self._session, [ 693.488976] env[61923]: ERROR nova.compute.manager [instance: 41b6744e-d548-442c-b6cc-e12bd403f229] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 693.489295] env[61923]: ERROR nova.compute.manager [instance: 41b6744e-d548-442c-b6cc-e12bd403f229] for vif in network_info: [ 693.489295] env[61923]: ERROR nova.compute.manager [instance: 41b6744e-d548-442c-b6cc-e12bd403f229] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 693.489295] env[61923]: ERROR nova.compute.manager [instance: 41b6744e-d548-442c-b6cc-e12bd403f229] return self._sync_wrapper(fn, *args, **kwargs) [ 693.489295] env[61923]: ERROR nova.compute.manager [instance: 41b6744e-d548-442c-b6cc-e12bd403f229] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 693.489295] env[61923]: ERROR nova.compute.manager [instance: 41b6744e-d548-442c-b6cc-e12bd403f229] self.wait() [ 693.489295] env[61923]: ERROR nova.compute.manager [instance: 41b6744e-d548-442c-b6cc-e12bd403f229] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 693.489295] env[61923]: ERROR nova.compute.manager [instance: 41b6744e-d548-442c-b6cc-e12bd403f229] self[:] = self._gt.wait() [ 693.489295] env[61923]: ERROR nova.compute.manager [instance: 41b6744e-d548-442c-b6cc-e12bd403f229] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 693.489295] env[61923]: ERROR nova.compute.manager [instance: 41b6744e-d548-442c-b6cc-e12bd403f229] return self._exit_event.wait() [ 693.489295] env[61923]: ERROR nova.compute.manager [instance: 41b6744e-d548-442c-b6cc-e12bd403f229] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 693.489295] env[61923]: ERROR nova.compute.manager [instance: 41b6744e-d548-442c-b6cc-e12bd403f229] result = hub.switch() [ 693.489295] env[61923]: ERROR nova.compute.manager [instance: 41b6744e-d548-442c-b6cc-e12bd403f229] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 693.489295] env[61923]: ERROR nova.compute.manager [instance: 41b6744e-d548-442c-b6cc-e12bd403f229] return self.greenlet.switch() [ 693.489602] env[61923]: ERROR nova.compute.manager [instance: 41b6744e-d548-442c-b6cc-e12bd403f229] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 693.489602] env[61923]: ERROR nova.compute.manager [instance: 41b6744e-d548-442c-b6cc-e12bd403f229] result = function(*args, **kwargs) [ 693.489602] env[61923]: ERROR nova.compute.manager [instance: 41b6744e-d548-442c-b6cc-e12bd403f229] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 693.489602] env[61923]: ERROR nova.compute.manager [instance: 41b6744e-d548-442c-b6cc-e12bd403f229] return func(*args, **kwargs) [ 693.489602] env[61923]: ERROR nova.compute.manager [instance: 41b6744e-d548-442c-b6cc-e12bd403f229] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 693.489602] env[61923]: ERROR nova.compute.manager [instance: 41b6744e-d548-442c-b6cc-e12bd403f229] raise e [ 693.489602] env[61923]: ERROR nova.compute.manager [instance: 41b6744e-d548-442c-b6cc-e12bd403f229] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 693.489602] env[61923]: ERROR nova.compute.manager [instance: 41b6744e-d548-442c-b6cc-e12bd403f229] nwinfo = self.network_api.allocate_for_instance( [ 693.489602] env[61923]: ERROR nova.compute.manager [instance: 41b6744e-d548-442c-b6cc-e12bd403f229] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 693.489602] env[61923]: ERROR nova.compute.manager [instance: 41b6744e-d548-442c-b6cc-e12bd403f229] created_port_ids = self._update_ports_for_instance( [ 693.489602] env[61923]: ERROR nova.compute.manager [instance: 41b6744e-d548-442c-b6cc-e12bd403f229] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 693.489602] env[61923]: ERROR nova.compute.manager [instance: 41b6744e-d548-442c-b6cc-e12bd403f229] with excutils.save_and_reraise_exception(): [ 693.489602] env[61923]: ERROR nova.compute.manager [instance: 41b6744e-d548-442c-b6cc-e12bd403f229] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 693.490017] env[61923]: ERROR nova.compute.manager [instance: 41b6744e-d548-442c-b6cc-e12bd403f229] self.force_reraise() [ 693.490017] env[61923]: ERROR nova.compute.manager [instance: 41b6744e-d548-442c-b6cc-e12bd403f229] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 693.490017] env[61923]: ERROR nova.compute.manager [instance: 41b6744e-d548-442c-b6cc-e12bd403f229] raise self.value [ 693.490017] env[61923]: ERROR nova.compute.manager [instance: 41b6744e-d548-442c-b6cc-e12bd403f229] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 693.490017] env[61923]: ERROR nova.compute.manager [instance: 41b6744e-d548-442c-b6cc-e12bd403f229] updated_port = self._update_port( [ 693.490017] env[61923]: ERROR nova.compute.manager [instance: 41b6744e-d548-442c-b6cc-e12bd403f229] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 693.490017] env[61923]: ERROR nova.compute.manager [instance: 41b6744e-d548-442c-b6cc-e12bd403f229] _ensure_no_port_binding_failure(port) [ 693.490017] env[61923]: ERROR nova.compute.manager [instance: 41b6744e-d548-442c-b6cc-e12bd403f229] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 693.490017] env[61923]: ERROR nova.compute.manager [instance: 41b6744e-d548-442c-b6cc-e12bd403f229] raise exception.PortBindingFailed(port_id=port['id']) [ 693.490017] env[61923]: ERROR nova.compute.manager [instance: 41b6744e-d548-442c-b6cc-e12bd403f229] nova.exception.PortBindingFailed: Binding failed for port 32a4c72c-9bfd-4f5f-a3e0-6a0ca4c9b6d2, please check neutron logs for more information. [ 693.490017] env[61923]: ERROR nova.compute.manager [instance: 41b6744e-d548-442c-b6cc-e12bd403f229] [ 693.490294] env[61923]: DEBUG nova.compute.utils [None req-859ffdff-aaee-46b4-bdd6-0a15cce9792f tempest-ServerRescueNegativeTestJSON-1435265126 tempest-ServerRescueNegativeTestJSON-1435265126-project-member] [instance: 41b6744e-d548-442c-b6cc-e12bd403f229] Binding failed for port 32a4c72c-9bfd-4f5f-a3e0-6a0ca4c9b6d2, please check neutron logs for more information. {{(pid=61923) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 693.491087] env[61923]: DEBUG oslo_concurrency.lockutils [None req-df4eb7d8-5b64-4db4-b042-683b16e0b991 tempest-ServersTestBootFromVolume-2104901628 tempest-ServersTestBootFromVolume-2104901628-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.384s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 693.492557] env[61923]: INFO nova.compute.claims [None req-df4eb7d8-5b64-4db4-b042-683b16e0b991 tempest-ServersTestBootFromVolume-2104901628 tempest-ServersTestBootFromVolume-2104901628-project-member] [instance: fb548893-2df5-4629-bf3a-3b0f807f6357] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 693.495525] env[61923]: DEBUG nova.compute.manager [None req-859ffdff-aaee-46b4-bdd6-0a15cce9792f tempest-ServerRescueNegativeTestJSON-1435265126 tempest-ServerRescueNegativeTestJSON-1435265126-project-member] [instance: 41b6744e-d548-442c-b6cc-e12bd403f229] Build of instance 41b6744e-d548-442c-b6cc-e12bd403f229 was re-scheduled: Binding failed for port 32a4c72c-9bfd-4f5f-a3e0-6a0ca4c9b6d2, please check neutron logs for more information. {{(pid=61923) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 693.495965] env[61923]: DEBUG nova.compute.manager [None req-859ffdff-aaee-46b4-bdd6-0a15cce9792f tempest-ServerRescueNegativeTestJSON-1435265126 tempest-ServerRescueNegativeTestJSON-1435265126-project-member] [instance: 41b6744e-d548-442c-b6cc-e12bd403f229] Unplugging VIFs for instance {{(pid=61923) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 693.496200] env[61923]: DEBUG oslo_concurrency.lockutils [None req-859ffdff-aaee-46b4-bdd6-0a15cce9792f tempest-ServerRescueNegativeTestJSON-1435265126 tempest-ServerRescueNegativeTestJSON-1435265126-project-member] Acquiring lock "refresh_cache-41b6744e-d548-442c-b6cc-e12bd403f229" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 693.496341] env[61923]: DEBUG oslo_concurrency.lockutils [None req-859ffdff-aaee-46b4-bdd6-0a15cce9792f tempest-ServerRescueNegativeTestJSON-1435265126 tempest-ServerRescueNegativeTestJSON-1435265126-project-member] Acquired lock "refresh_cache-41b6744e-d548-442c-b6cc-e12bd403f229" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 693.496493] env[61923]: DEBUG nova.network.neutron [None req-859ffdff-aaee-46b4-bdd6-0a15cce9792f tempest-ServerRescueNegativeTestJSON-1435265126 tempest-ServerRescueNegativeTestJSON-1435265126-project-member] [instance: 41b6744e-d548-442c-b6cc-e12bd403f229] Building network info cache for instance {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 693.589411] env[61923]: DEBUG nova.compute.manager [None req-6833b601-845b-439d-ad97-e2f9c0927775 tempest-ServerRescueNegativeTestJSON-1435265126 tempest-ServerRescueNegativeTestJSON-1435265126-project-member] [instance: 4f5348ec-f0b6-45e2-be04-31dbf1d49a3a] Start spawning the instance on the hypervisor. {{(pid=61923) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 693.621203] env[61923]: DEBUG nova.virt.hardware [None req-6833b601-845b-439d-ad97-e2f9c0927775 tempest-ServerRescueNegativeTestJSON-1435265126 tempest-ServerRescueNegativeTestJSON-1435265126-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-29T20:07:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-29T20:07:35Z,direct_url=,disk_format='vmdk',id=0f153f63-ae0a-45b1-b7f5-7f9b673c947f,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='4e7b5e3b3c3443c8bd5c70f8a15739f5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-29T20:07:36Z,virtual_size=,visibility=), allow threads: False {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 693.621479] env[61923]: DEBUG nova.virt.hardware [None req-6833b601-845b-439d-ad97-e2f9c0927775 tempest-ServerRescueNegativeTestJSON-1435265126 tempest-ServerRescueNegativeTestJSON-1435265126-project-member] Flavor limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 693.621628] env[61923]: DEBUG nova.virt.hardware [None req-6833b601-845b-439d-ad97-e2f9c0927775 tempest-ServerRescueNegativeTestJSON-1435265126 tempest-ServerRescueNegativeTestJSON-1435265126-project-member] Image limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 693.621819] env[61923]: DEBUG nova.virt.hardware [None req-6833b601-845b-439d-ad97-e2f9c0927775 tempest-ServerRescueNegativeTestJSON-1435265126 tempest-ServerRescueNegativeTestJSON-1435265126-project-member] Flavor pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 693.621959] env[61923]: DEBUG nova.virt.hardware [None req-6833b601-845b-439d-ad97-e2f9c0927775 tempest-ServerRescueNegativeTestJSON-1435265126 tempest-ServerRescueNegativeTestJSON-1435265126-project-member] Image pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 693.622286] env[61923]: DEBUG nova.virt.hardware [None req-6833b601-845b-439d-ad97-e2f9c0927775 tempest-ServerRescueNegativeTestJSON-1435265126 tempest-ServerRescueNegativeTestJSON-1435265126-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 693.622526] env[61923]: DEBUG nova.virt.hardware [None req-6833b601-845b-439d-ad97-e2f9c0927775 tempest-ServerRescueNegativeTestJSON-1435265126 tempest-ServerRescueNegativeTestJSON-1435265126-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 693.622686] env[61923]: DEBUG nova.virt.hardware [None req-6833b601-845b-439d-ad97-e2f9c0927775 tempest-ServerRescueNegativeTestJSON-1435265126 tempest-ServerRescueNegativeTestJSON-1435265126-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 693.622848] env[61923]: DEBUG nova.virt.hardware [None req-6833b601-845b-439d-ad97-e2f9c0927775 tempest-ServerRescueNegativeTestJSON-1435265126 tempest-ServerRescueNegativeTestJSON-1435265126-project-member] Got 1 possible topologies {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 693.623014] env[61923]: DEBUG nova.virt.hardware [None req-6833b601-845b-439d-ad97-e2f9c0927775 tempest-ServerRescueNegativeTestJSON-1435265126 tempest-ServerRescueNegativeTestJSON-1435265126-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 693.623207] env[61923]: DEBUG nova.virt.hardware [None req-6833b601-845b-439d-ad97-e2f9c0927775 tempest-ServerRescueNegativeTestJSON-1435265126 tempest-ServerRescueNegativeTestJSON-1435265126-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 693.624066] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-739262b8-50b0-47be-abdd-c30db1a4efa1 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.632025] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-337f3e85-2d7a-4eb6-8e24-c5da710a9672 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 693.684990] env[61923]: DEBUG nova.network.neutron [req-1dcab4c9-cedc-4382-86c9-e41adb92550c req-43b08bb8-5794-4d01-a9ef-589072a5838d service nova] [instance: 60ffc8b0-4a7b-4e2a-8774-d16e502d8a06] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 693.725802] env[61923]: DEBUG nova.network.neutron [-] [instance: 60ffc8b0-4a7b-4e2a-8774-d16e502d8a06] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 693.828536] env[61923]: DEBUG nova.network.neutron [req-1dcab4c9-cedc-4382-86c9-e41adb92550c req-43b08bb8-5794-4d01-a9ef-589072a5838d service nova] [instance: 60ffc8b0-4a7b-4e2a-8774-d16e502d8a06] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 693.889091] env[61923]: ERROR nova.compute.manager [None req-6833b601-845b-439d-ad97-e2f9c0927775 tempest-ServerRescueNegativeTestJSON-1435265126 tempest-ServerRescueNegativeTestJSON-1435265126-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port a616fa04-1fe8-475c-9237-a801d7430538, please check neutron logs for more information. [ 693.889091] env[61923]: ERROR nova.compute.manager Traceback (most recent call last): [ 693.889091] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 693.889091] env[61923]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 693.889091] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 693.889091] env[61923]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 693.889091] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 693.889091] env[61923]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 693.889091] env[61923]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 693.889091] env[61923]: ERROR nova.compute.manager self.force_reraise() [ 693.889091] env[61923]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 693.889091] env[61923]: ERROR nova.compute.manager raise self.value [ 693.889091] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 693.889091] env[61923]: ERROR nova.compute.manager updated_port = self._update_port( [ 693.889091] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 693.889091] env[61923]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 693.889545] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 693.889545] env[61923]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 693.889545] env[61923]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port a616fa04-1fe8-475c-9237-a801d7430538, please check neutron logs for more information. [ 693.889545] env[61923]: ERROR nova.compute.manager [ 693.889545] env[61923]: Traceback (most recent call last): [ 693.889545] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 693.889545] env[61923]: listener.cb(fileno) [ 693.889545] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 693.889545] env[61923]: result = function(*args, **kwargs) [ 693.889545] env[61923]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 693.889545] env[61923]: return func(*args, **kwargs) [ 693.889545] env[61923]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 693.889545] env[61923]: raise e [ 693.889545] env[61923]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 693.889545] env[61923]: nwinfo = self.network_api.allocate_for_instance( [ 693.889545] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 693.889545] env[61923]: created_port_ids = self._update_ports_for_instance( [ 693.889545] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 693.889545] env[61923]: with excutils.save_and_reraise_exception(): [ 693.889545] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 693.889545] env[61923]: self.force_reraise() [ 693.889545] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 693.889545] env[61923]: raise self.value [ 693.889545] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 693.889545] env[61923]: updated_port = self._update_port( [ 693.889545] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 693.889545] env[61923]: _ensure_no_port_binding_failure(port) [ 693.889545] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 693.889545] env[61923]: raise exception.PortBindingFailed(port_id=port['id']) [ 693.890232] env[61923]: nova.exception.PortBindingFailed: Binding failed for port a616fa04-1fe8-475c-9237-a801d7430538, please check neutron logs for more information. [ 693.890232] env[61923]: Removing descriptor: 14 [ 693.890232] env[61923]: ERROR nova.compute.manager [None req-6833b601-845b-439d-ad97-e2f9c0927775 tempest-ServerRescueNegativeTestJSON-1435265126 tempest-ServerRescueNegativeTestJSON-1435265126-project-member] [instance: 4f5348ec-f0b6-45e2-be04-31dbf1d49a3a] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port a616fa04-1fe8-475c-9237-a801d7430538, please check neutron logs for more information. [ 693.890232] env[61923]: ERROR nova.compute.manager [instance: 4f5348ec-f0b6-45e2-be04-31dbf1d49a3a] Traceback (most recent call last): [ 693.890232] env[61923]: ERROR nova.compute.manager [instance: 4f5348ec-f0b6-45e2-be04-31dbf1d49a3a] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 693.890232] env[61923]: ERROR nova.compute.manager [instance: 4f5348ec-f0b6-45e2-be04-31dbf1d49a3a] yield resources [ 693.890232] env[61923]: ERROR nova.compute.manager [instance: 4f5348ec-f0b6-45e2-be04-31dbf1d49a3a] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 693.890232] env[61923]: ERROR nova.compute.manager [instance: 4f5348ec-f0b6-45e2-be04-31dbf1d49a3a] self.driver.spawn(context, instance, image_meta, [ 693.890232] env[61923]: ERROR nova.compute.manager [instance: 4f5348ec-f0b6-45e2-be04-31dbf1d49a3a] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 693.890232] env[61923]: ERROR nova.compute.manager [instance: 4f5348ec-f0b6-45e2-be04-31dbf1d49a3a] self._vmops.spawn(context, instance, image_meta, injected_files, [ 693.890232] env[61923]: ERROR nova.compute.manager [instance: 4f5348ec-f0b6-45e2-be04-31dbf1d49a3a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 693.890232] env[61923]: ERROR nova.compute.manager [instance: 4f5348ec-f0b6-45e2-be04-31dbf1d49a3a] vm_ref = self.build_virtual_machine(instance, [ 693.890538] env[61923]: ERROR nova.compute.manager [instance: 4f5348ec-f0b6-45e2-be04-31dbf1d49a3a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 693.890538] env[61923]: ERROR nova.compute.manager [instance: 4f5348ec-f0b6-45e2-be04-31dbf1d49a3a] vif_infos = vmwarevif.get_vif_info(self._session, [ 693.890538] env[61923]: ERROR nova.compute.manager [instance: 4f5348ec-f0b6-45e2-be04-31dbf1d49a3a] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 693.890538] env[61923]: ERROR nova.compute.manager [instance: 4f5348ec-f0b6-45e2-be04-31dbf1d49a3a] for vif in network_info: [ 693.890538] env[61923]: ERROR nova.compute.manager [instance: 4f5348ec-f0b6-45e2-be04-31dbf1d49a3a] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 693.890538] env[61923]: ERROR nova.compute.manager [instance: 4f5348ec-f0b6-45e2-be04-31dbf1d49a3a] return self._sync_wrapper(fn, *args, **kwargs) [ 693.890538] env[61923]: ERROR nova.compute.manager [instance: 4f5348ec-f0b6-45e2-be04-31dbf1d49a3a] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 693.890538] env[61923]: ERROR nova.compute.manager [instance: 4f5348ec-f0b6-45e2-be04-31dbf1d49a3a] self.wait() [ 693.890538] env[61923]: ERROR nova.compute.manager [instance: 4f5348ec-f0b6-45e2-be04-31dbf1d49a3a] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 693.890538] env[61923]: ERROR nova.compute.manager [instance: 4f5348ec-f0b6-45e2-be04-31dbf1d49a3a] self[:] = self._gt.wait() [ 693.890538] env[61923]: ERROR nova.compute.manager [instance: 4f5348ec-f0b6-45e2-be04-31dbf1d49a3a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 693.890538] env[61923]: ERROR nova.compute.manager [instance: 4f5348ec-f0b6-45e2-be04-31dbf1d49a3a] return self._exit_event.wait() [ 693.890538] env[61923]: ERROR nova.compute.manager [instance: 4f5348ec-f0b6-45e2-be04-31dbf1d49a3a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 693.890862] env[61923]: ERROR nova.compute.manager [instance: 4f5348ec-f0b6-45e2-be04-31dbf1d49a3a] result = hub.switch() [ 693.890862] env[61923]: ERROR nova.compute.manager [instance: 4f5348ec-f0b6-45e2-be04-31dbf1d49a3a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 693.890862] env[61923]: ERROR nova.compute.manager [instance: 4f5348ec-f0b6-45e2-be04-31dbf1d49a3a] return self.greenlet.switch() [ 693.890862] env[61923]: ERROR nova.compute.manager [instance: 4f5348ec-f0b6-45e2-be04-31dbf1d49a3a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 693.890862] env[61923]: ERROR nova.compute.manager [instance: 4f5348ec-f0b6-45e2-be04-31dbf1d49a3a] result = function(*args, **kwargs) [ 693.890862] env[61923]: ERROR nova.compute.manager [instance: 4f5348ec-f0b6-45e2-be04-31dbf1d49a3a] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 693.890862] env[61923]: ERROR nova.compute.manager [instance: 4f5348ec-f0b6-45e2-be04-31dbf1d49a3a] return func(*args, **kwargs) [ 693.890862] env[61923]: ERROR nova.compute.manager [instance: 4f5348ec-f0b6-45e2-be04-31dbf1d49a3a] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 693.890862] env[61923]: ERROR nova.compute.manager [instance: 4f5348ec-f0b6-45e2-be04-31dbf1d49a3a] raise e [ 693.890862] env[61923]: ERROR nova.compute.manager [instance: 4f5348ec-f0b6-45e2-be04-31dbf1d49a3a] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 693.890862] env[61923]: ERROR nova.compute.manager [instance: 4f5348ec-f0b6-45e2-be04-31dbf1d49a3a] nwinfo = self.network_api.allocate_for_instance( [ 693.890862] env[61923]: ERROR nova.compute.manager [instance: 4f5348ec-f0b6-45e2-be04-31dbf1d49a3a] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 693.890862] env[61923]: ERROR nova.compute.manager [instance: 4f5348ec-f0b6-45e2-be04-31dbf1d49a3a] created_port_ids = self._update_ports_for_instance( [ 693.891267] env[61923]: ERROR nova.compute.manager [instance: 4f5348ec-f0b6-45e2-be04-31dbf1d49a3a] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 693.891267] env[61923]: ERROR nova.compute.manager [instance: 4f5348ec-f0b6-45e2-be04-31dbf1d49a3a] with excutils.save_and_reraise_exception(): [ 693.891267] env[61923]: ERROR nova.compute.manager [instance: 4f5348ec-f0b6-45e2-be04-31dbf1d49a3a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 693.891267] env[61923]: ERROR nova.compute.manager [instance: 4f5348ec-f0b6-45e2-be04-31dbf1d49a3a] self.force_reraise() [ 693.891267] env[61923]: ERROR nova.compute.manager [instance: 4f5348ec-f0b6-45e2-be04-31dbf1d49a3a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 693.891267] env[61923]: ERROR nova.compute.manager [instance: 4f5348ec-f0b6-45e2-be04-31dbf1d49a3a] raise self.value [ 693.891267] env[61923]: ERROR nova.compute.manager [instance: 4f5348ec-f0b6-45e2-be04-31dbf1d49a3a] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 693.891267] env[61923]: ERROR nova.compute.manager [instance: 4f5348ec-f0b6-45e2-be04-31dbf1d49a3a] updated_port = self._update_port( [ 693.891267] env[61923]: ERROR nova.compute.manager [instance: 4f5348ec-f0b6-45e2-be04-31dbf1d49a3a] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 693.891267] env[61923]: ERROR nova.compute.manager [instance: 4f5348ec-f0b6-45e2-be04-31dbf1d49a3a] _ensure_no_port_binding_failure(port) [ 693.891267] env[61923]: ERROR nova.compute.manager [instance: 4f5348ec-f0b6-45e2-be04-31dbf1d49a3a] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 693.891267] env[61923]: ERROR nova.compute.manager [instance: 4f5348ec-f0b6-45e2-be04-31dbf1d49a3a] raise exception.PortBindingFailed(port_id=port['id']) [ 693.891591] env[61923]: ERROR nova.compute.manager [instance: 4f5348ec-f0b6-45e2-be04-31dbf1d49a3a] nova.exception.PortBindingFailed: Binding failed for port a616fa04-1fe8-475c-9237-a801d7430538, please check neutron logs for more information. [ 693.891591] env[61923]: ERROR nova.compute.manager [instance: 4f5348ec-f0b6-45e2-be04-31dbf1d49a3a] [ 693.891591] env[61923]: INFO nova.compute.manager [None req-6833b601-845b-439d-ad97-e2f9c0927775 tempest-ServerRescueNegativeTestJSON-1435265126 tempest-ServerRescueNegativeTestJSON-1435265126-project-member] [instance: 4f5348ec-f0b6-45e2-be04-31dbf1d49a3a] Terminating instance [ 693.893047] env[61923]: DEBUG oslo_concurrency.lockutils [None req-6833b601-845b-439d-ad97-e2f9c0927775 tempest-ServerRescueNegativeTestJSON-1435265126 tempest-ServerRescueNegativeTestJSON-1435265126-project-member] Acquiring lock "refresh_cache-4f5348ec-f0b6-45e2-be04-31dbf1d49a3a" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 693.893047] env[61923]: DEBUG oslo_concurrency.lockutils [None req-6833b601-845b-439d-ad97-e2f9c0927775 tempest-ServerRescueNegativeTestJSON-1435265126 tempest-ServerRescueNegativeTestJSON-1435265126-project-member] Acquired lock "refresh_cache-4f5348ec-f0b6-45e2-be04-31dbf1d49a3a" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 693.893589] env[61923]: DEBUG nova.network.neutron [None req-6833b601-845b-439d-ad97-e2f9c0927775 tempest-ServerRescueNegativeTestJSON-1435265126 tempest-ServerRescueNegativeTestJSON-1435265126-project-member] [instance: 4f5348ec-f0b6-45e2-be04-31dbf1d49a3a] Building network info cache for instance {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 694.024243] env[61923]: DEBUG nova.network.neutron [None req-859ffdff-aaee-46b4-bdd6-0a15cce9792f tempest-ServerRescueNegativeTestJSON-1435265126 tempest-ServerRescueNegativeTestJSON-1435265126-project-member] [instance: 41b6744e-d548-442c-b6cc-e12bd403f229] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 694.135244] env[61923]: DEBUG nova.network.neutron [None req-859ffdff-aaee-46b4-bdd6-0a15cce9792f tempest-ServerRescueNegativeTestJSON-1435265126 tempest-ServerRescueNegativeTestJSON-1435265126-project-member] [instance: 41b6744e-d548-442c-b6cc-e12bd403f229] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 694.227797] env[61923]: INFO nova.compute.manager [-] [instance: 60ffc8b0-4a7b-4e2a-8774-d16e502d8a06] Took 1.04 seconds to deallocate network for instance. [ 694.230664] env[61923]: DEBUG nova.compute.claims [None req-08382130-d70f-42e8-a5b5-93e86f5fa49a tempest-AttachVolumeShelveTestJSON-1252849694 tempest-AttachVolumeShelveTestJSON-1252849694-project-member] [instance: 60ffc8b0-4a7b-4e2a-8774-d16e502d8a06] Aborting claim: {{(pid=61923) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 694.230898] env[61923]: DEBUG oslo_concurrency.lockutils [None req-08382130-d70f-42e8-a5b5-93e86f5fa49a tempest-AttachVolumeShelveTestJSON-1252849694 tempest-AttachVolumeShelveTestJSON-1252849694-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 694.331874] env[61923]: DEBUG oslo_concurrency.lockutils [req-1dcab4c9-cedc-4382-86c9-e41adb92550c req-43b08bb8-5794-4d01-a9ef-589072a5838d service nova] Releasing lock "refresh_cache-60ffc8b0-4a7b-4e2a-8774-d16e502d8a06" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 694.334152] env[61923]: DEBUG nova.compute.manager [req-1dcab4c9-cedc-4382-86c9-e41adb92550c req-43b08bb8-5794-4d01-a9ef-589072a5838d service nova] [instance: 60ffc8b0-4a7b-4e2a-8774-d16e502d8a06] Received event network-vif-deleted-c88b194c-774a-448e-8f31-5a89cfdbc187 {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 694.410623] env[61923]: DEBUG nova.network.neutron [None req-6833b601-845b-439d-ad97-e2f9c0927775 tempest-ServerRescueNegativeTestJSON-1435265126 tempest-ServerRescueNegativeTestJSON-1435265126-project-member] [instance: 4f5348ec-f0b6-45e2-be04-31dbf1d49a3a] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 694.487271] env[61923]: DEBUG nova.network.neutron [None req-6833b601-845b-439d-ad97-e2f9c0927775 tempest-ServerRescueNegativeTestJSON-1435265126 tempest-ServerRescueNegativeTestJSON-1435265126-project-member] [instance: 4f5348ec-f0b6-45e2-be04-31dbf1d49a3a] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 694.641101] env[61923]: DEBUG oslo_concurrency.lockutils [None req-859ffdff-aaee-46b4-bdd6-0a15cce9792f tempest-ServerRescueNegativeTestJSON-1435265126 tempest-ServerRescueNegativeTestJSON-1435265126-project-member] Releasing lock "refresh_cache-41b6744e-d548-442c-b6cc-e12bd403f229" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 694.641101] env[61923]: DEBUG nova.compute.manager [None req-859ffdff-aaee-46b4-bdd6-0a15cce9792f tempest-ServerRescueNegativeTestJSON-1435265126 tempest-ServerRescueNegativeTestJSON-1435265126-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61923) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 694.641101] env[61923]: DEBUG nova.compute.manager [None req-859ffdff-aaee-46b4-bdd6-0a15cce9792f tempest-ServerRescueNegativeTestJSON-1435265126 tempest-ServerRescueNegativeTestJSON-1435265126-project-member] [instance: 41b6744e-d548-442c-b6cc-e12bd403f229] Deallocating network for instance {{(pid=61923) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 694.641440] env[61923]: DEBUG nova.network.neutron [None req-859ffdff-aaee-46b4-bdd6-0a15cce9792f tempest-ServerRescueNegativeTestJSON-1435265126 tempest-ServerRescueNegativeTestJSON-1435265126-project-member] [instance: 41b6744e-d548-442c-b6cc-e12bd403f229] deallocate_for_instance() {{(pid=61923) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 694.658983] env[61923]: DEBUG nova.network.neutron [None req-859ffdff-aaee-46b4-bdd6-0a15cce9792f tempest-ServerRescueNegativeTestJSON-1435265126 tempest-ServerRescueNegativeTestJSON-1435265126-project-member] [instance: 41b6744e-d548-442c-b6cc-e12bd403f229] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 694.727915] env[61923]: DEBUG nova.compute.manager [req-247b60a0-27c3-4e5f-95ca-3f65e17e668e req-56b4785c-d8dd-48ce-8909-b5389a2eb9db service nova] [instance: 4f5348ec-f0b6-45e2-be04-31dbf1d49a3a] Received event network-changed-a616fa04-1fe8-475c-9237-a801d7430538 {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 694.728125] env[61923]: DEBUG nova.compute.manager [req-247b60a0-27c3-4e5f-95ca-3f65e17e668e req-56b4785c-d8dd-48ce-8909-b5389a2eb9db service nova] [instance: 4f5348ec-f0b6-45e2-be04-31dbf1d49a3a] Refreshing instance network info cache due to event network-changed-a616fa04-1fe8-475c-9237-a801d7430538. {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 694.728301] env[61923]: DEBUG oslo_concurrency.lockutils [req-247b60a0-27c3-4e5f-95ca-3f65e17e668e req-56b4785c-d8dd-48ce-8909-b5389a2eb9db service nova] Acquiring lock "refresh_cache-4f5348ec-f0b6-45e2-be04-31dbf1d49a3a" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 694.881617] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46fa3a9a-21ca-4a2b-9464-c9262bbab7b3 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.890494] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6da6aeb3-9837-4023-8f26-79ad0313b4ef {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.921735] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a33b35d8-c991-4914-98e4-71b806e0e19f {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.929055] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22d8d9be-52a9-49e7-a1a7-91b387e4e550 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.942726] env[61923]: DEBUG nova.compute.provider_tree [None req-df4eb7d8-5b64-4db4-b042-683b16e0b991 tempest-ServersTestBootFromVolume-2104901628 tempest-ServersTestBootFromVolume-2104901628-project-member] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 694.990722] env[61923]: DEBUG oslo_concurrency.lockutils [None req-6833b601-845b-439d-ad97-e2f9c0927775 tempest-ServerRescueNegativeTestJSON-1435265126 tempest-ServerRescueNegativeTestJSON-1435265126-project-member] Releasing lock "refresh_cache-4f5348ec-f0b6-45e2-be04-31dbf1d49a3a" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 694.991154] env[61923]: DEBUG nova.compute.manager [None req-6833b601-845b-439d-ad97-e2f9c0927775 tempest-ServerRescueNegativeTestJSON-1435265126 tempest-ServerRescueNegativeTestJSON-1435265126-project-member] [instance: 4f5348ec-f0b6-45e2-be04-31dbf1d49a3a] Start destroying the instance on the hypervisor. {{(pid=61923) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 694.991340] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-6833b601-845b-439d-ad97-e2f9c0927775 tempest-ServerRescueNegativeTestJSON-1435265126 tempest-ServerRescueNegativeTestJSON-1435265126-project-member] [instance: 4f5348ec-f0b6-45e2-be04-31dbf1d49a3a] Destroying instance {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 694.991635] env[61923]: DEBUG oslo_concurrency.lockutils [req-247b60a0-27c3-4e5f-95ca-3f65e17e668e req-56b4785c-d8dd-48ce-8909-b5389a2eb9db service nova] Acquired lock "refresh_cache-4f5348ec-f0b6-45e2-be04-31dbf1d49a3a" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 694.991797] env[61923]: DEBUG nova.network.neutron [req-247b60a0-27c3-4e5f-95ca-3f65e17e668e req-56b4785c-d8dd-48ce-8909-b5389a2eb9db service nova] [instance: 4f5348ec-f0b6-45e2-be04-31dbf1d49a3a] Refreshing network info cache for port a616fa04-1fe8-475c-9237-a801d7430538 {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 694.992880] env[61923]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-8bb17cde-9714-4890-a09f-40e73f940422 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.006056] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23c9aca9-7fa3-4aa7-b187-d4517d0527f1 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 695.028679] env[61923]: WARNING nova.virt.vmwareapi.vmops [None req-6833b601-845b-439d-ad97-e2f9c0927775 tempest-ServerRescueNegativeTestJSON-1435265126 tempest-ServerRescueNegativeTestJSON-1435265126-project-member] [instance: 4f5348ec-f0b6-45e2-be04-31dbf1d49a3a] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 4f5348ec-f0b6-45e2-be04-31dbf1d49a3a could not be found. [ 695.028935] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-6833b601-845b-439d-ad97-e2f9c0927775 tempest-ServerRescueNegativeTestJSON-1435265126 tempest-ServerRescueNegativeTestJSON-1435265126-project-member] [instance: 4f5348ec-f0b6-45e2-be04-31dbf1d49a3a] Instance destroyed {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 695.029143] env[61923]: INFO nova.compute.manager [None req-6833b601-845b-439d-ad97-e2f9c0927775 tempest-ServerRescueNegativeTestJSON-1435265126 tempest-ServerRescueNegativeTestJSON-1435265126-project-member] [instance: 4f5348ec-f0b6-45e2-be04-31dbf1d49a3a] Took 0.04 seconds to destroy the instance on the hypervisor. [ 695.029618] env[61923]: DEBUG oslo.service.loopingcall [None req-6833b601-845b-439d-ad97-e2f9c0927775 tempest-ServerRescueNegativeTestJSON-1435265126 tempest-ServerRescueNegativeTestJSON-1435265126-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61923) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 695.029865] env[61923]: DEBUG nova.compute.manager [-] [instance: 4f5348ec-f0b6-45e2-be04-31dbf1d49a3a] Deallocating network for instance {{(pid=61923) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 695.029992] env[61923]: DEBUG nova.network.neutron [-] [instance: 4f5348ec-f0b6-45e2-be04-31dbf1d49a3a] deallocate_for_instance() {{(pid=61923) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 695.163733] env[61923]: DEBUG nova.network.neutron [None req-859ffdff-aaee-46b4-bdd6-0a15cce9792f tempest-ServerRescueNegativeTestJSON-1435265126 tempest-ServerRescueNegativeTestJSON-1435265126-project-member] [instance: 41b6744e-d548-442c-b6cc-e12bd403f229] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 695.203473] env[61923]: DEBUG nova.network.neutron [-] [instance: 4f5348ec-f0b6-45e2-be04-31dbf1d49a3a] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 695.445811] env[61923]: DEBUG nova.scheduler.client.report [None req-df4eb7d8-5b64-4db4-b042-683b16e0b991 tempest-ServersTestBootFromVolume-2104901628 tempest-ServersTestBootFromVolume-2104901628-project-member] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 695.523786] env[61923]: DEBUG nova.network.neutron [req-247b60a0-27c3-4e5f-95ca-3f65e17e668e req-56b4785c-d8dd-48ce-8909-b5389a2eb9db service nova] [instance: 4f5348ec-f0b6-45e2-be04-31dbf1d49a3a] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 695.635238] env[61923]: DEBUG nova.network.neutron [req-247b60a0-27c3-4e5f-95ca-3f65e17e668e req-56b4785c-d8dd-48ce-8909-b5389a2eb9db service nova] [instance: 4f5348ec-f0b6-45e2-be04-31dbf1d49a3a] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 695.666113] env[61923]: INFO nova.compute.manager [None req-859ffdff-aaee-46b4-bdd6-0a15cce9792f tempest-ServerRescueNegativeTestJSON-1435265126 tempest-ServerRescueNegativeTestJSON-1435265126-project-member] [instance: 41b6744e-d548-442c-b6cc-e12bd403f229] Took 1.02 seconds to deallocate network for instance. [ 695.706202] env[61923]: DEBUG nova.network.neutron [-] [instance: 4f5348ec-f0b6-45e2-be04-31dbf1d49a3a] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 695.954657] env[61923]: DEBUG oslo_concurrency.lockutils [None req-df4eb7d8-5b64-4db4-b042-683b16e0b991 tempest-ServersTestBootFromVolume-2104901628 tempest-ServersTestBootFromVolume-2104901628-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.464s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 695.955123] env[61923]: DEBUG nova.compute.manager [None req-df4eb7d8-5b64-4db4-b042-683b16e0b991 tempest-ServersTestBootFromVolume-2104901628 tempest-ServersTestBootFromVolume-2104901628-project-member] [instance: fb548893-2df5-4629-bf3a-3b0f807f6357] Start building networks asynchronously for instance. {{(pid=61923) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 695.957900] env[61923]: DEBUG oslo_concurrency.lockutils [None req-44a76586-0f76-448f-af41-a5d846bfda81 tempest-ServerActionsV293TestJSON-849444864 tempest-ServerActionsV293TestJSON-849444864-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.945s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 695.959525] env[61923]: INFO nova.compute.claims [None req-44a76586-0f76-448f-af41-a5d846bfda81 tempest-ServerActionsV293TestJSON-849444864 tempest-ServerActionsV293TestJSON-849444864-project-member] [instance: 62144317-0354-4751-8f2b-4eae588c9da7] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 696.139174] env[61923]: DEBUG oslo_concurrency.lockutils [req-247b60a0-27c3-4e5f-95ca-3f65e17e668e req-56b4785c-d8dd-48ce-8909-b5389a2eb9db service nova] Releasing lock "refresh_cache-4f5348ec-f0b6-45e2-be04-31dbf1d49a3a" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 696.139446] env[61923]: DEBUG nova.compute.manager [req-247b60a0-27c3-4e5f-95ca-3f65e17e668e req-56b4785c-d8dd-48ce-8909-b5389a2eb9db service nova] [instance: 4f5348ec-f0b6-45e2-be04-31dbf1d49a3a] Received event network-vif-deleted-a616fa04-1fe8-475c-9237-a801d7430538 {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 696.208570] env[61923]: INFO nova.compute.manager [-] [instance: 4f5348ec-f0b6-45e2-be04-31dbf1d49a3a] Took 1.18 seconds to deallocate network for instance. [ 696.210995] env[61923]: DEBUG nova.compute.claims [None req-6833b601-845b-439d-ad97-e2f9c0927775 tempest-ServerRescueNegativeTestJSON-1435265126 tempest-ServerRescueNegativeTestJSON-1435265126-project-member] [instance: 4f5348ec-f0b6-45e2-be04-31dbf1d49a3a] Aborting claim: {{(pid=61923) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 696.210995] env[61923]: DEBUG oslo_concurrency.lockutils [None req-6833b601-845b-439d-ad97-e2f9c0927775 tempest-ServerRescueNegativeTestJSON-1435265126 tempest-ServerRescueNegativeTestJSON-1435265126-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 696.464574] env[61923]: DEBUG nova.compute.utils [None req-df4eb7d8-5b64-4db4-b042-683b16e0b991 tempest-ServersTestBootFromVolume-2104901628 tempest-ServersTestBootFromVolume-2104901628-project-member] Using /dev/sd instead of None {{(pid=61923) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 696.467715] env[61923]: DEBUG nova.compute.manager [None req-df4eb7d8-5b64-4db4-b042-683b16e0b991 tempest-ServersTestBootFromVolume-2104901628 tempest-ServersTestBootFromVolume-2104901628-project-member] [instance: fb548893-2df5-4629-bf3a-3b0f807f6357] Allocating IP information in the background. {{(pid=61923) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 696.468184] env[61923]: DEBUG nova.network.neutron [None req-df4eb7d8-5b64-4db4-b042-683b16e0b991 tempest-ServersTestBootFromVolume-2104901628 tempest-ServersTestBootFromVolume-2104901628-project-member] [instance: fb548893-2df5-4629-bf3a-3b0f807f6357] allocate_for_instance() {{(pid=61923) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 696.525852] env[61923]: DEBUG nova.policy [None req-df4eb7d8-5b64-4db4-b042-683b16e0b991 tempest-ServersTestBootFromVolume-2104901628 tempest-ServersTestBootFromVolume-2104901628-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd62f71e5308c42b598fa5ada6110eca4', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '2e8415c6f16f472eb8bd4283afadc2b0', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61923) authorize /opt/stack/nova/nova/policy.py:201}} [ 696.697308] env[61923]: INFO nova.scheduler.client.report [None req-859ffdff-aaee-46b4-bdd6-0a15cce9792f tempest-ServerRescueNegativeTestJSON-1435265126 tempest-ServerRescueNegativeTestJSON-1435265126-project-member] Deleted allocations for instance 41b6744e-d548-442c-b6cc-e12bd403f229 [ 696.839252] env[61923]: DEBUG oslo_concurrency.lockutils [None req-a9ea97fa-71e5-4463-a1f3-ab69743d14a4 tempest-ServerActionsTestJSON-1295690556 tempest-ServerActionsTestJSON-1295690556-project-member] Acquiring lock "ad3b3933-5bd9-444a-af7b-c81e868e0622" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 696.839252] env[61923]: DEBUG oslo_concurrency.lockutils [None req-a9ea97fa-71e5-4463-a1f3-ab69743d14a4 tempest-ServerActionsTestJSON-1295690556 tempest-ServerActionsTestJSON-1295690556-project-member] Lock "ad3b3933-5bd9-444a-af7b-c81e868e0622" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 696.933779] env[61923]: DEBUG nova.network.neutron [None req-df4eb7d8-5b64-4db4-b042-683b16e0b991 tempest-ServersTestBootFromVolume-2104901628 tempest-ServersTestBootFromVolume-2104901628-project-member] [instance: fb548893-2df5-4629-bf3a-3b0f807f6357] Successfully created port: 02c3270e-c070-4203-bf7e-d9f1db17ad7e {{(pid=61923) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 696.969263] env[61923]: DEBUG nova.compute.manager [None req-df4eb7d8-5b64-4db4-b042-683b16e0b991 tempest-ServersTestBootFromVolume-2104901628 tempest-ServersTestBootFromVolume-2104901628-project-member] [instance: fb548893-2df5-4629-bf3a-3b0f807f6357] Start building block device mappings for instance. {{(pid=61923) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 697.212993] env[61923]: DEBUG oslo_concurrency.lockutils [None req-859ffdff-aaee-46b4-bdd6-0a15cce9792f tempest-ServerRescueNegativeTestJSON-1435265126 tempest-ServerRescueNegativeTestJSON-1435265126-project-member] Lock "41b6744e-d548-442c-b6cc-e12bd403f229" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 121.546s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 697.338018] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5169fe08-4a5c-4ff3-bed6-36358e779ee1 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.344200] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a91b00ea-9992-4c90-9647-f7866fcd2512 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.381332] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22c57d28-14d3-4e71-b7a3-eb5641e627cd {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.389025] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e371dd8d-4082-49ca-80da-f4047fc97080 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.402963] env[61923]: DEBUG nova.compute.provider_tree [None req-44a76586-0f76-448f-af41-a5d846bfda81 tempest-ServerActionsV293TestJSON-849444864 tempest-ServerActionsV293TestJSON-849444864-project-member] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 697.477433] env[61923]: INFO nova.virt.block_device [None req-df4eb7d8-5b64-4db4-b042-683b16e0b991 tempest-ServersTestBootFromVolume-2104901628 tempest-ServersTestBootFromVolume-2104901628-project-member] [instance: fb548893-2df5-4629-bf3a-3b0f807f6357] Booting with volume 5e72b245-48a6-4832-a2bc-efa0bb1b6c11 at /dev/sda [ 697.523708] env[61923]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-635dfc72-9da4-4949-b7cf-a210a2c7325a {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.532322] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-752f7655-894e-4884-b895-93d33806e999 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.554131] env[61923]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ef9fb9a3-5e3f-4f9e-b1f2-ea2833db77a8 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.562564] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-917e1628-360c-494a-9969-f74178ba1f08 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.584549] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83fdb6e9-5e97-47d3-8ede-e0c73bf48602 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.591045] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5557e9d-8ad8-4fc5-aa18-3c868fdb9a72 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 697.606185] env[61923]: DEBUG nova.virt.block_device [None req-df4eb7d8-5b64-4db4-b042-683b16e0b991 tempest-ServersTestBootFromVolume-2104901628 tempest-ServersTestBootFromVolume-2104901628-project-member] [instance: fb548893-2df5-4629-bf3a-3b0f807f6357] Updating existing volume attachment record: a079b3f3-cead-4319-98ed-e70a37d7a48c {{(pid=61923) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 697.714756] env[61923]: DEBUG nova.compute.manager [None req-425a3606-915c-4f9b-8507-50f1c8d0d8d9 tempest-ServerRescueTestJSON-521767381 tempest-ServerRescueTestJSON-521767381-project-member] [instance: 74b32299-983f-41f7-b3f4-3a12815a83e6] Starting instance... {{(pid=61923) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 697.913176] env[61923]: DEBUG nova.scheduler.client.report [None req-44a76586-0f76-448f-af41-a5d846bfda81 tempest-ServerActionsV293TestJSON-849444864 tempest-ServerActionsV293TestJSON-849444864-project-member] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 697.934660] env[61923]: DEBUG nova.compute.manager [req-f7233aab-2d7d-4932-941a-6164985c2db1 req-8c926b96-731e-49f4-9ba2-c34ca252eade service nova] [instance: fb548893-2df5-4629-bf3a-3b0f807f6357] Received event network-changed-02c3270e-c070-4203-bf7e-d9f1db17ad7e {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 697.934863] env[61923]: DEBUG nova.compute.manager [req-f7233aab-2d7d-4932-941a-6164985c2db1 req-8c926b96-731e-49f4-9ba2-c34ca252eade service nova] [instance: fb548893-2df5-4629-bf3a-3b0f807f6357] Refreshing instance network info cache due to event network-changed-02c3270e-c070-4203-bf7e-d9f1db17ad7e. {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 697.935089] env[61923]: DEBUG oslo_concurrency.lockutils [req-f7233aab-2d7d-4932-941a-6164985c2db1 req-8c926b96-731e-49f4-9ba2-c34ca252eade service nova] Acquiring lock "refresh_cache-fb548893-2df5-4629-bf3a-3b0f807f6357" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 697.935233] env[61923]: DEBUG oslo_concurrency.lockutils [req-f7233aab-2d7d-4932-941a-6164985c2db1 req-8c926b96-731e-49f4-9ba2-c34ca252eade service nova] Acquired lock "refresh_cache-fb548893-2df5-4629-bf3a-3b0f807f6357" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 697.935390] env[61923]: DEBUG nova.network.neutron [req-f7233aab-2d7d-4932-941a-6164985c2db1 req-8c926b96-731e-49f4-9ba2-c34ca252eade service nova] [instance: fb548893-2df5-4629-bf3a-3b0f807f6357] Refreshing network info cache for port 02c3270e-c070-4203-bf7e-d9f1db17ad7e {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 698.127166] env[61923]: ERROR nova.compute.manager [None req-df4eb7d8-5b64-4db4-b042-683b16e0b991 tempest-ServersTestBootFromVolume-2104901628 tempest-ServersTestBootFromVolume-2104901628-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 02c3270e-c070-4203-bf7e-d9f1db17ad7e, please check neutron logs for more information. [ 698.127166] env[61923]: ERROR nova.compute.manager Traceback (most recent call last): [ 698.127166] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 698.127166] env[61923]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 698.127166] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 698.127166] env[61923]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 698.127166] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 698.127166] env[61923]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 698.127166] env[61923]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 698.127166] env[61923]: ERROR nova.compute.manager self.force_reraise() [ 698.127166] env[61923]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 698.127166] env[61923]: ERROR nova.compute.manager raise self.value [ 698.127166] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 698.127166] env[61923]: ERROR nova.compute.manager updated_port = self._update_port( [ 698.127166] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 698.127166] env[61923]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 698.127665] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 698.127665] env[61923]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 698.127665] env[61923]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 02c3270e-c070-4203-bf7e-d9f1db17ad7e, please check neutron logs for more information. [ 698.127665] env[61923]: ERROR nova.compute.manager [ 698.127665] env[61923]: Traceback (most recent call last): [ 698.127665] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 698.127665] env[61923]: listener.cb(fileno) [ 698.127665] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 698.127665] env[61923]: result = function(*args, **kwargs) [ 698.127665] env[61923]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 698.127665] env[61923]: return func(*args, **kwargs) [ 698.127665] env[61923]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 698.127665] env[61923]: raise e [ 698.127665] env[61923]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 698.127665] env[61923]: nwinfo = self.network_api.allocate_for_instance( [ 698.127665] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 698.127665] env[61923]: created_port_ids = self._update_ports_for_instance( [ 698.127665] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 698.127665] env[61923]: with excutils.save_and_reraise_exception(): [ 698.127665] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 698.127665] env[61923]: self.force_reraise() [ 698.127665] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 698.127665] env[61923]: raise self.value [ 698.127665] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 698.127665] env[61923]: updated_port = self._update_port( [ 698.127665] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 698.127665] env[61923]: _ensure_no_port_binding_failure(port) [ 698.127665] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 698.127665] env[61923]: raise exception.PortBindingFailed(port_id=port['id']) [ 698.128535] env[61923]: nova.exception.PortBindingFailed: Binding failed for port 02c3270e-c070-4203-bf7e-d9f1db17ad7e, please check neutron logs for more information. [ 698.128535] env[61923]: Removing descriptor: 14 [ 698.238501] env[61923]: DEBUG oslo_concurrency.lockutils [None req-425a3606-915c-4f9b-8507-50f1c8d0d8d9 tempest-ServerRescueTestJSON-521767381 tempest-ServerRescueTestJSON-521767381-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 698.417582] env[61923]: DEBUG oslo_concurrency.lockutils [None req-44a76586-0f76-448f-af41-a5d846bfda81 tempest-ServerActionsV293TestJSON-849444864 tempest-ServerActionsV293TestJSON-849444864-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.460s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 698.418060] env[61923]: DEBUG nova.compute.manager [None req-44a76586-0f76-448f-af41-a5d846bfda81 tempest-ServerActionsV293TestJSON-849444864 tempest-ServerActionsV293TestJSON-849444864-project-member] [instance: 62144317-0354-4751-8f2b-4eae588c9da7] Start building networks asynchronously for instance. {{(pid=61923) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 698.420928] env[61923]: DEBUG oslo_concurrency.lockutils [None req-c09a27e9-8e33-411f-b17b-7cd8d66a175c tempest-ServerMetadataNegativeTestJSON-1976245157 tempest-ServerMetadataNegativeTestJSON-1976245157-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.457s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 698.422330] env[61923]: INFO nova.compute.claims [None req-c09a27e9-8e33-411f-b17b-7cd8d66a175c tempest-ServerMetadataNegativeTestJSON-1976245157 tempest-ServerMetadataNegativeTestJSON-1976245157-project-member] [instance: 25b62db6-6e64-4e38-bfd3-82a2e9a4883a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 698.454251] env[61923]: DEBUG nova.network.neutron [req-f7233aab-2d7d-4932-941a-6164985c2db1 req-8c926b96-731e-49f4-9ba2-c34ca252eade service nova] [instance: fb548893-2df5-4629-bf3a-3b0f807f6357] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 698.563719] env[61923]: DEBUG nova.network.neutron [req-f7233aab-2d7d-4932-941a-6164985c2db1 req-8c926b96-731e-49f4-9ba2-c34ca252eade service nova] [instance: fb548893-2df5-4629-bf3a-3b0f807f6357] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 698.929568] env[61923]: DEBUG nova.compute.utils [None req-44a76586-0f76-448f-af41-a5d846bfda81 tempest-ServerActionsV293TestJSON-849444864 tempest-ServerActionsV293TestJSON-849444864-project-member] Using /dev/sd instead of None {{(pid=61923) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 698.930929] env[61923]: DEBUG nova.compute.manager [None req-44a76586-0f76-448f-af41-a5d846bfda81 tempest-ServerActionsV293TestJSON-849444864 tempest-ServerActionsV293TestJSON-849444864-project-member] [instance: 62144317-0354-4751-8f2b-4eae588c9da7] Allocating IP information in the background. {{(pid=61923) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 698.931108] env[61923]: DEBUG nova.network.neutron [None req-44a76586-0f76-448f-af41-a5d846bfda81 tempest-ServerActionsV293TestJSON-849444864 tempest-ServerActionsV293TestJSON-849444864-project-member] [instance: 62144317-0354-4751-8f2b-4eae588c9da7] allocate_for_instance() {{(pid=61923) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 698.978942] env[61923]: DEBUG nova.policy [None req-44a76586-0f76-448f-af41-a5d846bfda81 tempest-ServerActionsV293TestJSON-849444864 tempest-ServerActionsV293TestJSON-849444864-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'fd29f14d28074b45a257d8772cfe6e6f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'fea9f06542bd46dcbe15841a081f03bc', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61923) authorize /opt/stack/nova/nova/policy.py:201}} [ 699.068339] env[61923]: DEBUG oslo_concurrency.lockutils [req-f7233aab-2d7d-4932-941a-6164985c2db1 req-8c926b96-731e-49f4-9ba2-c34ca252eade service nova] Releasing lock "refresh_cache-fb548893-2df5-4629-bf3a-3b0f807f6357" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 699.328867] env[61923]: DEBUG nova.network.neutron [None req-44a76586-0f76-448f-af41-a5d846bfda81 tempest-ServerActionsV293TestJSON-849444864 tempest-ServerActionsV293TestJSON-849444864-project-member] [instance: 62144317-0354-4751-8f2b-4eae588c9da7] Successfully created port: 6059f47a-1eef-4b32-9c0c-25cdd6ab2609 {{(pid=61923) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 699.435206] env[61923]: DEBUG nova.compute.manager [None req-44a76586-0f76-448f-af41-a5d846bfda81 tempest-ServerActionsV293TestJSON-849444864 tempest-ServerActionsV293TestJSON-849444864-project-member] [instance: 62144317-0354-4751-8f2b-4eae588c9da7] Start building block device mappings for instance. {{(pid=61923) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 699.714101] env[61923]: DEBUG nova.compute.manager [None req-df4eb7d8-5b64-4db4-b042-683b16e0b991 tempest-ServersTestBootFromVolume-2104901628 tempest-ServersTestBootFromVolume-2104901628-project-member] [instance: fb548893-2df5-4629-bf3a-3b0f807f6357] Start spawning the instance on the hypervisor. {{(pid=61923) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 699.714639] env[61923]: DEBUG nova.virt.hardware [None req-df4eb7d8-5b64-4db4-b042-683b16e0b991 tempest-ServersTestBootFromVolume-2104901628 tempest-ServersTestBootFromVolume-2104901628-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-29T20:07:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=0,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=1073741824,status='active',tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 699.714856] env[61923]: DEBUG nova.virt.hardware [None req-df4eb7d8-5b64-4db4-b042-683b16e0b991 tempest-ServersTestBootFromVolume-2104901628 tempest-ServersTestBootFromVolume-2104901628-project-member] Flavor limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 699.715012] env[61923]: DEBUG nova.virt.hardware [None req-df4eb7d8-5b64-4db4-b042-683b16e0b991 tempest-ServersTestBootFromVolume-2104901628 tempest-ServersTestBootFromVolume-2104901628-project-member] Image limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 699.715198] env[61923]: DEBUG nova.virt.hardware [None req-df4eb7d8-5b64-4db4-b042-683b16e0b991 tempest-ServersTestBootFromVolume-2104901628 tempest-ServersTestBootFromVolume-2104901628-project-member] Flavor pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 699.715349] env[61923]: DEBUG nova.virt.hardware [None req-df4eb7d8-5b64-4db4-b042-683b16e0b991 tempest-ServersTestBootFromVolume-2104901628 tempest-ServersTestBootFromVolume-2104901628-project-member] Image pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 699.715491] env[61923]: DEBUG nova.virt.hardware [None req-df4eb7d8-5b64-4db4-b042-683b16e0b991 tempest-ServersTestBootFromVolume-2104901628 tempest-ServersTestBootFromVolume-2104901628-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 699.715710] env[61923]: DEBUG nova.virt.hardware [None req-df4eb7d8-5b64-4db4-b042-683b16e0b991 tempest-ServersTestBootFromVolume-2104901628 tempest-ServersTestBootFromVolume-2104901628-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 699.715878] env[61923]: DEBUG nova.virt.hardware [None req-df4eb7d8-5b64-4db4-b042-683b16e0b991 tempest-ServersTestBootFromVolume-2104901628 tempest-ServersTestBootFromVolume-2104901628-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 699.716796] env[61923]: DEBUG nova.virt.hardware [None req-df4eb7d8-5b64-4db4-b042-683b16e0b991 tempest-ServersTestBootFromVolume-2104901628 tempest-ServersTestBootFromVolume-2104901628-project-member] Got 1 possible topologies {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 699.716796] env[61923]: DEBUG nova.virt.hardware [None req-df4eb7d8-5b64-4db4-b042-683b16e0b991 tempest-ServersTestBootFromVolume-2104901628 tempest-ServersTestBootFromVolume-2104901628-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 699.716796] env[61923]: DEBUG nova.virt.hardware [None req-df4eb7d8-5b64-4db4-b042-683b16e0b991 tempest-ServersTestBootFromVolume-2104901628 tempest-ServersTestBootFromVolume-2104901628-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 699.717307] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31763688-e734-4944-937b-5c2a7cd533b9 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.728656] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec027cbb-2bf5-4659-bea2-9f8f6ba195ba {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.744893] env[61923]: ERROR nova.compute.manager [None req-df4eb7d8-5b64-4db4-b042-683b16e0b991 tempest-ServersTestBootFromVolume-2104901628 tempest-ServersTestBootFromVolume-2104901628-project-member] [instance: fb548893-2df5-4629-bf3a-3b0f807f6357] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 02c3270e-c070-4203-bf7e-d9f1db17ad7e, please check neutron logs for more information. [ 699.744893] env[61923]: ERROR nova.compute.manager [instance: fb548893-2df5-4629-bf3a-3b0f807f6357] Traceback (most recent call last): [ 699.744893] env[61923]: ERROR nova.compute.manager [instance: fb548893-2df5-4629-bf3a-3b0f807f6357] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 699.744893] env[61923]: ERROR nova.compute.manager [instance: fb548893-2df5-4629-bf3a-3b0f807f6357] yield resources [ 699.744893] env[61923]: ERROR nova.compute.manager [instance: fb548893-2df5-4629-bf3a-3b0f807f6357] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 699.744893] env[61923]: ERROR nova.compute.manager [instance: fb548893-2df5-4629-bf3a-3b0f807f6357] self.driver.spawn(context, instance, image_meta, [ 699.744893] env[61923]: ERROR nova.compute.manager [instance: fb548893-2df5-4629-bf3a-3b0f807f6357] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 699.744893] env[61923]: ERROR nova.compute.manager [instance: fb548893-2df5-4629-bf3a-3b0f807f6357] self._vmops.spawn(context, instance, image_meta, injected_files, [ 699.744893] env[61923]: ERROR nova.compute.manager [instance: fb548893-2df5-4629-bf3a-3b0f807f6357] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 699.744893] env[61923]: ERROR nova.compute.manager [instance: fb548893-2df5-4629-bf3a-3b0f807f6357] vm_ref = self.build_virtual_machine(instance, [ 699.744893] env[61923]: ERROR nova.compute.manager [instance: fb548893-2df5-4629-bf3a-3b0f807f6357] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 699.745298] env[61923]: ERROR nova.compute.manager [instance: fb548893-2df5-4629-bf3a-3b0f807f6357] vif_infos = vmwarevif.get_vif_info(self._session, [ 699.745298] env[61923]: ERROR nova.compute.manager [instance: fb548893-2df5-4629-bf3a-3b0f807f6357] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 699.745298] env[61923]: ERROR nova.compute.manager [instance: fb548893-2df5-4629-bf3a-3b0f807f6357] for vif in network_info: [ 699.745298] env[61923]: ERROR nova.compute.manager [instance: fb548893-2df5-4629-bf3a-3b0f807f6357] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 699.745298] env[61923]: ERROR nova.compute.manager [instance: fb548893-2df5-4629-bf3a-3b0f807f6357] return self._sync_wrapper(fn, *args, **kwargs) [ 699.745298] env[61923]: ERROR nova.compute.manager [instance: fb548893-2df5-4629-bf3a-3b0f807f6357] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 699.745298] env[61923]: ERROR nova.compute.manager [instance: fb548893-2df5-4629-bf3a-3b0f807f6357] self.wait() [ 699.745298] env[61923]: ERROR nova.compute.manager [instance: fb548893-2df5-4629-bf3a-3b0f807f6357] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 699.745298] env[61923]: ERROR nova.compute.manager [instance: fb548893-2df5-4629-bf3a-3b0f807f6357] self[:] = self._gt.wait() [ 699.745298] env[61923]: ERROR nova.compute.manager [instance: fb548893-2df5-4629-bf3a-3b0f807f6357] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 699.745298] env[61923]: ERROR nova.compute.manager [instance: fb548893-2df5-4629-bf3a-3b0f807f6357] return self._exit_event.wait() [ 699.745298] env[61923]: ERROR nova.compute.manager [instance: fb548893-2df5-4629-bf3a-3b0f807f6357] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 699.745298] env[61923]: ERROR nova.compute.manager [instance: fb548893-2df5-4629-bf3a-3b0f807f6357] current.throw(*self._exc) [ 699.745708] env[61923]: ERROR nova.compute.manager [instance: fb548893-2df5-4629-bf3a-3b0f807f6357] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 699.745708] env[61923]: ERROR nova.compute.manager [instance: fb548893-2df5-4629-bf3a-3b0f807f6357] result = function(*args, **kwargs) [ 699.745708] env[61923]: ERROR nova.compute.manager [instance: fb548893-2df5-4629-bf3a-3b0f807f6357] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 699.745708] env[61923]: ERROR nova.compute.manager [instance: fb548893-2df5-4629-bf3a-3b0f807f6357] return func(*args, **kwargs) [ 699.745708] env[61923]: ERROR nova.compute.manager [instance: fb548893-2df5-4629-bf3a-3b0f807f6357] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 699.745708] env[61923]: ERROR nova.compute.manager [instance: fb548893-2df5-4629-bf3a-3b0f807f6357] raise e [ 699.745708] env[61923]: ERROR nova.compute.manager [instance: fb548893-2df5-4629-bf3a-3b0f807f6357] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 699.745708] env[61923]: ERROR nova.compute.manager [instance: fb548893-2df5-4629-bf3a-3b0f807f6357] nwinfo = self.network_api.allocate_for_instance( [ 699.745708] env[61923]: ERROR nova.compute.manager [instance: fb548893-2df5-4629-bf3a-3b0f807f6357] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 699.745708] env[61923]: ERROR nova.compute.manager [instance: fb548893-2df5-4629-bf3a-3b0f807f6357] created_port_ids = self._update_ports_for_instance( [ 699.745708] env[61923]: ERROR nova.compute.manager [instance: fb548893-2df5-4629-bf3a-3b0f807f6357] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 699.745708] env[61923]: ERROR nova.compute.manager [instance: fb548893-2df5-4629-bf3a-3b0f807f6357] with excutils.save_and_reraise_exception(): [ 699.745708] env[61923]: ERROR nova.compute.manager [instance: fb548893-2df5-4629-bf3a-3b0f807f6357] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 699.746108] env[61923]: ERROR nova.compute.manager [instance: fb548893-2df5-4629-bf3a-3b0f807f6357] self.force_reraise() [ 699.746108] env[61923]: ERROR nova.compute.manager [instance: fb548893-2df5-4629-bf3a-3b0f807f6357] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 699.746108] env[61923]: ERROR nova.compute.manager [instance: fb548893-2df5-4629-bf3a-3b0f807f6357] raise self.value [ 699.746108] env[61923]: ERROR nova.compute.manager [instance: fb548893-2df5-4629-bf3a-3b0f807f6357] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 699.746108] env[61923]: ERROR nova.compute.manager [instance: fb548893-2df5-4629-bf3a-3b0f807f6357] updated_port = self._update_port( [ 699.746108] env[61923]: ERROR nova.compute.manager [instance: fb548893-2df5-4629-bf3a-3b0f807f6357] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 699.746108] env[61923]: ERROR nova.compute.manager [instance: fb548893-2df5-4629-bf3a-3b0f807f6357] _ensure_no_port_binding_failure(port) [ 699.746108] env[61923]: ERROR nova.compute.manager [instance: fb548893-2df5-4629-bf3a-3b0f807f6357] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 699.746108] env[61923]: ERROR nova.compute.manager [instance: fb548893-2df5-4629-bf3a-3b0f807f6357] raise exception.PortBindingFailed(port_id=port['id']) [ 699.746108] env[61923]: ERROR nova.compute.manager [instance: fb548893-2df5-4629-bf3a-3b0f807f6357] nova.exception.PortBindingFailed: Binding failed for port 02c3270e-c070-4203-bf7e-d9f1db17ad7e, please check neutron logs for more information. [ 699.746108] env[61923]: ERROR nova.compute.manager [instance: fb548893-2df5-4629-bf3a-3b0f807f6357] [ 699.746108] env[61923]: INFO nova.compute.manager [None req-df4eb7d8-5b64-4db4-b042-683b16e0b991 tempest-ServersTestBootFromVolume-2104901628 tempest-ServersTestBootFromVolume-2104901628-project-member] [instance: fb548893-2df5-4629-bf3a-3b0f807f6357] Terminating instance [ 699.749210] env[61923]: DEBUG oslo_concurrency.lockutils [None req-df4eb7d8-5b64-4db4-b042-683b16e0b991 tempest-ServersTestBootFromVolume-2104901628 tempest-ServersTestBootFromVolume-2104901628-project-member] Acquiring lock "refresh_cache-fb548893-2df5-4629-bf3a-3b0f807f6357" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 699.749887] env[61923]: DEBUG oslo_concurrency.lockutils [None req-df4eb7d8-5b64-4db4-b042-683b16e0b991 tempest-ServersTestBootFromVolume-2104901628 tempest-ServersTestBootFromVolume-2104901628-project-member] Acquired lock "refresh_cache-fb548893-2df5-4629-bf3a-3b0f807f6357" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 699.749887] env[61923]: DEBUG nova.network.neutron [None req-df4eb7d8-5b64-4db4-b042-683b16e0b991 tempest-ServersTestBootFromVolume-2104901628 tempest-ServersTestBootFromVolume-2104901628-project-member] [instance: fb548893-2df5-4629-bf3a-3b0f807f6357] Building network info cache for instance {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 699.805216] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4fe7de30-bbeb-4afa-bd6b-231dd4812f10 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.812604] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c5dffce-64d8-4bac-b117-6ed4f325e04f {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.841761] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16119fa2-63b7-40be-9f0c-101d304c1a9a {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.848614] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ee42679-ac3a-4680-8b8c-9842626ddc89 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.861174] env[61923]: DEBUG nova.compute.provider_tree [None req-c09a27e9-8e33-411f-b17b-7cd8d66a175c tempest-ServerMetadataNegativeTestJSON-1976245157 tempest-ServerMetadataNegativeTestJSON-1976245157-project-member] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 699.941458] env[61923]: INFO nova.virt.block_device [None req-44a76586-0f76-448f-af41-a5d846bfda81 tempest-ServerActionsV293TestJSON-849444864 tempest-ServerActionsV293TestJSON-849444864-project-member] [instance: 62144317-0354-4751-8f2b-4eae588c9da7] Booting with volume 42937a2d-433b-4f1d-b30f-60189aa93e20 at /dev/sda [ 699.974907] env[61923]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-1b283e44-a114-4f40-886e-fc5f2be72719 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.987591] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d3b31bd-ba41-49f4-8923-a48d54788cbd {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 699.999821] env[61923]: DEBUG nova.compute.manager [req-eeca530e-93db-427b-bf51-3df5df0d2904 req-9594d589-8f0a-4617-b8c2-20fbe1656a08 service nova] [instance: fb548893-2df5-4629-bf3a-3b0f807f6357] Received event network-vif-deleted-02c3270e-c070-4203-bf7e-d9f1db17ad7e {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 700.013225] env[61923]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-621ca485-33ee-4f18-965c-124e923fcb2f {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.020527] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e95e0b7-c76c-4a8a-9c70-fbe5b70cc205 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.042084] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ce13c21-dc05-43e0-8e87-e711c89e78a1 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.048559] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-731c3960-5ddd-4caf-9342-b319eeb3580e {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.060906] env[61923]: DEBUG nova.virt.block_device [None req-44a76586-0f76-448f-af41-a5d846bfda81 tempest-ServerActionsV293TestJSON-849444864 tempest-ServerActionsV293TestJSON-849444864-project-member] [instance: 62144317-0354-4751-8f2b-4eae588c9da7] Updating existing volume attachment record: 0d6b198e-67aa-466b-9ff1-4174dadf0076 {{(pid=61923) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 700.271814] env[61923]: DEBUG nova.network.neutron [None req-df4eb7d8-5b64-4db4-b042-683b16e0b991 tempest-ServersTestBootFromVolume-2104901628 tempest-ServersTestBootFromVolume-2104901628-project-member] [instance: fb548893-2df5-4629-bf3a-3b0f807f6357] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 700.366878] env[61923]: DEBUG nova.scheduler.client.report [None req-c09a27e9-8e33-411f-b17b-7cd8d66a175c tempest-ServerMetadataNegativeTestJSON-1976245157 tempest-ServerMetadataNegativeTestJSON-1976245157-project-member] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 700.460467] env[61923]: DEBUG nova.network.neutron [None req-df4eb7d8-5b64-4db4-b042-683b16e0b991 tempest-ServersTestBootFromVolume-2104901628 tempest-ServersTestBootFromVolume-2104901628-project-member] [instance: fb548893-2df5-4629-bf3a-3b0f807f6357] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 700.572292] env[61923]: ERROR nova.compute.manager [None req-44a76586-0f76-448f-af41-a5d846bfda81 tempest-ServerActionsV293TestJSON-849444864 tempest-ServerActionsV293TestJSON-849444864-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 6059f47a-1eef-4b32-9c0c-25cdd6ab2609, please check neutron logs for more information. [ 700.572292] env[61923]: ERROR nova.compute.manager Traceback (most recent call last): [ 700.572292] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 700.572292] env[61923]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 700.572292] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 700.572292] env[61923]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 700.572292] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 700.572292] env[61923]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 700.572292] env[61923]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 700.572292] env[61923]: ERROR nova.compute.manager self.force_reraise() [ 700.572292] env[61923]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 700.572292] env[61923]: ERROR nova.compute.manager raise self.value [ 700.572292] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 700.572292] env[61923]: ERROR nova.compute.manager updated_port = self._update_port( [ 700.572292] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 700.572292] env[61923]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 700.572874] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 700.572874] env[61923]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 700.572874] env[61923]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 6059f47a-1eef-4b32-9c0c-25cdd6ab2609, please check neutron logs for more information. [ 700.572874] env[61923]: ERROR nova.compute.manager [ 700.572874] env[61923]: Traceback (most recent call last): [ 700.572874] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 700.572874] env[61923]: listener.cb(fileno) [ 700.572874] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 700.572874] env[61923]: result = function(*args, **kwargs) [ 700.572874] env[61923]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 700.572874] env[61923]: return func(*args, **kwargs) [ 700.572874] env[61923]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 700.572874] env[61923]: raise e [ 700.572874] env[61923]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 700.572874] env[61923]: nwinfo = self.network_api.allocate_for_instance( [ 700.572874] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 700.572874] env[61923]: created_port_ids = self._update_ports_for_instance( [ 700.572874] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 700.572874] env[61923]: with excutils.save_and_reraise_exception(): [ 700.572874] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 700.572874] env[61923]: self.force_reraise() [ 700.572874] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 700.572874] env[61923]: raise self.value [ 700.572874] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 700.572874] env[61923]: updated_port = self._update_port( [ 700.572874] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 700.572874] env[61923]: _ensure_no_port_binding_failure(port) [ 700.572874] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 700.572874] env[61923]: raise exception.PortBindingFailed(port_id=port['id']) [ 700.573719] env[61923]: nova.exception.PortBindingFailed: Binding failed for port 6059f47a-1eef-4b32-9c0c-25cdd6ab2609, please check neutron logs for more information. [ 700.573719] env[61923]: Removing descriptor: 14 [ 700.869299] env[61923]: DEBUG oslo_concurrency.lockutils [None req-c09a27e9-8e33-411f-b17b-7cd8d66a175c tempest-ServerMetadataNegativeTestJSON-1976245157 tempest-ServerMetadataNegativeTestJSON-1976245157-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.448s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 700.869835] env[61923]: DEBUG nova.compute.manager [None req-c09a27e9-8e33-411f-b17b-7cd8d66a175c tempest-ServerMetadataNegativeTestJSON-1976245157 tempest-ServerMetadataNegativeTestJSON-1976245157-project-member] [instance: 25b62db6-6e64-4e38-bfd3-82a2e9a4883a] Start building networks asynchronously for instance. {{(pid=61923) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 700.873023] env[61923]: DEBUG oslo_concurrency.lockutils [None req-1f24f04c-cf4b-4940-9447-eacc11fa5e66 tempest-ServersTestFqdnHostnames-345291565 tempest-ServersTestFqdnHostnames-345291565-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.976s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 700.961801] env[61923]: DEBUG oslo_concurrency.lockutils [None req-df4eb7d8-5b64-4db4-b042-683b16e0b991 tempest-ServersTestBootFromVolume-2104901628 tempest-ServersTestBootFromVolume-2104901628-project-member] Releasing lock "refresh_cache-fb548893-2df5-4629-bf3a-3b0f807f6357" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 700.962415] env[61923]: DEBUG nova.compute.manager [None req-df4eb7d8-5b64-4db4-b042-683b16e0b991 tempest-ServersTestBootFromVolume-2104901628 tempest-ServersTestBootFromVolume-2104901628-project-member] [instance: fb548893-2df5-4629-bf3a-3b0f807f6357] Start destroying the instance on the hypervisor. {{(pid=61923) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 700.962739] env[61923]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-040e3f96-05e9-43d7-a06d-8a800bf30afa {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.971714] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bad7be5b-ee9f-496f-92db-0fc6a1610678 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 700.993877] env[61923]: WARNING nova.virt.vmwareapi.driver [None req-df4eb7d8-5b64-4db4-b042-683b16e0b991 tempest-ServersTestBootFromVolume-2104901628 tempest-ServersTestBootFromVolume-2104901628-project-member] [instance: fb548893-2df5-4629-bf3a-3b0f807f6357] Instance does not exists. Proceeding to delete instance properties on datastore: nova.exception.InstanceNotFound: Instance fb548893-2df5-4629-bf3a-3b0f807f6357 could not be found. [ 700.994115] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-df4eb7d8-5b64-4db4-b042-683b16e0b991 tempest-ServersTestBootFromVolume-2104901628 tempest-ServersTestBootFromVolume-2104901628-project-member] [instance: fb548893-2df5-4629-bf3a-3b0f807f6357] Destroying instance {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 700.994397] env[61923]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ef2fa0ca-01af-4828-88ee-9cb4aea7592c {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.001772] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c71a01c-a210-496e-b30c-e573d1c6878e {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.023285] env[61923]: WARNING nova.virt.vmwareapi.vmops [None req-df4eb7d8-5b64-4db4-b042-683b16e0b991 tempest-ServersTestBootFromVolume-2104901628 tempest-ServersTestBootFromVolume-2104901628-project-member] [instance: fb548893-2df5-4629-bf3a-3b0f807f6357] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance fb548893-2df5-4629-bf3a-3b0f807f6357 could not be found. [ 701.023518] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-df4eb7d8-5b64-4db4-b042-683b16e0b991 tempest-ServersTestBootFromVolume-2104901628 tempest-ServersTestBootFromVolume-2104901628-project-member] [instance: fb548893-2df5-4629-bf3a-3b0f807f6357] Instance destroyed {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 701.023701] env[61923]: INFO nova.compute.manager [None req-df4eb7d8-5b64-4db4-b042-683b16e0b991 tempest-ServersTestBootFromVolume-2104901628 tempest-ServersTestBootFromVolume-2104901628-project-member] [instance: fb548893-2df5-4629-bf3a-3b0f807f6357] Took 0.06 seconds to destroy the instance on the hypervisor. [ 701.023942] env[61923]: DEBUG oslo.service.loopingcall [None req-df4eb7d8-5b64-4db4-b042-683b16e0b991 tempest-ServersTestBootFromVolume-2104901628 tempest-ServersTestBootFromVolume-2104901628-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61923) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 701.024178] env[61923]: DEBUG nova.compute.manager [-] [instance: fb548893-2df5-4629-bf3a-3b0f807f6357] Deallocating network for instance {{(pid=61923) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 701.024278] env[61923]: DEBUG nova.network.neutron [-] [instance: fb548893-2df5-4629-bf3a-3b0f807f6357] deallocate_for_instance() {{(pid=61923) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 701.141808] env[61923]: DEBUG nova.network.neutron [-] [instance: fb548893-2df5-4629-bf3a-3b0f807f6357] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 701.377329] env[61923]: DEBUG nova.compute.utils [None req-c09a27e9-8e33-411f-b17b-7cd8d66a175c tempest-ServerMetadataNegativeTestJSON-1976245157 tempest-ServerMetadataNegativeTestJSON-1976245157-project-member] Using /dev/sd instead of None {{(pid=61923) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 701.382009] env[61923]: DEBUG nova.compute.manager [None req-c09a27e9-8e33-411f-b17b-7cd8d66a175c tempest-ServerMetadataNegativeTestJSON-1976245157 tempest-ServerMetadataNegativeTestJSON-1976245157-project-member] [instance: 25b62db6-6e64-4e38-bfd3-82a2e9a4883a] Allocating IP information in the background. {{(pid=61923) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 701.382009] env[61923]: DEBUG nova.network.neutron [None req-c09a27e9-8e33-411f-b17b-7cd8d66a175c tempest-ServerMetadataNegativeTestJSON-1976245157 tempest-ServerMetadataNegativeTestJSON-1976245157-project-member] [instance: 25b62db6-6e64-4e38-bfd3-82a2e9a4883a] allocate_for_instance() {{(pid=61923) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 701.429929] env[61923]: DEBUG nova.policy [None req-c09a27e9-8e33-411f-b17b-7cd8d66a175c tempest-ServerMetadataNegativeTestJSON-1976245157 tempest-ServerMetadataNegativeTestJSON-1976245157-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '1be633de635c474da4211a1d17fc628e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b7d4b45a6d8e4ab89f78ef68ac1d02ac', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61923) authorize /opt/stack/nova/nova/policy.py:201}} [ 701.644309] env[61923]: DEBUG nova.network.neutron [-] [instance: fb548893-2df5-4629-bf3a-3b0f807f6357] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 701.714995] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02ff7bdd-c138-407f-ba06-e6ad3838bdc3 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.722022] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8abdb423-9f4c-47d8-a4d1-6dc190b51d8b {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.752384] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5d13b80-b4e1-47b8-ac67-70f428a9062e {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.759798] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5522835f-146f-44ae-aa43-91ece39619b0 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 701.773145] env[61923]: DEBUG nova.compute.provider_tree [None req-1f24f04c-cf4b-4940-9447-eacc11fa5e66 tempest-ServersTestFqdnHostnames-345291565 tempest-ServersTestFqdnHostnames-345291565-project-member] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 701.811728] env[61923]: DEBUG nova.network.neutron [None req-c09a27e9-8e33-411f-b17b-7cd8d66a175c tempest-ServerMetadataNegativeTestJSON-1976245157 tempest-ServerMetadataNegativeTestJSON-1976245157-project-member] [instance: 25b62db6-6e64-4e38-bfd3-82a2e9a4883a] Successfully created port: bf236ea4-19e2-487c-ad15-61be750ff9c5 {{(pid=61923) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 701.882650] env[61923]: DEBUG nova.compute.manager [None req-c09a27e9-8e33-411f-b17b-7cd8d66a175c tempest-ServerMetadataNegativeTestJSON-1976245157 tempest-ServerMetadataNegativeTestJSON-1976245157-project-member] [instance: 25b62db6-6e64-4e38-bfd3-82a2e9a4883a] Start building block device mappings for instance. {{(pid=61923) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 702.147809] env[61923]: DEBUG nova.compute.manager [None req-44a76586-0f76-448f-af41-a5d846bfda81 tempest-ServerActionsV293TestJSON-849444864 tempest-ServerActionsV293TestJSON-849444864-project-member] [instance: 62144317-0354-4751-8f2b-4eae588c9da7] Start spawning the instance on the hypervisor. {{(pid=61923) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 702.148367] env[61923]: DEBUG nova.virt.hardware [None req-44a76586-0f76-448f-af41-a5d846bfda81 tempest-ServerActionsV293TestJSON-849444864 tempest-ServerActionsV293TestJSON-849444864-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-29T20:07:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=0,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=1073741824,status='active',tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 702.148576] env[61923]: DEBUG nova.virt.hardware [None req-44a76586-0f76-448f-af41-a5d846bfda81 tempest-ServerActionsV293TestJSON-849444864 tempest-ServerActionsV293TestJSON-849444864-project-member] Flavor limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 702.148728] env[61923]: DEBUG nova.virt.hardware [None req-44a76586-0f76-448f-af41-a5d846bfda81 tempest-ServerActionsV293TestJSON-849444864 tempest-ServerActionsV293TestJSON-849444864-project-member] Image limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 702.148902] env[61923]: DEBUG nova.virt.hardware [None req-44a76586-0f76-448f-af41-a5d846bfda81 tempest-ServerActionsV293TestJSON-849444864 tempest-ServerActionsV293TestJSON-849444864-project-member] Flavor pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 702.149053] env[61923]: DEBUG nova.virt.hardware [None req-44a76586-0f76-448f-af41-a5d846bfda81 tempest-ServerActionsV293TestJSON-849444864 tempest-ServerActionsV293TestJSON-849444864-project-member] Image pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 702.149197] env[61923]: DEBUG nova.virt.hardware [None req-44a76586-0f76-448f-af41-a5d846bfda81 tempest-ServerActionsV293TestJSON-849444864 tempest-ServerActionsV293TestJSON-849444864-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 702.149393] env[61923]: DEBUG nova.virt.hardware [None req-44a76586-0f76-448f-af41-a5d846bfda81 tempest-ServerActionsV293TestJSON-849444864 tempest-ServerActionsV293TestJSON-849444864-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 702.149545] env[61923]: DEBUG nova.virt.hardware [None req-44a76586-0f76-448f-af41-a5d846bfda81 tempest-ServerActionsV293TestJSON-849444864 tempest-ServerActionsV293TestJSON-849444864-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 702.149707] env[61923]: DEBUG nova.virt.hardware [None req-44a76586-0f76-448f-af41-a5d846bfda81 tempest-ServerActionsV293TestJSON-849444864 tempest-ServerActionsV293TestJSON-849444864-project-member] Got 1 possible topologies {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 702.149864] env[61923]: DEBUG nova.virt.hardware [None req-44a76586-0f76-448f-af41-a5d846bfda81 tempest-ServerActionsV293TestJSON-849444864 tempest-ServerActionsV293TestJSON-849444864-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 702.150038] env[61923]: DEBUG nova.virt.hardware [None req-44a76586-0f76-448f-af41-a5d846bfda81 tempest-ServerActionsV293TestJSON-849444864 tempest-ServerActionsV293TestJSON-849444864-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 702.150420] env[61923]: INFO nova.compute.manager [-] [instance: fb548893-2df5-4629-bf3a-3b0f807f6357] Took 1.13 seconds to deallocate network for instance. [ 702.151247] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4725550-4b4c-4666-acc1-922ea01f3d5d {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.167291] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4865c7ba-c630-4611-a27d-18de98ca6e22 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.188119] env[61923]: ERROR nova.compute.manager [None req-44a76586-0f76-448f-af41-a5d846bfda81 tempest-ServerActionsV293TestJSON-849444864 tempest-ServerActionsV293TestJSON-849444864-project-member] [instance: 62144317-0354-4751-8f2b-4eae588c9da7] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 6059f47a-1eef-4b32-9c0c-25cdd6ab2609, please check neutron logs for more information. [ 702.188119] env[61923]: ERROR nova.compute.manager [instance: 62144317-0354-4751-8f2b-4eae588c9da7] Traceback (most recent call last): [ 702.188119] env[61923]: ERROR nova.compute.manager [instance: 62144317-0354-4751-8f2b-4eae588c9da7] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 702.188119] env[61923]: ERROR nova.compute.manager [instance: 62144317-0354-4751-8f2b-4eae588c9da7] yield resources [ 702.188119] env[61923]: ERROR nova.compute.manager [instance: 62144317-0354-4751-8f2b-4eae588c9da7] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 702.188119] env[61923]: ERROR nova.compute.manager [instance: 62144317-0354-4751-8f2b-4eae588c9da7] self.driver.spawn(context, instance, image_meta, [ 702.188119] env[61923]: ERROR nova.compute.manager [instance: 62144317-0354-4751-8f2b-4eae588c9da7] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 702.188119] env[61923]: ERROR nova.compute.manager [instance: 62144317-0354-4751-8f2b-4eae588c9da7] self._vmops.spawn(context, instance, image_meta, injected_files, [ 702.188119] env[61923]: ERROR nova.compute.manager [instance: 62144317-0354-4751-8f2b-4eae588c9da7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 702.188119] env[61923]: ERROR nova.compute.manager [instance: 62144317-0354-4751-8f2b-4eae588c9da7] vm_ref = self.build_virtual_machine(instance, [ 702.188119] env[61923]: ERROR nova.compute.manager [instance: 62144317-0354-4751-8f2b-4eae588c9da7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 702.188455] env[61923]: ERROR nova.compute.manager [instance: 62144317-0354-4751-8f2b-4eae588c9da7] vif_infos = vmwarevif.get_vif_info(self._session, [ 702.188455] env[61923]: ERROR nova.compute.manager [instance: 62144317-0354-4751-8f2b-4eae588c9da7] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 702.188455] env[61923]: ERROR nova.compute.manager [instance: 62144317-0354-4751-8f2b-4eae588c9da7] for vif in network_info: [ 702.188455] env[61923]: ERROR nova.compute.manager [instance: 62144317-0354-4751-8f2b-4eae588c9da7] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 702.188455] env[61923]: ERROR nova.compute.manager [instance: 62144317-0354-4751-8f2b-4eae588c9da7] return self._sync_wrapper(fn, *args, **kwargs) [ 702.188455] env[61923]: ERROR nova.compute.manager [instance: 62144317-0354-4751-8f2b-4eae588c9da7] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 702.188455] env[61923]: ERROR nova.compute.manager [instance: 62144317-0354-4751-8f2b-4eae588c9da7] self.wait() [ 702.188455] env[61923]: ERROR nova.compute.manager [instance: 62144317-0354-4751-8f2b-4eae588c9da7] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 702.188455] env[61923]: ERROR nova.compute.manager [instance: 62144317-0354-4751-8f2b-4eae588c9da7] self[:] = self._gt.wait() [ 702.188455] env[61923]: ERROR nova.compute.manager [instance: 62144317-0354-4751-8f2b-4eae588c9da7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 702.188455] env[61923]: ERROR nova.compute.manager [instance: 62144317-0354-4751-8f2b-4eae588c9da7] return self._exit_event.wait() [ 702.188455] env[61923]: ERROR nova.compute.manager [instance: 62144317-0354-4751-8f2b-4eae588c9da7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 702.188455] env[61923]: ERROR nova.compute.manager [instance: 62144317-0354-4751-8f2b-4eae588c9da7] current.throw(*self._exc) [ 702.188840] env[61923]: ERROR nova.compute.manager [instance: 62144317-0354-4751-8f2b-4eae588c9da7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 702.188840] env[61923]: ERROR nova.compute.manager [instance: 62144317-0354-4751-8f2b-4eae588c9da7] result = function(*args, **kwargs) [ 702.188840] env[61923]: ERROR nova.compute.manager [instance: 62144317-0354-4751-8f2b-4eae588c9da7] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 702.188840] env[61923]: ERROR nova.compute.manager [instance: 62144317-0354-4751-8f2b-4eae588c9da7] return func(*args, **kwargs) [ 702.188840] env[61923]: ERROR nova.compute.manager [instance: 62144317-0354-4751-8f2b-4eae588c9da7] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 702.188840] env[61923]: ERROR nova.compute.manager [instance: 62144317-0354-4751-8f2b-4eae588c9da7] raise e [ 702.188840] env[61923]: ERROR nova.compute.manager [instance: 62144317-0354-4751-8f2b-4eae588c9da7] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 702.188840] env[61923]: ERROR nova.compute.manager [instance: 62144317-0354-4751-8f2b-4eae588c9da7] nwinfo = self.network_api.allocate_for_instance( [ 702.188840] env[61923]: ERROR nova.compute.manager [instance: 62144317-0354-4751-8f2b-4eae588c9da7] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 702.188840] env[61923]: ERROR nova.compute.manager [instance: 62144317-0354-4751-8f2b-4eae588c9da7] created_port_ids = self._update_ports_for_instance( [ 702.188840] env[61923]: ERROR nova.compute.manager [instance: 62144317-0354-4751-8f2b-4eae588c9da7] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 702.188840] env[61923]: ERROR nova.compute.manager [instance: 62144317-0354-4751-8f2b-4eae588c9da7] with excutils.save_and_reraise_exception(): [ 702.188840] env[61923]: ERROR nova.compute.manager [instance: 62144317-0354-4751-8f2b-4eae588c9da7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 702.189210] env[61923]: ERROR nova.compute.manager [instance: 62144317-0354-4751-8f2b-4eae588c9da7] self.force_reraise() [ 702.189210] env[61923]: ERROR nova.compute.manager [instance: 62144317-0354-4751-8f2b-4eae588c9da7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 702.189210] env[61923]: ERROR nova.compute.manager [instance: 62144317-0354-4751-8f2b-4eae588c9da7] raise self.value [ 702.189210] env[61923]: ERROR nova.compute.manager [instance: 62144317-0354-4751-8f2b-4eae588c9da7] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 702.189210] env[61923]: ERROR nova.compute.manager [instance: 62144317-0354-4751-8f2b-4eae588c9da7] updated_port = self._update_port( [ 702.189210] env[61923]: ERROR nova.compute.manager [instance: 62144317-0354-4751-8f2b-4eae588c9da7] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 702.189210] env[61923]: ERROR nova.compute.manager [instance: 62144317-0354-4751-8f2b-4eae588c9da7] _ensure_no_port_binding_failure(port) [ 702.189210] env[61923]: ERROR nova.compute.manager [instance: 62144317-0354-4751-8f2b-4eae588c9da7] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 702.189210] env[61923]: ERROR nova.compute.manager [instance: 62144317-0354-4751-8f2b-4eae588c9da7] raise exception.PortBindingFailed(port_id=port['id']) [ 702.189210] env[61923]: ERROR nova.compute.manager [instance: 62144317-0354-4751-8f2b-4eae588c9da7] nova.exception.PortBindingFailed: Binding failed for port 6059f47a-1eef-4b32-9c0c-25cdd6ab2609, please check neutron logs for more information. [ 702.189210] env[61923]: ERROR nova.compute.manager [instance: 62144317-0354-4751-8f2b-4eae588c9da7] [ 702.189210] env[61923]: INFO nova.compute.manager [None req-44a76586-0f76-448f-af41-a5d846bfda81 tempest-ServerActionsV293TestJSON-849444864 tempest-ServerActionsV293TestJSON-849444864-project-member] [instance: 62144317-0354-4751-8f2b-4eae588c9da7] Terminating instance [ 702.190576] env[61923]: DEBUG oslo_concurrency.lockutils [None req-44a76586-0f76-448f-af41-a5d846bfda81 tempest-ServerActionsV293TestJSON-849444864 tempest-ServerActionsV293TestJSON-849444864-project-member] Acquiring lock "refresh_cache-62144317-0354-4751-8f2b-4eae588c9da7" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 702.190945] env[61923]: DEBUG oslo_concurrency.lockutils [None req-44a76586-0f76-448f-af41-a5d846bfda81 tempest-ServerActionsV293TestJSON-849444864 tempest-ServerActionsV293TestJSON-849444864-project-member] Acquired lock "refresh_cache-62144317-0354-4751-8f2b-4eae588c9da7" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 702.191146] env[61923]: DEBUG nova.network.neutron [None req-44a76586-0f76-448f-af41-a5d846bfda81 tempest-ServerActionsV293TestJSON-849444864 tempest-ServerActionsV293TestJSON-849444864-project-member] [instance: 62144317-0354-4751-8f2b-4eae588c9da7] Building network info cache for instance {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 702.210813] env[61923]: DEBUG nova.compute.manager [req-b9f2787e-ad0b-4abf-9b7b-2d13ff5feb43 req-1012b622-79a5-4dad-9068-333911395cfe service nova] [instance: 62144317-0354-4751-8f2b-4eae588c9da7] Received event network-changed-6059f47a-1eef-4b32-9c0c-25cdd6ab2609 {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 702.211018] env[61923]: DEBUG nova.compute.manager [req-b9f2787e-ad0b-4abf-9b7b-2d13ff5feb43 req-1012b622-79a5-4dad-9068-333911395cfe service nova] [instance: 62144317-0354-4751-8f2b-4eae588c9da7] Refreshing instance network info cache due to event network-changed-6059f47a-1eef-4b32-9c0c-25cdd6ab2609. {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 702.211206] env[61923]: DEBUG oslo_concurrency.lockutils [req-b9f2787e-ad0b-4abf-9b7b-2d13ff5feb43 req-1012b622-79a5-4dad-9068-333911395cfe service nova] Acquiring lock "refresh_cache-62144317-0354-4751-8f2b-4eae588c9da7" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 702.276108] env[61923]: DEBUG nova.scheduler.client.report [None req-1f24f04c-cf4b-4940-9447-eacc11fa5e66 tempest-ServersTestFqdnHostnames-345291565 tempest-ServersTestFqdnHostnames-345291565-project-member] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 702.726202] env[61923]: DEBUG nova.network.neutron [None req-44a76586-0f76-448f-af41-a5d846bfda81 tempest-ServerActionsV293TestJSON-849444864 tempest-ServerActionsV293TestJSON-849444864-project-member] [instance: 62144317-0354-4751-8f2b-4eae588c9da7] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 702.745022] env[61923]: INFO nova.compute.manager [None req-df4eb7d8-5b64-4db4-b042-683b16e0b991 tempest-ServersTestBootFromVolume-2104901628 tempest-ServersTestBootFromVolume-2104901628-project-member] [instance: fb548893-2df5-4629-bf3a-3b0f807f6357] Took 0.59 seconds to detach 1 volumes for instance. [ 702.747569] env[61923]: DEBUG nova.compute.claims [None req-df4eb7d8-5b64-4db4-b042-683b16e0b991 tempest-ServersTestBootFromVolume-2104901628 tempest-ServersTestBootFromVolume-2104901628-project-member] [instance: fb548893-2df5-4629-bf3a-3b0f807f6357] Aborting claim: {{(pid=61923) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 702.748028] env[61923]: DEBUG oslo_concurrency.lockutils [None req-df4eb7d8-5b64-4db4-b042-683b16e0b991 tempest-ServersTestBootFromVolume-2104901628 tempest-ServersTestBootFromVolume-2104901628-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 702.780788] env[61923]: DEBUG oslo_concurrency.lockutils [None req-1f24f04c-cf4b-4940-9447-eacc11fa5e66 tempest-ServersTestFqdnHostnames-345291565 tempest-ServersTestFqdnHostnames-345291565-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.908s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 702.781433] env[61923]: ERROR nova.compute.manager [None req-1f24f04c-cf4b-4940-9447-eacc11fa5e66 tempest-ServersTestFqdnHostnames-345291565 tempest-ServersTestFqdnHostnames-345291565-project-member] [instance: c94e8e46-7697-426c-ae2f-aece493fa8f4] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 70370c2b-9eb9-48b7-8acc-8a7eafc1d41d, please check neutron logs for more information. [ 702.781433] env[61923]: ERROR nova.compute.manager [instance: c94e8e46-7697-426c-ae2f-aece493fa8f4] Traceback (most recent call last): [ 702.781433] env[61923]: ERROR nova.compute.manager [instance: c94e8e46-7697-426c-ae2f-aece493fa8f4] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 702.781433] env[61923]: ERROR nova.compute.manager [instance: c94e8e46-7697-426c-ae2f-aece493fa8f4] self.driver.spawn(context, instance, image_meta, [ 702.781433] env[61923]: ERROR nova.compute.manager [instance: c94e8e46-7697-426c-ae2f-aece493fa8f4] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 702.781433] env[61923]: ERROR nova.compute.manager [instance: c94e8e46-7697-426c-ae2f-aece493fa8f4] self._vmops.spawn(context, instance, image_meta, injected_files, [ 702.781433] env[61923]: ERROR nova.compute.manager [instance: c94e8e46-7697-426c-ae2f-aece493fa8f4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 702.781433] env[61923]: ERROR nova.compute.manager [instance: c94e8e46-7697-426c-ae2f-aece493fa8f4] vm_ref = self.build_virtual_machine(instance, [ 702.781433] env[61923]: ERROR nova.compute.manager [instance: c94e8e46-7697-426c-ae2f-aece493fa8f4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 702.781433] env[61923]: ERROR nova.compute.manager [instance: c94e8e46-7697-426c-ae2f-aece493fa8f4] vif_infos = vmwarevif.get_vif_info(self._session, [ 702.781433] env[61923]: ERROR nova.compute.manager [instance: c94e8e46-7697-426c-ae2f-aece493fa8f4] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 702.781927] env[61923]: ERROR nova.compute.manager [instance: c94e8e46-7697-426c-ae2f-aece493fa8f4] for vif in network_info: [ 702.781927] env[61923]: ERROR nova.compute.manager [instance: c94e8e46-7697-426c-ae2f-aece493fa8f4] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 702.781927] env[61923]: ERROR nova.compute.manager [instance: c94e8e46-7697-426c-ae2f-aece493fa8f4] return self._sync_wrapper(fn, *args, **kwargs) [ 702.781927] env[61923]: ERROR nova.compute.manager [instance: c94e8e46-7697-426c-ae2f-aece493fa8f4] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 702.781927] env[61923]: ERROR nova.compute.manager [instance: c94e8e46-7697-426c-ae2f-aece493fa8f4] self.wait() [ 702.781927] env[61923]: ERROR nova.compute.manager [instance: c94e8e46-7697-426c-ae2f-aece493fa8f4] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 702.781927] env[61923]: ERROR nova.compute.manager [instance: c94e8e46-7697-426c-ae2f-aece493fa8f4] self[:] = self._gt.wait() [ 702.781927] env[61923]: ERROR nova.compute.manager [instance: c94e8e46-7697-426c-ae2f-aece493fa8f4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 702.781927] env[61923]: ERROR nova.compute.manager [instance: c94e8e46-7697-426c-ae2f-aece493fa8f4] return self._exit_event.wait() [ 702.781927] env[61923]: ERROR nova.compute.manager [instance: c94e8e46-7697-426c-ae2f-aece493fa8f4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 702.781927] env[61923]: ERROR nova.compute.manager [instance: c94e8e46-7697-426c-ae2f-aece493fa8f4] result = hub.switch() [ 702.781927] env[61923]: ERROR nova.compute.manager [instance: c94e8e46-7697-426c-ae2f-aece493fa8f4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 702.781927] env[61923]: ERROR nova.compute.manager [instance: c94e8e46-7697-426c-ae2f-aece493fa8f4] return self.greenlet.switch() [ 702.782550] env[61923]: ERROR nova.compute.manager [instance: c94e8e46-7697-426c-ae2f-aece493fa8f4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 702.782550] env[61923]: ERROR nova.compute.manager [instance: c94e8e46-7697-426c-ae2f-aece493fa8f4] result = function(*args, **kwargs) [ 702.782550] env[61923]: ERROR nova.compute.manager [instance: c94e8e46-7697-426c-ae2f-aece493fa8f4] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 702.782550] env[61923]: ERROR nova.compute.manager [instance: c94e8e46-7697-426c-ae2f-aece493fa8f4] return func(*args, **kwargs) [ 702.782550] env[61923]: ERROR nova.compute.manager [instance: c94e8e46-7697-426c-ae2f-aece493fa8f4] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 702.782550] env[61923]: ERROR nova.compute.manager [instance: c94e8e46-7697-426c-ae2f-aece493fa8f4] raise e [ 702.782550] env[61923]: ERROR nova.compute.manager [instance: c94e8e46-7697-426c-ae2f-aece493fa8f4] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 702.782550] env[61923]: ERROR nova.compute.manager [instance: c94e8e46-7697-426c-ae2f-aece493fa8f4] nwinfo = self.network_api.allocate_for_instance( [ 702.782550] env[61923]: ERROR nova.compute.manager [instance: c94e8e46-7697-426c-ae2f-aece493fa8f4] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 702.782550] env[61923]: ERROR nova.compute.manager [instance: c94e8e46-7697-426c-ae2f-aece493fa8f4] created_port_ids = self._update_ports_for_instance( [ 702.782550] env[61923]: ERROR nova.compute.manager [instance: c94e8e46-7697-426c-ae2f-aece493fa8f4] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 702.782550] env[61923]: ERROR nova.compute.manager [instance: c94e8e46-7697-426c-ae2f-aece493fa8f4] with excutils.save_and_reraise_exception(): [ 702.782550] env[61923]: ERROR nova.compute.manager [instance: c94e8e46-7697-426c-ae2f-aece493fa8f4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 702.783473] env[61923]: ERROR nova.compute.manager [instance: c94e8e46-7697-426c-ae2f-aece493fa8f4] self.force_reraise() [ 702.783473] env[61923]: ERROR nova.compute.manager [instance: c94e8e46-7697-426c-ae2f-aece493fa8f4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 702.783473] env[61923]: ERROR nova.compute.manager [instance: c94e8e46-7697-426c-ae2f-aece493fa8f4] raise self.value [ 702.783473] env[61923]: ERROR nova.compute.manager [instance: c94e8e46-7697-426c-ae2f-aece493fa8f4] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 702.783473] env[61923]: ERROR nova.compute.manager [instance: c94e8e46-7697-426c-ae2f-aece493fa8f4] updated_port = self._update_port( [ 702.783473] env[61923]: ERROR nova.compute.manager [instance: c94e8e46-7697-426c-ae2f-aece493fa8f4] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 702.783473] env[61923]: ERROR nova.compute.manager [instance: c94e8e46-7697-426c-ae2f-aece493fa8f4] _ensure_no_port_binding_failure(port) [ 702.783473] env[61923]: ERROR nova.compute.manager [instance: c94e8e46-7697-426c-ae2f-aece493fa8f4] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 702.783473] env[61923]: ERROR nova.compute.manager [instance: c94e8e46-7697-426c-ae2f-aece493fa8f4] raise exception.PortBindingFailed(port_id=port['id']) [ 702.783473] env[61923]: ERROR nova.compute.manager [instance: c94e8e46-7697-426c-ae2f-aece493fa8f4] nova.exception.PortBindingFailed: Binding failed for port 70370c2b-9eb9-48b7-8acc-8a7eafc1d41d, please check neutron logs for more information. [ 702.783473] env[61923]: ERROR nova.compute.manager [instance: c94e8e46-7697-426c-ae2f-aece493fa8f4] [ 702.784184] env[61923]: DEBUG nova.compute.utils [None req-1f24f04c-cf4b-4940-9447-eacc11fa5e66 tempest-ServersTestFqdnHostnames-345291565 tempest-ServersTestFqdnHostnames-345291565-project-member] [instance: c94e8e46-7697-426c-ae2f-aece493fa8f4] Binding failed for port 70370c2b-9eb9-48b7-8acc-8a7eafc1d41d, please check neutron logs for more information. {{(pid=61923) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 702.786519] env[61923]: DEBUG nova.compute.manager [None req-1f24f04c-cf4b-4940-9447-eacc11fa5e66 tempest-ServersTestFqdnHostnames-345291565 tempest-ServersTestFqdnHostnames-345291565-project-member] [instance: c94e8e46-7697-426c-ae2f-aece493fa8f4] Build of instance c94e8e46-7697-426c-ae2f-aece493fa8f4 was re-scheduled: Binding failed for port 70370c2b-9eb9-48b7-8acc-8a7eafc1d41d, please check neutron logs for more information. {{(pid=61923) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 702.786939] env[61923]: DEBUG nova.compute.manager [None req-1f24f04c-cf4b-4940-9447-eacc11fa5e66 tempest-ServersTestFqdnHostnames-345291565 tempest-ServersTestFqdnHostnames-345291565-project-member] [instance: c94e8e46-7697-426c-ae2f-aece493fa8f4] Unplugging VIFs for instance {{(pid=61923) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 702.787180] env[61923]: DEBUG oslo_concurrency.lockutils [None req-1f24f04c-cf4b-4940-9447-eacc11fa5e66 tempest-ServersTestFqdnHostnames-345291565 tempest-ServersTestFqdnHostnames-345291565-project-member] Acquiring lock "refresh_cache-c94e8e46-7697-426c-ae2f-aece493fa8f4" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 702.787324] env[61923]: DEBUG oslo_concurrency.lockutils [None req-1f24f04c-cf4b-4940-9447-eacc11fa5e66 tempest-ServersTestFqdnHostnames-345291565 tempest-ServersTestFqdnHostnames-345291565-project-member] Acquired lock "refresh_cache-c94e8e46-7697-426c-ae2f-aece493fa8f4" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 702.787478] env[61923]: DEBUG nova.network.neutron [None req-1f24f04c-cf4b-4940-9447-eacc11fa5e66 tempest-ServersTestFqdnHostnames-345291565 tempest-ServersTestFqdnHostnames-345291565-project-member] [instance: c94e8e46-7697-426c-ae2f-aece493fa8f4] Building network info cache for instance {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 702.788580] env[61923]: DEBUG oslo_concurrency.lockutils [None req-24801cc4-3974-42da-978f-a22cb8f70af0 tempest-SecurityGroupsTestJSON-509490211 tempest-SecurityGroupsTestJSON-509490211-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.817s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 702.793018] env[61923]: INFO nova.compute.claims [None req-24801cc4-3974-42da-978f-a22cb8f70af0 tempest-SecurityGroupsTestJSON-509490211 tempest-SecurityGroupsTestJSON-509490211-project-member] [instance: 2939d1e5-07fe-48b3-bc87-5f67fa1a7ae4] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 702.892528] env[61923]: DEBUG nova.compute.manager [None req-c09a27e9-8e33-411f-b17b-7cd8d66a175c tempest-ServerMetadataNegativeTestJSON-1976245157 tempest-ServerMetadataNegativeTestJSON-1976245157-project-member] [instance: 25b62db6-6e64-4e38-bfd3-82a2e9a4883a] Start spawning the instance on the hypervisor. {{(pid=61923) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 702.914160] env[61923]: DEBUG nova.network.neutron [None req-44a76586-0f76-448f-af41-a5d846bfda81 tempest-ServerActionsV293TestJSON-849444864 tempest-ServerActionsV293TestJSON-849444864-project-member] [instance: 62144317-0354-4751-8f2b-4eae588c9da7] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 702.923201] env[61923]: DEBUG nova.virt.hardware [None req-c09a27e9-8e33-411f-b17b-7cd8d66a175c tempest-ServerMetadataNegativeTestJSON-1976245157 tempest-ServerMetadataNegativeTestJSON-1976245157-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-29T20:07:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-29T20:07:35Z,direct_url=,disk_format='vmdk',id=0f153f63-ae0a-45b1-b7f5-7f9b673c947f,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='4e7b5e3b3c3443c8bd5c70f8a15739f5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-29T20:07:36Z,virtual_size=,visibility=), allow threads: False {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 702.923201] env[61923]: DEBUG nova.virt.hardware [None req-c09a27e9-8e33-411f-b17b-7cd8d66a175c tempest-ServerMetadataNegativeTestJSON-1976245157 tempest-ServerMetadataNegativeTestJSON-1976245157-project-member] Flavor limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 702.923388] env[61923]: DEBUG nova.virt.hardware [None req-c09a27e9-8e33-411f-b17b-7cd8d66a175c tempest-ServerMetadataNegativeTestJSON-1976245157 tempest-ServerMetadataNegativeTestJSON-1976245157-project-member] Image limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 702.923464] env[61923]: DEBUG nova.virt.hardware [None req-c09a27e9-8e33-411f-b17b-7cd8d66a175c tempest-ServerMetadataNegativeTestJSON-1976245157 tempest-ServerMetadataNegativeTestJSON-1976245157-project-member] Flavor pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 702.923598] env[61923]: DEBUG nova.virt.hardware [None req-c09a27e9-8e33-411f-b17b-7cd8d66a175c tempest-ServerMetadataNegativeTestJSON-1976245157 tempest-ServerMetadataNegativeTestJSON-1976245157-project-member] Image pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 702.923738] env[61923]: DEBUG nova.virt.hardware [None req-c09a27e9-8e33-411f-b17b-7cd8d66a175c tempest-ServerMetadataNegativeTestJSON-1976245157 tempest-ServerMetadataNegativeTestJSON-1976245157-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 702.923937] env[61923]: DEBUG nova.virt.hardware [None req-c09a27e9-8e33-411f-b17b-7cd8d66a175c tempest-ServerMetadataNegativeTestJSON-1976245157 tempest-ServerMetadataNegativeTestJSON-1976245157-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 702.924106] env[61923]: DEBUG nova.virt.hardware [None req-c09a27e9-8e33-411f-b17b-7cd8d66a175c tempest-ServerMetadataNegativeTestJSON-1976245157 tempest-ServerMetadataNegativeTestJSON-1976245157-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 702.924297] env[61923]: DEBUG nova.virt.hardware [None req-c09a27e9-8e33-411f-b17b-7cd8d66a175c tempest-ServerMetadataNegativeTestJSON-1976245157 tempest-ServerMetadataNegativeTestJSON-1976245157-project-member] Got 1 possible topologies {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 702.924418] env[61923]: DEBUG nova.virt.hardware [None req-c09a27e9-8e33-411f-b17b-7cd8d66a175c tempest-ServerMetadataNegativeTestJSON-1976245157 tempest-ServerMetadataNegativeTestJSON-1976245157-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 702.924584] env[61923]: DEBUG nova.virt.hardware [None req-c09a27e9-8e33-411f-b17b-7cd8d66a175c tempest-ServerMetadataNegativeTestJSON-1976245157 tempest-ServerMetadataNegativeTestJSON-1976245157-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 702.925469] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9c72277-ae25-4a2d-b93c-96833edb6c52 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 702.935109] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ba0ea70-7eba-4e47-a9e5-83ea1cf95f70 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.055550] env[61923]: ERROR nova.compute.manager [None req-c09a27e9-8e33-411f-b17b-7cd8d66a175c tempest-ServerMetadataNegativeTestJSON-1976245157 tempest-ServerMetadataNegativeTestJSON-1976245157-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port bf236ea4-19e2-487c-ad15-61be750ff9c5, please check neutron logs for more information. [ 703.055550] env[61923]: ERROR nova.compute.manager Traceback (most recent call last): [ 703.055550] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 703.055550] env[61923]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 703.055550] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 703.055550] env[61923]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 703.055550] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 703.055550] env[61923]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 703.055550] env[61923]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 703.055550] env[61923]: ERROR nova.compute.manager self.force_reraise() [ 703.055550] env[61923]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 703.055550] env[61923]: ERROR nova.compute.manager raise self.value [ 703.055550] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 703.055550] env[61923]: ERROR nova.compute.manager updated_port = self._update_port( [ 703.055550] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 703.055550] env[61923]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 703.056133] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 703.056133] env[61923]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 703.056133] env[61923]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port bf236ea4-19e2-487c-ad15-61be750ff9c5, please check neutron logs for more information. [ 703.056133] env[61923]: ERROR nova.compute.manager [ 703.056133] env[61923]: Traceback (most recent call last): [ 703.056133] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 703.056133] env[61923]: listener.cb(fileno) [ 703.056133] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 703.056133] env[61923]: result = function(*args, **kwargs) [ 703.056133] env[61923]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 703.056133] env[61923]: return func(*args, **kwargs) [ 703.056133] env[61923]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 703.056133] env[61923]: raise e [ 703.056133] env[61923]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 703.056133] env[61923]: nwinfo = self.network_api.allocate_for_instance( [ 703.056133] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 703.056133] env[61923]: created_port_ids = self._update_ports_for_instance( [ 703.056133] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 703.056133] env[61923]: with excutils.save_and_reraise_exception(): [ 703.056133] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 703.056133] env[61923]: self.force_reraise() [ 703.056133] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 703.056133] env[61923]: raise self.value [ 703.056133] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 703.056133] env[61923]: updated_port = self._update_port( [ 703.056133] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 703.056133] env[61923]: _ensure_no_port_binding_failure(port) [ 703.056133] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 703.056133] env[61923]: raise exception.PortBindingFailed(port_id=port['id']) [ 703.056977] env[61923]: nova.exception.PortBindingFailed: Binding failed for port bf236ea4-19e2-487c-ad15-61be750ff9c5, please check neutron logs for more information. [ 703.056977] env[61923]: Removing descriptor: 14 [ 703.056977] env[61923]: ERROR nova.compute.manager [None req-c09a27e9-8e33-411f-b17b-7cd8d66a175c tempest-ServerMetadataNegativeTestJSON-1976245157 tempest-ServerMetadataNegativeTestJSON-1976245157-project-member] [instance: 25b62db6-6e64-4e38-bfd3-82a2e9a4883a] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port bf236ea4-19e2-487c-ad15-61be750ff9c5, please check neutron logs for more information. [ 703.056977] env[61923]: ERROR nova.compute.manager [instance: 25b62db6-6e64-4e38-bfd3-82a2e9a4883a] Traceback (most recent call last): [ 703.056977] env[61923]: ERROR nova.compute.manager [instance: 25b62db6-6e64-4e38-bfd3-82a2e9a4883a] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 703.056977] env[61923]: ERROR nova.compute.manager [instance: 25b62db6-6e64-4e38-bfd3-82a2e9a4883a] yield resources [ 703.056977] env[61923]: ERROR nova.compute.manager [instance: 25b62db6-6e64-4e38-bfd3-82a2e9a4883a] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 703.056977] env[61923]: ERROR nova.compute.manager [instance: 25b62db6-6e64-4e38-bfd3-82a2e9a4883a] self.driver.spawn(context, instance, image_meta, [ 703.056977] env[61923]: ERROR nova.compute.manager [instance: 25b62db6-6e64-4e38-bfd3-82a2e9a4883a] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 703.056977] env[61923]: ERROR nova.compute.manager [instance: 25b62db6-6e64-4e38-bfd3-82a2e9a4883a] self._vmops.spawn(context, instance, image_meta, injected_files, [ 703.056977] env[61923]: ERROR nova.compute.manager [instance: 25b62db6-6e64-4e38-bfd3-82a2e9a4883a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 703.056977] env[61923]: ERROR nova.compute.manager [instance: 25b62db6-6e64-4e38-bfd3-82a2e9a4883a] vm_ref = self.build_virtual_machine(instance, [ 703.057344] env[61923]: ERROR nova.compute.manager [instance: 25b62db6-6e64-4e38-bfd3-82a2e9a4883a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 703.057344] env[61923]: ERROR nova.compute.manager [instance: 25b62db6-6e64-4e38-bfd3-82a2e9a4883a] vif_infos = vmwarevif.get_vif_info(self._session, [ 703.057344] env[61923]: ERROR nova.compute.manager [instance: 25b62db6-6e64-4e38-bfd3-82a2e9a4883a] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 703.057344] env[61923]: ERROR nova.compute.manager [instance: 25b62db6-6e64-4e38-bfd3-82a2e9a4883a] for vif in network_info: [ 703.057344] env[61923]: ERROR nova.compute.manager [instance: 25b62db6-6e64-4e38-bfd3-82a2e9a4883a] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 703.057344] env[61923]: ERROR nova.compute.manager [instance: 25b62db6-6e64-4e38-bfd3-82a2e9a4883a] return self._sync_wrapper(fn, *args, **kwargs) [ 703.057344] env[61923]: ERROR nova.compute.manager [instance: 25b62db6-6e64-4e38-bfd3-82a2e9a4883a] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 703.057344] env[61923]: ERROR nova.compute.manager [instance: 25b62db6-6e64-4e38-bfd3-82a2e9a4883a] self.wait() [ 703.057344] env[61923]: ERROR nova.compute.manager [instance: 25b62db6-6e64-4e38-bfd3-82a2e9a4883a] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 703.057344] env[61923]: ERROR nova.compute.manager [instance: 25b62db6-6e64-4e38-bfd3-82a2e9a4883a] self[:] = self._gt.wait() [ 703.057344] env[61923]: ERROR nova.compute.manager [instance: 25b62db6-6e64-4e38-bfd3-82a2e9a4883a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 703.057344] env[61923]: ERROR nova.compute.manager [instance: 25b62db6-6e64-4e38-bfd3-82a2e9a4883a] return self._exit_event.wait() [ 703.057344] env[61923]: ERROR nova.compute.manager [instance: 25b62db6-6e64-4e38-bfd3-82a2e9a4883a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 703.057700] env[61923]: ERROR nova.compute.manager [instance: 25b62db6-6e64-4e38-bfd3-82a2e9a4883a] result = hub.switch() [ 703.057700] env[61923]: ERROR nova.compute.manager [instance: 25b62db6-6e64-4e38-bfd3-82a2e9a4883a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 703.057700] env[61923]: ERROR nova.compute.manager [instance: 25b62db6-6e64-4e38-bfd3-82a2e9a4883a] return self.greenlet.switch() [ 703.057700] env[61923]: ERROR nova.compute.manager [instance: 25b62db6-6e64-4e38-bfd3-82a2e9a4883a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 703.057700] env[61923]: ERROR nova.compute.manager [instance: 25b62db6-6e64-4e38-bfd3-82a2e9a4883a] result = function(*args, **kwargs) [ 703.057700] env[61923]: ERROR nova.compute.manager [instance: 25b62db6-6e64-4e38-bfd3-82a2e9a4883a] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 703.057700] env[61923]: ERROR nova.compute.manager [instance: 25b62db6-6e64-4e38-bfd3-82a2e9a4883a] return func(*args, **kwargs) [ 703.057700] env[61923]: ERROR nova.compute.manager [instance: 25b62db6-6e64-4e38-bfd3-82a2e9a4883a] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 703.057700] env[61923]: ERROR nova.compute.manager [instance: 25b62db6-6e64-4e38-bfd3-82a2e9a4883a] raise e [ 703.057700] env[61923]: ERROR nova.compute.manager [instance: 25b62db6-6e64-4e38-bfd3-82a2e9a4883a] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 703.057700] env[61923]: ERROR nova.compute.manager [instance: 25b62db6-6e64-4e38-bfd3-82a2e9a4883a] nwinfo = self.network_api.allocate_for_instance( [ 703.057700] env[61923]: ERROR nova.compute.manager [instance: 25b62db6-6e64-4e38-bfd3-82a2e9a4883a] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 703.057700] env[61923]: ERROR nova.compute.manager [instance: 25b62db6-6e64-4e38-bfd3-82a2e9a4883a] created_port_ids = self._update_ports_for_instance( [ 703.058089] env[61923]: ERROR nova.compute.manager [instance: 25b62db6-6e64-4e38-bfd3-82a2e9a4883a] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 703.058089] env[61923]: ERROR nova.compute.manager [instance: 25b62db6-6e64-4e38-bfd3-82a2e9a4883a] with excutils.save_and_reraise_exception(): [ 703.058089] env[61923]: ERROR nova.compute.manager [instance: 25b62db6-6e64-4e38-bfd3-82a2e9a4883a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 703.058089] env[61923]: ERROR nova.compute.manager [instance: 25b62db6-6e64-4e38-bfd3-82a2e9a4883a] self.force_reraise() [ 703.058089] env[61923]: ERROR nova.compute.manager [instance: 25b62db6-6e64-4e38-bfd3-82a2e9a4883a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 703.058089] env[61923]: ERROR nova.compute.manager [instance: 25b62db6-6e64-4e38-bfd3-82a2e9a4883a] raise self.value [ 703.058089] env[61923]: ERROR nova.compute.manager [instance: 25b62db6-6e64-4e38-bfd3-82a2e9a4883a] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 703.058089] env[61923]: ERROR nova.compute.manager [instance: 25b62db6-6e64-4e38-bfd3-82a2e9a4883a] updated_port = self._update_port( [ 703.058089] env[61923]: ERROR nova.compute.manager [instance: 25b62db6-6e64-4e38-bfd3-82a2e9a4883a] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 703.058089] env[61923]: ERROR nova.compute.manager [instance: 25b62db6-6e64-4e38-bfd3-82a2e9a4883a] _ensure_no_port_binding_failure(port) [ 703.058089] env[61923]: ERROR nova.compute.manager [instance: 25b62db6-6e64-4e38-bfd3-82a2e9a4883a] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 703.058089] env[61923]: ERROR nova.compute.manager [instance: 25b62db6-6e64-4e38-bfd3-82a2e9a4883a] raise exception.PortBindingFailed(port_id=port['id']) [ 703.058419] env[61923]: ERROR nova.compute.manager [instance: 25b62db6-6e64-4e38-bfd3-82a2e9a4883a] nova.exception.PortBindingFailed: Binding failed for port bf236ea4-19e2-487c-ad15-61be750ff9c5, please check neutron logs for more information. [ 703.058419] env[61923]: ERROR nova.compute.manager [instance: 25b62db6-6e64-4e38-bfd3-82a2e9a4883a] [ 703.058419] env[61923]: INFO nova.compute.manager [None req-c09a27e9-8e33-411f-b17b-7cd8d66a175c tempest-ServerMetadataNegativeTestJSON-1976245157 tempest-ServerMetadataNegativeTestJSON-1976245157-project-member] [instance: 25b62db6-6e64-4e38-bfd3-82a2e9a4883a] Terminating instance [ 703.059142] env[61923]: DEBUG oslo_concurrency.lockutils [None req-c09a27e9-8e33-411f-b17b-7cd8d66a175c tempest-ServerMetadataNegativeTestJSON-1976245157 tempest-ServerMetadataNegativeTestJSON-1976245157-project-member] Acquiring lock "refresh_cache-25b62db6-6e64-4e38-bfd3-82a2e9a4883a" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 703.059311] env[61923]: DEBUG oslo_concurrency.lockutils [None req-c09a27e9-8e33-411f-b17b-7cd8d66a175c tempest-ServerMetadataNegativeTestJSON-1976245157 tempest-ServerMetadataNegativeTestJSON-1976245157-project-member] Acquired lock "refresh_cache-25b62db6-6e64-4e38-bfd3-82a2e9a4883a" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 703.059473] env[61923]: DEBUG nova.network.neutron [None req-c09a27e9-8e33-411f-b17b-7cd8d66a175c tempest-ServerMetadataNegativeTestJSON-1976245157 tempest-ServerMetadataNegativeTestJSON-1976245157-project-member] [instance: 25b62db6-6e64-4e38-bfd3-82a2e9a4883a] Building network info cache for instance {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 703.320523] env[61923]: DEBUG nova.network.neutron [None req-1f24f04c-cf4b-4940-9447-eacc11fa5e66 tempest-ServersTestFqdnHostnames-345291565 tempest-ServersTestFqdnHostnames-345291565-project-member] [instance: c94e8e46-7697-426c-ae2f-aece493fa8f4] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 703.400588] env[61923]: DEBUG nova.network.neutron [None req-1f24f04c-cf4b-4940-9447-eacc11fa5e66 tempest-ServersTestFqdnHostnames-345291565 tempest-ServersTestFqdnHostnames-345291565-project-member] [instance: c94e8e46-7697-426c-ae2f-aece493fa8f4] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 703.416558] env[61923]: DEBUG oslo_concurrency.lockutils [None req-44a76586-0f76-448f-af41-a5d846bfda81 tempest-ServerActionsV293TestJSON-849444864 tempest-ServerActionsV293TestJSON-849444864-project-member] Releasing lock "refresh_cache-62144317-0354-4751-8f2b-4eae588c9da7" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 703.417123] env[61923]: DEBUG nova.compute.manager [None req-44a76586-0f76-448f-af41-a5d846bfda81 tempest-ServerActionsV293TestJSON-849444864 tempest-ServerActionsV293TestJSON-849444864-project-member] [instance: 62144317-0354-4751-8f2b-4eae588c9da7] Start destroying the instance on the hypervisor. {{(pid=61923) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 703.417452] env[61923]: DEBUG oslo_concurrency.lockutils [req-b9f2787e-ad0b-4abf-9b7b-2d13ff5feb43 req-1012b622-79a5-4dad-9068-333911395cfe service nova] Acquired lock "refresh_cache-62144317-0354-4751-8f2b-4eae588c9da7" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 703.417801] env[61923]: DEBUG nova.network.neutron [req-b9f2787e-ad0b-4abf-9b7b-2d13ff5feb43 req-1012b622-79a5-4dad-9068-333911395cfe service nova] [instance: 62144317-0354-4751-8f2b-4eae588c9da7] Refreshing network info cache for port 6059f47a-1eef-4b32-9c0c-25cdd6ab2609 {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 703.418980] env[61923]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ab3b28e0-792b-4dc0-bd7f-2c1623d47f85 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.429255] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55a6feb9-bee7-49fc-b180-b022402f1b5f {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.453873] env[61923]: WARNING nova.virt.vmwareapi.driver [None req-44a76586-0f76-448f-af41-a5d846bfda81 tempest-ServerActionsV293TestJSON-849444864 tempest-ServerActionsV293TestJSON-849444864-project-member] [instance: 62144317-0354-4751-8f2b-4eae588c9da7] Instance does not exists. Proceeding to delete instance properties on datastore: nova.exception.InstanceNotFound: Instance 62144317-0354-4751-8f2b-4eae588c9da7 could not be found. [ 703.454118] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-44a76586-0f76-448f-af41-a5d846bfda81 tempest-ServerActionsV293TestJSON-849444864 tempest-ServerActionsV293TestJSON-849444864-project-member] [instance: 62144317-0354-4751-8f2b-4eae588c9da7] Destroying instance {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 703.454405] env[61923]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-1c2633c7-d89a-48c1-abbd-753ae5d5ae7f {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.463021] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad52f2f1-cab2-4142-bae2-b47b269a3223 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 703.484486] env[61923]: WARNING nova.virt.vmwareapi.vmops [None req-44a76586-0f76-448f-af41-a5d846bfda81 tempest-ServerActionsV293TestJSON-849444864 tempest-ServerActionsV293TestJSON-849444864-project-member] [instance: 62144317-0354-4751-8f2b-4eae588c9da7] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 62144317-0354-4751-8f2b-4eae588c9da7 could not be found. [ 703.484760] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-44a76586-0f76-448f-af41-a5d846bfda81 tempest-ServerActionsV293TestJSON-849444864 tempest-ServerActionsV293TestJSON-849444864-project-member] [instance: 62144317-0354-4751-8f2b-4eae588c9da7] Instance destroyed {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 703.484982] env[61923]: INFO nova.compute.manager [None req-44a76586-0f76-448f-af41-a5d846bfda81 tempest-ServerActionsV293TestJSON-849444864 tempest-ServerActionsV293TestJSON-849444864-project-member] [instance: 62144317-0354-4751-8f2b-4eae588c9da7] Took 0.07 seconds to destroy the instance on the hypervisor. [ 703.485243] env[61923]: DEBUG oslo.service.loopingcall [None req-44a76586-0f76-448f-af41-a5d846bfda81 tempest-ServerActionsV293TestJSON-849444864 tempest-ServerActionsV293TestJSON-849444864-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61923) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 703.485463] env[61923]: DEBUG nova.compute.manager [-] [instance: 62144317-0354-4751-8f2b-4eae588c9da7] Deallocating network for instance {{(pid=61923) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 703.485558] env[61923]: DEBUG nova.network.neutron [-] [instance: 62144317-0354-4751-8f2b-4eae588c9da7] deallocate_for_instance() {{(pid=61923) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 703.501969] env[61923]: DEBUG nova.network.neutron [-] [instance: 62144317-0354-4751-8f2b-4eae588c9da7] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 703.575806] env[61923]: DEBUG nova.network.neutron [None req-c09a27e9-8e33-411f-b17b-7cd8d66a175c tempest-ServerMetadataNegativeTestJSON-1976245157 tempest-ServerMetadataNegativeTestJSON-1976245157-project-member] [instance: 25b62db6-6e64-4e38-bfd3-82a2e9a4883a] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 703.688355] env[61923]: DEBUG nova.network.neutron [None req-c09a27e9-8e33-411f-b17b-7cd8d66a175c tempest-ServerMetadataNegativeTestJSON-1976245157 tempest-ServerMetadataNegativeTestJSON-1976245157-project-member] [instance: 25b62db6-6e64-4e38-bfd3-82a2e9a4883a] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 703.904074] env[61923]: DEBUG oslo_concurrency.lockutils [None req-1f24f04c-cf4b-4940-9447-eacc11fa5e66 tempest-ServersTestFqdnHostnames-345291565 tempest-ServersTestFqdnHostnames-345291565-project-member] Releasing lock "refresh_cache-c94e8e46-7697-426c-ae2f-aece493fa8f4" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 703.904074] env[61923]: DEBUG nova.compute.manager [None req-1f24f04c-cf4b-4940-9447-eacc11fa5e66 tempest-ServersTestFqdnHostnames-345291565 tempest-ServersTestFqdnHostnames-345291565-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61923) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 703.904074] env[61923]: DEBUG nova.compute.manager [None req-1f24f04c-cf4b-4940-9447-eacc11fa5e66 tempest-ServersTestFqdnHostnames-345291565 tempest-ServersTestFqdnHostnames-345291565-project-member] [instance: c94e8e46-7697-426c-ae2f-aece493fa8f4] Deallocating network for instance {{(pid=61923) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 703.904074] env[61923]: DEBUG nova.network.neutron [None req-1f24f04c-cf4b-4940-9447-eacc11fa5e66 tempest-ServersTestFqdnHostnames-345291565 tempest-ServersTestFqdnHostnames-345291565-project-member] [instance: c94e8e46-7697-426c-ae2f-aece493fa8f4] deallocate_for_instance() {{(pid=61923) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 703.927558] env[61923]: DEBUG nova.network.neutron [None req-1f24f04c-cf4b-4940-9447-eacc11fa5e66 tempest-ServersTestFqdnHostnames-345291565 tempest-ServersTestFqdnHostnames-345291565-project-member] [instance: c94e8e46-7697-426c-ae2f-aece493fa8f4] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 703.947987] env[61923]: DEBUG nova.network.neutron [req-b9f2787e-ad0b-4abf-9b7b-2d13ff5feb43 req-1012b622-79a5-4dad-9068-333911395cfe service nova] [instance: 62144317-0354-4751-8f2b-4eae588c9da7] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 704.006021] env[61923]: DEBUG nova.network.neutron [-] [instance: 62144317-0354-4751-8f2b-4eae588c9da7] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 704.073560] env[61923]: DEBUG nova.network.neutron [req-b9f2787e-ad0b-4abf-9b7b-2d13ff5feb43 req-1012b622-79a5-4dad-9068-333911395cfe service nova] [instance: 62144317-0354-4751-8f2b-4eae588c9da7] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 704.147797] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76ddc005-1e6a-49f4-a44f-941d62e80a08 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.155805] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23cc50dd-860e-410a-a21f-84f71a73d241 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.186137] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8bd56059-34c2-41b5-a65c-a6f7da1e8ca6 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.192297] env[61923]: DEBUG oslo_concurrency.lockutils [None req-c09a27e9-8e33-411f-b17b-7cd8d66a175c tempest-ServerMetadataNegativeTestJSON-1976245157 tempest-ServerMetadataNegativeTestJSON-1976245157-project-member] Releasing lock "refresh_cache-25b62db6-6e64-4e38-bfd3-82a2e9a4883a" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 704.192670] env[61923]: DEBUG nova.compute.manager [None req-c09a27e9-8e33-411f-b17b-7cd8d66a175c tempest-ServerMetadataNegativeTestJSON-1976245157 tempest-ServerMetadataNegativeTestJSON-1976245157-project-member] [instance: 25b62db6-6e64-4e38-bfd3-82a2e9a4883a] Start destroying the instance on the hypervisor. {{(pid=61923) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 704.192852] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-c09a27e9-8e33-411f-b17b-7cd8d66a175c tempest-ServerMetadataNegativeTestJSON-1976245157 tempest-ServerMetadataNegativeTestJSON-1976245157-project-member] [instance: 25b62db6-6e64-4e38-bfd3-82a2e9a4883a] Destroying instance {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 704.193130] env[61923]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-25b855ef-7521-4a61-8ff4-4614b6ec7304 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.195721] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef4425d3-2d00-4acf-9333-0e4b9e5b7c7c {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.208937] env[61923]: DEBUG nova.compute.provider_tree [None req-24801cc4-3974-42da-978f-a22cb8f70af0 tempest-SecurityGroupsTestJSON-509490211 tempest-SecurityGroupsTestJSON-509490211-project-member] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 704.213640] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bbecc3fe-f15e-4911-8e87-ca4ca64791fd {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.235265] env[61923]: WARNING nova.virt.vmwareapi.vmops [None req-c09a27e9-8e33-411f-b17b-7cd8d66a175c tempest-ServerMetadataNegativeTestJSON-1976245157 tempest-ServerMetadataNegativeTestJSON-1976245157-project-member] [instance: 25b62db6-6e64-4e38-bfd3-82a2e9a4883a] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 25b62db6-6e64-4e38-bfd3-82a2e9a4883a could not be found. [ 704.235477] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-c09a27e9-8e33-411f-b17b-7cd8d66a175c tempest-ServerMetadataNegativeTestJSON-1976245157 tempest-ServerMetadataNegativeTestJSON-1976245157-project-member] [instance: 25b62db6-6e64-4e38-bfd3-82a2e9a4883a] Instance destroyed {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 704.235657] env[61923]: INFO nova.compute.manager [None req-c09a27e9-8e33-411f-b17b-7cd8d66a175c tempest-ServerMetadataNegativeTestJSON-1976245157 tempest-ServerMetadataNegativeTestJSON-1976245157-project-member] [instance: 25b62db6-6e64-4e38-bfd3-82a2e9a4883a] Took 0.04 seconds to destroy the instance on the hypervisor. [ 704.235899] env[61923]: DEBUG oslo.service.loopingcall [None req-c09a27e9-8e33-411f-b17b-7cd8d66a175c tempest-ServerMetadataNegativeTestJSON-1976245157 tempest-ServerMetadataNegativeTestJSON-1976245157-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61923) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 704.236921] env[61923]: DEBUG nova.compute.manager [-] [instance: 25b62db6-6e64-4e38-bfd3-82a2e9a4883a] Deallocating network for instance {{(pid=61923) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 704.237030] env[61923]: DEBUG nova.network.neutron [-] [instance: 25b62db6-6e64-4e38-bfd3-82a2e9a4883a] deallocate_for_instance() {{(pid=61923) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 704.239501] env[61923]: DEBUG nova.compute.manager [req-2c2fe3c4-9243-4164-833c-2bd3b3b839d7 req-9503d4ea-5ff3-4131-a184-c3bbb79acf38 service nova] [instance: 25b62db6-6e64-4e38-bfd3-82a2e9a4883a] Received event network-changed-bf236ea4-19e2-487c-ad15-61be750ff9c5 {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 704.239713] env[61923]: DEBUG nova.compute.manager [req-2c2fe3c4-9243-4164-833c-2bd3b3b839d7 req-9503d4ea-5ff3-4131-a184-c3bbb79acf38 service nova] [instance: 25b62db6-6e64-4e38-bfd3-82a2e9a4883a] Refreshing instance network info cache due to event network-changed-bf236ea4-19e2-487c-ad15-61be750ff9c5. {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 704.239916] env[61923]: DEBUG oslo_concurrency.lockutils [req-2c2fe3c4-9243-4164-833c-2bd3b3b839d7 req-9503d4ea-5ff3-4131-a184-c3bbb79acf38 service nova] Acquiring lock "refresh_cache-25b62db6-6e64-4e38-bfd3-82a2e9a4883a" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 704.240179] env[61923]: DEBUG oslo_concurrency.lockutils [req-2c2fe3c4-9243-4164-833c-2bd3b3b839d7 req-9503d4ea-5ff3-4131-a184-c3bbb79acf38 service nova] Acquired lock "refresh_cache-25b62db6-6e64-4e38-bfd3-82a2e9a4883a" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 704.240238] env[61923]: DEBUG nova.network.neutron [req-2c2fe3c4-9243-4164-833c-2bd3b3b839d7 req-9503d4ea-5ff3-4131-a184-c3bbb79acf38 service nova] [instance: 25b62db6-6e64-4e38-bfd3-82a2e9a4883a] Refreshing network info cache for port bf236ea4-19e2-487c-ad15-61be750ff9c5 {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 704.254272] env[61923]: DEBUG nova.network.neutron [-] [instance: 25b62db6-6e64-4e38-bfd3-82a2e9a4883a] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 704.430021] env[61923]: DEBUG nova.network.neutron [None req-1f24f04c-cf4b-4940-9447-eacc11fa5e66 tempest-ServersTestFqdnHostnames-345291565 tempest-ServersTestFqdnHostnames-345291565-project-member] [instance: c94e8e46-7697-426c-ae2f-aece493fa8f4] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 704.506206] env[61923]: INFO nova.compute.manager [-] [instance: 62144317-0354-4751-8f2b-4eae588c9da7] Took 1.02 seconds to deallocate network for instance. [ 704.577964] env[61923]: DEBUG oslo_concurrency.lockutils [req-b9f2787e-ad0b-4abf-9b7b-2d13ff5feb43 req-1012b622-79a5-4dad-9068-333911395cfe service nova] Releasing lock "refresh_cache-62144317-0354-4751-8f2b-4eae588c9da7" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 704.579194] env[61923]: DEBUG nova.compute.manager [req-b9f2787e-ad0b-4abf-9b7b-2d13ff5feb43 req-1012b622-79a5-4dad-9068-333911395cfe service nova] [instance: 62144317-0354-4751-8f2b-4eae588c9da7] Received event network-vif-deleted-6059f47a-1eef-4b32-9c0c-25cdd6ab2609 {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 704.712809] env[61923]: DEBUG nova.scheduler.client.report [None req-24801cc4-3974-42da-978f-a22cb8f70af0 tempest-SecurityGroupsTestJSON-509490211 tempest-SecurityGroupsTestJSON-509490211-project-member] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 704.756095] env[61923]: DEBUG nova.network.neutron [req-2c2fe3c4-9243-4164-833c-2bd3b3b839d7 req-9503d4ea-5ff3-4131-a184-c3bbb79acf38 service nova] [instance: 25b62db6-6e64-4e38-bfd3-82a2e9a4883a] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 704.757609] env[61923]: DEBUG nova.network.neutron [-] [instance: 25b62db6-6e64-4e38-bfd3-82a2e9a4883a] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 704.824675] env[61923]: DEBUG nova.network.neutron [req-2c2fe3c4-9243-4164-833c-2bd3b3b839d7 req-9503d4ea-5ff3-4131-a184-c3bbb79acf38 service nova] [instance: 25b62db6-6e64-4e38-bfd3-82a2e9a4883a] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 704.932502] env[61923]: INFO nova.compute.manager [None req-1f24f04c-cf4b-4940-9447-eacc11fa5e66 tempest-ServersTestFqdnHostnames-345291565 tempest-ServersTestFqdnHostnames-345291565-project-member] [instance: c94e8e46-7697-426c-ae2f-aece493fa8f4] Took 1.03 seconds to deallocate network for instance. [ 705.065097] env[61923]: INFO nova.compute.manager [None req-44a76586-0f76-448f-af41-a5d846bfda81 tempest-ServerActionsV293TestJSON-849444864 tempest-ServerActionsV293TestJSON-849444864-project-member] [instance: 62144317-0354-4751-8f2b-4eae588c9da7] Took 0.56 seconds to detach 1 volumes for instance. [ 705.067544] env[61923]: DEBUG nova.compute.claims [None req-44a76586-0f76-448f-af41-a5d846bfda81 tempest-ServerActionsV293TestJSON-849444864 tempest-ServerActionsV293TestJSON-849444864-project-member] [instance: 62144317-0354-4751-8f2b-4eae588c9da7] Aborting claim: {{(pid=61923) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 705.067544] env[61923]: DEBUG oslo_concurrency.lockutils [None req-44a76586-0f76-448f-af41-a5d846bfda81 tempest-ServerActionsV293TestJSON-849444864 tempest-ServerActionsV293TestJSON-849444864-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 705.217218] env[61923]: DEBUG oslo_concurrency.lockutils [None req-24801cc4-3974-42da-978f-a22cb8f70af0 tempest-SecurityGroupsTestJSON-509490211 tempest-SecurityGroupsTestJSON-509490211-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.429s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 705.217783] env[61923]: DEBUG nova.compute.manager [None req-24801cc4-3974-42da-978f-a22cb8f70af0 tempest-SecurityGroupsTestJSON-509490211 tempest-SecurityGroupsTestJSON-509490211-project-member] [instance: 2939d1e5-07fe-48b3-bc87-5f67fa1a7ae4] Start building networks asynchronously for instance. {{(pid=61923) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 705.222041] env[61923]: DEBUG oslo_concurrency.lockutils [None req-aadda1bb-f770-4742-8c34-cd25e3f7e7cb tempest-InstanceActionsNegativeTestJSON-1700332362 tempest-InstanceActionsNegativeTestJSON-1700332362-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.527s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 705.259990] env[61923]: INFO nova.compute.manager [-] [instance: 25b62db6-6e64-4e38-bfd3-82a2e9a4883a] Took 1.02 seconds to deallocate network for instance. [ 705.262199] env[61923]: DEBUG nova.compute.claims [None req-c09a27e9-8e33-411f-b17b-7cd8d66a175c tempest-ServerMetadataNegativeTestJSON-1976245157 tempest-ServerMetadataNegativeTestJSON-1976245157-project-member] [instance: 25b62db6-6e64-4e38-bfd3-82a2e9a4883a] Aborting claim: {{(pid=61923) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 705.262376] env[61923]: DEBUG oslo_concurrency.lockutils [None req-c09a27e9-8e33-411f-b17b-7cd8d66a175c tempest-ServerMetadataNegativeTestJSON-1976245157 tempest-ServerMetadataNegativeTestJSON-1976245157-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 705.327519] env[61923]: DEBUG oslo_concurrency.lockutils [req-2c2fe3c4-9243-4164-833c-2bd3b3b839d7 req-9503d4ea-5ff3-4131-a184-c3bbb79acf38 service nova] Releasing lock "refresh_cache-25b62db6-6e64-4e38-bfd3-82a2e9a4883a" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 705.327971] env[61923]: DEBUG nova.compute.manager [req-2c2fe3c4-9243-4164-833c-2bd3b3b839d7 req-9503d4ea-5ff3-4131-a184-c3bbb79acf38 service nova] [instance: 25b62db6-6e64-4e38-bfd3-82a2e9a4883a] Received event network-vif-deleted-bf236ea4-19e2-487c-ad15-61be750ff9c5 {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 705.726535] env[61923]: DEBUG nova.compute.utils [None req-24801cc4-3974-42da-978f-a22cb8f70af0 tempest-SecurityGroupsTestJSON-509490211 tempest-SecurityGroupsTestJSON-509490211-project-member] Using /dev/sd instead of None {{(pid=61923) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 705.731413] env[61923]: DEBUG nova.compute.manager [None req-24801cc4-3974-42da-978f-a22cb8f70af0 tempest-SecurityGroupsTestJSON-509490211 tempest-SecurityGroupsTestJSON-509490211-project-member] [instance: 2939d1e5-07fe-48b3-bc87-5f67fa1a7ae4] Allocating IP information in the background. {{(pid=61923) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 705.731413] env[61923]: DEBUG nova.network.neutron [None req-24801cc4-3974-42da-978f-a22cb8f70af0 tempest-SecurityGroupsTestJSON-509490211 tempest-SecurityGroupsTestJSON-509490211-project-member] [instance: 2939d1e5-07fe-48b3-bc87-5f67fa1a7ae4] allocate_for_instance() {{(pid=61923) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 705.788522] env[61923]: DEBUG nova.policy [None req-24801cc4-3974-42da-978f-a22cb8f70af0 tempest-SecurityGroupsTestJSON-509490211 tempest-SecurityGroupsTestJSON-509490211-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0237dcc9ff3e4b5e919f8c921d16268e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ebd1ac0807c34d0597971c074f4b6853', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61923) authorize /opt/stack/nova/nova/policy.py:201}} [ 705.960476] env[61923]: INFO nova.scheduler.client.report [None req-1f24f04c-cf4b-4940-9447-eacc11fa5e66 tempest-ServersTestFqdnHostnames-345291565 tempest-ServersTestFqdnHostnames-345291565-project-member] Deleted allocations for instance c94e8e46-7697-426c-ae2f-aece493fa8f4 [ 706.084988] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00cbdc68-05d5-46e2-8e42-60796cb7849d {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.092749] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9351bc68-e9d8-497d-a698-1a5ee79ec7c2 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.121746] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bcbd91c5-29d3-49da-bdd0-028543fc179e {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.129139] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-223b9aa9-6d89-47f6-8cc3-8736825fadc6 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.142697] env[61923]: DEBUG nova.compute.provider_tree [None req-aadda1bb-f770-4742-8c34-cd25e3f7e7cb tempest-InstanceActionsNegativeTestJSON-1700332362 tempest-InstanceActionsNegativeTestJSON-1700332362-project-member] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 706.231432] env[61923]: DEBUG nova.compute.manager [None req-24801cc4-3974-42da-978f-a22cb8f70af0 tempest-SecurityGroupsTestJSON-509490211 tempest-SecurityGroupsTestJSON-509490211-project-member] [instance: 2939d1e5-07fe-48b3-bc87-5f67fa1a7ae4] Start building block device mappings for instance. {{(pid=61923) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 706.235247] env[61923]: DEBUG nova.network.neutron [None req-24801cc4-3974-42da-978f-a22cb8f70af0 tempest-SecurityGroupsTestJSON-509490211 tempest-SecurityGroupsTestJSON-509490211-project-member] [instance: 2939d1e5-07fe-48b3-bc87-5f67fa1a7ae4] Successfully created port: cb32985d-46e9-42d3-a1cd-7e7620873e5e {{(pid=61923) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 706.472883] env[61923]: DEBUG oslo_concurrency.lockutils [None req-1f24f04c-cf4b-4940-9447-eacc11fa5e66 tempest-ServersTestFqdnHostnames-345291565 tempest-ServersTestFqdnHostnames-345291565-project-member] Lock "c94e8e46-7697-426c-ae2f-aece493fa8f4" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 129.894s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 706.647382] env[61923]: DEBUG nova.scheduler.client.report [None req-aadda1bb-f770-4742-8c34-cd25e3f7e7cb tempest-InstanceActionsNegativeTestJSON-1700332362 tempest-InstanceActionsNegativeTestJSON-1700332362-project-member] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 706.974071] env[61923]: DEBUG nova.compute.manager [None req-40b011d4-39ab-4684-a124-b3ba87d489df tempest-ServerPasswordTestJSON-595643156 tempest-ServerPasswordTestJSON-595643156-project-member] [instance: 7c0341f9-7401-42ee-9512-afd832c76940] Starting instance... {{(pid=61923) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 707.152938] env[61923]: DEBUG oslo_concurrency.lockutils [None req-aadda1bb-f770-4742-8c34-cd25e3f7e7cb tempest-InstanceActionsNegativeTestJSON-1700332362 tempest-InstanceActionsNegativeTestJSON-1700332362-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.930s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 707.152938] env[61923]: ERROR nova.compute.manager [None req-aadda1bb-f770-4742-8c34-cd25e3f7e7cb tempest-InstanceActionsNegativeTestJSON-1700332362 tempest-InstanceActionsNegativeTestJSON-1700332362-project-member] [instance: 64584976-b3f3-4da5-a76b-a05cf2ed6aa4] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 13363b75-27bb-4922-b5da-febd2dc08e25, please check neutron logs for more information. [ 707.152938] env[61923]: ERROR nova.compute.manager [instance: 64584976-b3f3-4da5-a76b-a05cf2ed6aa4] Traceback (most recent call last): [ 707.152938] env[61923]: ERROR nova.compute.manager [instance: 64584976-b3f3-4da5-a76b-a05cf2ed6aa4] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 707.152938] env[61923]: ERROR nova.compute.manager [instance: 64584976-b3f3-4da5-a76b-a05cf2ed6aa4] self.driver.spawn(context, instance, image_meta, [ 707.152938] env[61923]: ERROR nova.compute.manager [instance: 64584976-b3f3-4da5-a76b-a05cf2ed6aa4] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 707.152938] env[61923]: ERROR nova.compute.manager [instance: 64584976-b3f3-4da5-a76b-a05cf2ed6aa4] self._vmops.spawn(context, instance, image_meta, injected_files, [ 707.152938] env[61923]: ERROR nova.compute.manager [instance: 64584976-b3f3-4da5-a76b-a05cf2ed6aa4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 707.152938] env[61923]: ERROR nova.compute.manager [instance: 64584976-b3f3-4da5-a76b-a05cf2ed6aa4] vm_ref = self.build_virtual_machine(instance, [ 707.153311] env[61923]: ERROR nova.compute.manager [instance: 64584976-b3f3-4da5-a76b-a05cf2ed6aa4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 707.153311] env[61923]: ERROR nova.compute.manager [instance: 64584976-b3f3-4da5-a76b-a05cf2ed6aa4] vif_infos = vmwarevif.get_vif_info(self._session, [ 707.153311] env[61923]: ERROR nova.compute.manager [instance: 64584976-b3f3-4da5-a76b-a05cf2ed6aa4] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 707.153311] env[61923]: ERROR nova.compute.manager [instance: 64584976-b3f3-4da5-a76b-a05cf2ed6aa4] for vif in network_info: [ 707.153311] env[61923]: ERROR nova.compute.manager [instance: 64584976-b3f3-4da5-a76b-a05cf2ed6aa4] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 707.153311] env[61923]: ERROR nova.compute.manager [instance: 64584976-b3f3-4da5-a76b-a05cf2ed6aa4] return self._sync_wrapper(fn, *args, **kwargs) [ 707.153311] env[61923]: ERROR nova.compute.manager [instance: 64584976-b3f3-4da5-a76b-a05cf2ed6aa4] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 707.153311] env[61923]: ERROR nova.compute.manager [instance: 64584976-b3f3-4da5-a76b-a05cf2ed6aa4] self.wait() [ 707.153311] env[61923]: ERROR nova.compute.manager [instance: 64584976-b3f3-4da5-a76b-a05cf2ed6aa4] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 707.153311] env[61923]: ERROR nova.compute.manager [instance: 64584976-b3f3-4da5-a76b-a05cf2ed6aa4] self[:] = self._gt.wait() [ 707.153311] env[61923]: ERROR nova.compute.manager [instance: 64584976-b3f3-4da5-a76b-a05cf2ed6aa4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 707.153311] env[61923]: ERROR nova.compute.manager [instance: 64584976-b3f3-4da5-a76b-a05cf2ed6aa4] return self._exit_event.wait() [ 707.153311] env[61923]: ERROR nova.compute.manager [instance: 64584976-b3f3-4da5-a76b-a05cf2ed6aa4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 707.153753] env[61923]: ERROR nova.compute.manager [instance: 64584976-b3f3-4da5-a76b-a05cf2ed6aa4] result = hub.switch() [ 707.153753] env[61923]: ERROR nova.compute.manager [instance: 64584976-b3f3-4da5-a76b-a05cf2ed6aa4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 707.153753] env[61923]: ERROR nova.compute.manager [instance: 64584976-b3f3-4da5-a76b-a05cf2ed6aa4] return self.greenlet.switch() [ 707.153753] env[61923]: ERROR nova.compute.manager [instance: 64584976-b3f3-4da5-a76b-a05cf2ed6aa4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 707.153753] env[61923]: ERROR nova.compute.manager [instance: 64584976-b3f3-4da5-a76b-a05cf2ed6aa4] result = function(*args, **kwargs) [ 707.153753] env[61923]: ERROR nova.compute.manager [instance: 64584976-b3f3-4da5-a76b-a05cf2ed6aa4] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 707.153753] env[61923]: ERROR nova.compute.manager [instance: 64584976-b3f3-4da5-a76b-a05cf2ed6aa4] return func(*args, **kwargs) [ 707.153753] env[61923]: ERROR nova.compute.manager [instance: 64584976-b3f3-4da5-a76b-a05cf2ed6aa4] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 707.153753] env[61923]: ERROR nova.compute.manager [instance: 64584976-b3f3-4da5-a76b-a05cf2ed6aa4] raise e [ 707.153753] env[61923]: ERROR nova.compute.manager [instance: 64584976-b3f3-4da5-a76b-a05cf2ed6aa4] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 707.153753] env[61923]: ERROR nova.compute.manager [instance: 64584976-b3f3-4da5-a76b-a05cf2ed6aa4] nwinfo = self.network_api.allocate_for_instance( [ 707.153753] env[61923]: ERROR nova.compute.manager [instance: 64584976-b3f3-4da5-a76b-a05cf2ed6aa4] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 707.153753] env[61923]: ERROR nova.compute.manager [instance: 64584976-b3f3-4da5-a76b-a05cf2ed6aa4] created_port_ids = self._update_ports_for_instance( [ 707.154117] env[61923]: ERROR nova.compute.manager [instance: 64584976-b3f3-4da5-a76b-a05cf2ed6aa4] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 707.154117] env[61923]: ERROR nova.compute.manager [instance: 64584976-b3f3-4da5-a76b-a05cf2ed6aa4] with excutils.save_and_reraise_exception(): [ 707.154117] env[61923]: ERROR nova.compute.manager [instance: 64584976-b3f3-4da5-a76b-a05cf2ed6aa4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 707.154117] env[61923]: ERROR nova.compute.manager [instance: 64584976-b3f3-4da5-a76b-a05cf2ed6aa4] self.force_reraise() [ 707.154117] env[61923]: ERROR nova.compute.manager [instance: 64584976-b3f3-4da5-a76b-a05cf2ed6aa4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 707.154117] env[61923]: ERROR nova.compute.manager [instance: 64584976-b3f3-4da5-a76b-a05cf2ed6aa4] raise self.value [ 707.154117] env[61923]: ERROR nova.compute.manager [instance: 64584976-b3f3-4da5-a76b-a05cf2ed6aa4] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 707.154117] env[61923]: ERROR nova.compute.manager [instance: 64584976-b3f3-4da5-a76b-a05cf2ed6aa4] updated_port = self._update_port( [ 707.154117] env[61923]: ERROR nova.compute.manager [instance: 64584976-b3f3-4da5-a76b-a05cf2ed6aa4] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 707.154117] env[61923]: ERROR nova.compute.manager [instance: 64584976-b3f3-4da5-a76b-a05cf2ed6aa4] _ensure_no_port_binding_failure(port) [ 707.154117] env[61923]: ERROR nova.compute.manager [instance: 64584976-b3f3-4da5-a76b-a05cf2ed6aa4] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 707.154117] env[61923]: ERROR nova.compute.manager [instance: 64584976-b3f3-4da5-a76b-a05cf2ed6aa4] raise exception.PortBindingFailed(port_id=port['id']) [ 707.154486] env[61923]: ERROR nova.compute.manager [instance: 64584976-b3f3-4da5-a76b-a05cf2ed6aa4] nova.exception.PortBindingFailed: Binding failed for port 13363b75-27bb-4922-b5da-febd2dc08e25, please check neutron logs for more information. [ 707.154486] env[61923]: ERROR nova.compute.manager [instance: 64584976-b3f3-4da5-a76b-a05cf2ed6aa4] [ 707.154486] env[61923]: DEBUG nova.compute.utils [None req-aadda1bb-f770-4742-8c34-cd25e3f7e7cb tempest-InstanceActionsNegativeTestJSON-1700332362 tempest-InstanceActionsNegativeTestJSON-1700332362-project-member] [instance: 64584976-b3f3-4da5-a76b-a05cf2ed6aa4] Binding failed for port 13363b75-27bb-4922-b5da-febd2dc08e25, please check neutron logs for more information. {{(pid=61923) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 707.155913] env[61923]: DEBUG oslo_concurrency.lockutils [None req-5f08bfba-63c5-4161-be1b-a884c0ae71af tempest-VolumesAdminNegativeTest-1422120048 tempest-VolumesAdminNegativeTest-1422120048-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.672s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 707.159484] env[61923]: DEBUG nova.compute.manager [None req-aadda1bb-f770-4742-8c34-cd25e3f7e7cb tempest-InstanceActionsNegativeTestJSON-1700332362 tempest-InstanceActionsNegativeTestJSON-1700332362-project-member] [instance: 64584976-b3f3-4da5-a76b-a05cf2ed6aa4] Build of instance 64584976-b3f3-4da5-a76b-a05cf2ed6aa4 was re-scheduled: Binding failed for port 13363b75-27bb-4922-b5da-febd2dc08e25, please check neutron logs for more information. {{(pid=61923) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 707.159484] env[61923]: DEBUG nova.compute.manager [None req-aadda1bb-f770-4742-8c34-cd25e3f7e7cb tempest-InstanceActionsNegativeTestJSON-1700332362 tempest-InstanceActionsNegativeTestJSON-1700332362-project-member] [instance: 64584976-b3f3-4da5-a76b-a05cf2ed6aa4] Unplugging VIFs for instance {{(pid=61923) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 707.159484] env[61923]: DEBUG oslo_concurrency.lockutils [None req-aadda1bb-f770-4742-8c34-cd25e3f7e7cb tempest-InstanceActionsNegativeTestJSON-1700332362 tempest-InstanceActionsNegativeTestJSON-1700332362-project-member] Acquiring lock "refresh_cache-64584976-b3f3-4da5-a76b-a05cf2ed6aa4" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 707.159484] env[61923]: DEBUG oslo_concurrency.lockutils [None req-aadda1bb-f770-4742-8c34-cd25e3f7e7cb tempest-InstanceActionsNegativeTestJSON-1700332362 tempest-InstanceActionsNegativeTestJSON-1700332362-project-member] Acquired lock "refresh_cache-64584976-b3f3-4da5-a76b-a05cf2ed6aa4" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 707.159708] env[61923]: DEBUG nova.network.neutron [None req-aadda1bb-f770-4742-8c34-cd25e3f7e7cb tempest-InstanceActionsNegativeTestJSON-1700332362 tempest-InstanceActionsNegativeTestJSON-1700332362-project-member] [instance: 64584976-b3f3-4da5-a76b-a05cf2ed6aa4] Building network info cache for instance {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 707.244599] env[61923]: DEBUG nova.compute.manager [None req-24801cc4-3974-42da-978f-a22cb8f70af0 tempest-SecurityGroupsTestJSON-509490211 tempest-SecurityGroupsTestJSON-509490211-project-member] [instance: 2939d1e5-07fe-48b3-bc87-5f67fa1a7ae4] Start spawning the instance on the hypervisor. {{(pid=61923) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 707.270893] env[61923]: DEBUG nova.virt.hardware [None req-24801cc4-3974-42da-978f-a22cb8f70af0 tempest-SecurityGroupsTestJSON-509490211 tempest-SecurityGroupsTestJSON-509490211-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-29T20:07:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-29T20:07:35Z,direct_url=,disk_format='vmdk',id=0f153f63-ae0a-45b1-b7f5-7f9b673c947f,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='4e7b5e3b3c3443c8bd5c70f8a15739f5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-29T20:07:36Z,virtual_size=,visibility=), allow threads: False {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 707.271131] env[61923]: DEBUG nova.virt.hardware [None req-24801cc4-3974-42da-978f-a22cb8f70af0 tempest-SecurityGroupsTestJSON-509490211 tempest-SecurityGroupsTestJSON-509490211-project-member] Flavor limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 707.271279] env[61923]: DEBUG nova.virt.hardware [None req-24801cc4-3974-42da-978f-a22cb8f70af0 tempest-SecurityGroupsTestJSON-509490211 tempest-SecurityGroupsTestJSON-509490211-project-member] Image limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 707.271452] env[61923]: DEBUG nova.virt.hardware [None req-24801cc4-3974-42da-978f-a22cb8f70af0 tempest-SecurityGroupsTestJSON-509490211 tempest-SecurityGroupsTestJSON-509490211-project-member] Flavor pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 707.271589] env[61923]: DEBUG nova.virt.hardware [None req-24801cc4-3974-42da-978f-a22cb8f70af0 tempest-SecurityGroupsTestJSON-509490211 tempest-SecurityGroupsTestJSON-509490211-project-member] Image pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 707.271728] env[61923]: DEBUG nova.virt.hardware [None req-24801cc4-3974-42da-978f-a22cb8f70af0 tempest-SecurityGroupsTestJSON-509490211 tempest-SecurityGroupsTestJSON-509490211-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 707.271927] env[61923]: DEBUG nova.virt.hardware [None req-24801cc4-3974-42da-978f-a22cb8f70af0 tempest-SecurityGroupsTestJSON-509490211 tempest-SecurityGroupsTestJSON-509490211-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 707.272090] env[61923]: DEBUG nova.virt.hardware [None req-24801cc4-3974-42da-978f-a22cb8f70af0 tempest-SecurityGroupsTestJSON-509490211 tempest-SecurityGroupsTestJSON-509490211-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 707.272252] env[61923]: DEBUG nova.virt.hardware [None req-24801cc4-3974-42da-978f-a22cb8f70af0 tempest-SecurityGroupsTestJSON-509490211 tempest-SecurityGroupsTestJSON-509490211-project-member] Got 1 possible topologies {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 707.272405] env[61923]: DEBUG nova.virt.hardware [None req-24801cc4-3974-42da-978f-a22cb8f70af0 tempest-SecurityGroupsTestJSON-509490211 tempest-SecurityGroupsTestJSON-509490211-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 707.272567] env[61923]: DEBUG nova.virt.hardware [None req-24801cc4-3974-42da-978f-a22cb8f70af0 tempest-SecurityGroupsTestJSON-509490211 tempest-SecurityGroupsTestJSON-509490211-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 707.273534] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef4d37d0-d494-4130-bf63-c1293ffb5f01 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.281565] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0590733-e860-4c64-a68a-79d5a44f1bf3 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 707.346359] env[61923]: DEBUG nova.compute.manager [req-7f736d39-3ccb-4328-925e-2aac55186c77 req-49221ae2-2f1d-4094-bd5f-1fdf5ce18a29 service nova] [instance: 2939d1e5-07fe-48b3-bc87-5f67fa1a7ae4] Received event network-changed-cb32985d-46e9-42d3-a1cd-7e7620873e5e {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 707.346563] env[61923]: DEBUG nova.compute.manager [req-7f736d39-3ccb-4328-925e-2aac55186c77 req-49221ae2-2f1d-4094-bd5f-1fdf5ce18a29 service nova] [instance: 2939d1e5-07fe-48b3-bc87-5f67fa1a7ae4] Refreshing instance network info cache due to event network-changed-cb32985d-46e9-42d3-a1cd-7e7620873e5e. {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 707.346784] env[61923]: DEBUG oslo_concurrency.lockutils [req-7f736d39-3ccb-4328-925e-2aac55186c77 req-49221ae2-2f1d-4094-bd5f-1fdf5ce18a29 service nova] Acquiring lock "refresh_cache-2939d1e5-07fe-48b3-bc87-5f67fa1a7ae4" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 707.346924] env[61923]: DEBUG oslo_concurrency.lockutils [req-7f736d39-3ccb-4328-925e-2aac55186c77 req-49221ae2-2f1d-4094-bd5f-1fdf5ce18a29 service nova] Acquired lock "refresh_cache-2939d1e5-07fe-48b3-bc87-5f67fa1a7ae4" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 707.347094] env[61923]: DEBUG nova.network.neutron [req-7f736d39-3ccb-4328-925e-2aac55186c77 req-49221ae2-2f1d-4094-bd5f-1fdf5ce18a29 service nova] [instance: 2939d1e5-07fe-48b3-bc87-5f67fa1a7ae4] Refreshing network info cache for port cb32985d-46e9-42d3-a1cd-7e7620873e5e {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 707.501577] env[61923]: DEBUG oslo_concurrency.lockutils [None req-40b011d4-39ab-4684-a124-b3ba87d489df tempest-ServerPasswordTestJSON-595643156 tempest-ServerPasswordTestJSON-595643156-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 707.595158] env[61923]: ERROR nova.compute.manager [None req-24801cc4-3974-42da-978f-a22cb8f70af0 tempest-SecurityGroupsTestJSON-509490211 tempest-SecurityGroupsTestJSON-509490211-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port cb32985d-46e9-42d3-a1cd-7e7620873e5e, please check neutron logs for more information. [ 707.595158] env[61923]: ERROR nova.compute.manager Traceback (most recent call last): [ 707.595158] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 707.595158] env[61923]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 707.595158] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 707.595158] env[61923]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 707.595158] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 707.595158] env[61923]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 707.595158] env[61923]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 707.595158] env[61923]: ERROR nova.compute.manager self.force_reraise() [ 707.595158] env[61923]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 707.595158] env[61923]: ERROR nova.compute.manager raise self.value [ 707.595158] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 707.595158] env[61923]: ERROR nova.compute.manager updated_port = self._update_port( [ 707.595158] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 707.595158] env[61923]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 707.595561] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 707.595561] env[61923]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 707.595561] env[61923]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port cb32985d-46e9-42d3-a1cd-7e7620873e5e, please check neutron logs for more information. [ 707.595561] env[61923]: ERROR nova.compute.manager [ 707.595561] env[61923]: Traceback (most recent call last): [ 707.595561] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 707.595561] env[61923]: listener.cb(fileno) [ 707.595561] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 707.595561] env[61923]: result = function(*args, **kwargs) [ 707.595561] env[61923]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 707.595561] env[61923]: return func(*args, **kwargs) [ 707.595561] env[61923]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 707.595561] env[61923]: raise e [ 707.595561] env[61923]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 707.595561] env[61923]: nwinfo = self.network_api.allocate_for_instance( [ 707.595561] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 707.595561] env[61923]: created_port_ids = self._update_ports_for_instance( [ 707.595561] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 707.595561] env[61923]: with excutils.save_and_reraise_exception(): [ 707.595561] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 707.595561] env[61923]: self.force_reraise() [ 707.595561] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 707.595561] env[61923]: raise self.value [ 707.595561] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 707.595561] env[61923]: updated_port = self._update_port( [ 707.595561] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 707.595561] env[61923]: _ensure_no_port_binding_failure(port) [ 707.595561] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 707.595561] env[61923]: raise exception.PortBindingFailed(port_id=port['id']) [ 707.596293] env[61923]: nova.exception.PortBindingFailed: Binding failed for port cb32985d-46e9-42d3-a1cd-7e7620873e5e, please check neutron logs for more information. [ 707.596293] env[61923]: Removing descriptor: 17 [ 707.596293] env[61923]: ERROR nova.compute.manager [None req-24801cc4-3974-42da-978f-a22cb8f70af0 tempest-SecurityGroupsTestJSON-509490211 tempest-SecurityGroupsTestJSON-509490211-project-member] [instance: 2939d1e5-07fe-48b3-bc87-5f67fa1a7ae4] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port cb32985d-46e9-42d3-a1cd-7e7620873e5e, please check neutron logs for more information. [ 707.596293] env[61923]: ERROR nova.compute.manager [instance: 2939d1e5-07fe-48b3-bc87-5f67fa1a7ae4] Traceback (most recent call last): [ 707.596293] env[61923]: ERROR nova.compute.manager [instance: 2939d1e5-07fe-48b3-bc87-5f67fa1a7ae4] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 707.596293] env[61923]: ERROR nova.compute.manager [instance: 2939d1e5-07fe-48b3-bc87-5f67fa1a7ae4] yield resources [ 707.596293] env[61923]: ERROR nova.compute.manager [instance: 2939d1e5-07fe-48b3-bc87-5f67fa1a7ae4] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 707.596293] env[61923]: ERROR nova.compute.manager [instance: 2939d1e5-07fe-48b3-bc87-5f67fa1a7ae4] self.driver.spawn(context, instance, image_meta, [ 707.596293] env[61923]: ERROR nova.compute.manager [instance: 2939d1e5-07fe-48b3-bc87-5f67fa1a7ae4] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 707.596293] env[61923]: ERROR nova.compute.manager [instance: 2939d1e5-07fe-48b3-bc87-5f67fa1a7ae4] self._vmops.spawn(context, instance, image_meta, injected_files, [ 707.596293] env[61923]: ERROR nova.compute.manager [instance: 2939d1e5-07fe-48b3-bc87-5f67fa1a7ae4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 707.596293] env[61923]: ERROR nova.compute.manager [instance: 2939d1e5-07fe-48b3-bc87-5f67fa1a7ae4] vm_ref = self.build_virtual_machine(instance, [ 707.596687] env[61923]: ERROR nova.compute.manager [instance: 2939d1e5-07fe-48b3-bc87-5f67fa1a7ae4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 707.596687] env[61923]: ERROR nova.compute.manager [instance: 2939d1e5-07fe-48b3-bc87-5f67fa1a7ae4] vif_infos = vmwarevif.get_vif_info(self._session, [ 707.596687] env[61923]: ERROR nova.compute.manager [instance: 2939d1e5-07fe-48b3-bc87-5f67fa1a7ae4] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 707.596687] env[61923]: ERROR nova.compute.manager [instance: 2939d1e5-07fe-48b3-bc87-5f67fa1a7ae4] for vif in network_info: [ 707.596687] env[61923]: ERROR nova.compute.manager [instance: 2939d1e5-07fe-48b3-bc87-5f67fa1a7ae4] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 707.596687] env[61923]: ERROR nova.compute.manager [instance: 2939d1e5-07fe-48b3-bc87-5f67fa1a7ae4] return self._sync_wrapper(fn, *args, **kwargs) [ 707.596687] env[61923]: ERROR nova.compute.manager [instance: 2939d1e5-07fe-48b3-bc87-5f67fa1a7ae4] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 707.596687] env[61923]: ERROR nova.compute.manager [instance: 2939d1e5-07fe-48b3-bc87-5f67fa1a7ae4] self.wait() [ 707.596687] env[61923]: ERROR nova.compute.manager [instance: 2939d1e5-07fe-48b3-bc87-5f67fa1a7ae4] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 707.596687] env[61923]: ERROR nova.compute.manager [instance: 2939d1e5-07fe-48b3-bc87-5f67fa1a7ae4] self[:] = self._gt.wait() [ 707.596687] env[61923]: ERROR nova.compute.manager [instance: 2939d1e5-07fe-48b3-bc87-5f67fa1a7ae4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 707.596687] env[61923]: ERROR nova.compute.manager [instance: 2939d1e5-07fe-48b3-bc87-5f67fa1a7ae4] return self._exit_event.wait() [ 707.596687] env[61923]: ERROR nova.compute.manager [instance: 2939d1e5-07fe-48b3-bc87-5f67fa1a7ae4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 707.597008] env[61923]: ERROR nova.compute.manager [instance: 2939d1e5-07fe-48b3-bc87-5f67fa1a7ae4] result = hub.switch() [ 707.597008] env[61923]: ERROR nova.compute.manager [instance: 2939d1e5-07fe-48b3-bc87-5f67fa1a7ae4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 707.597008] env[61923]: ERROR nova.compute.manager [instance: 2939d1e5-07fe-48b3-bc87-5f67fa1a7ae4] return self.greenlet.switch() [ 707.597008] env[61923]: ERROR nova.compute.manager [instance: 2939d1e5-07fe-48b3-bc87-5f67fa1a7ae4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 707.597008] env[61923]: ERROR nova.compute.manager [instance: 2939d1e5-07fe-48b3-bc87-5f67fa1a7ae4] result = function(*args, **kwargs) [ 707.597008] env[61923]: ERROR nova.compute.manager [instance: 2939d1e5-07fe-48b3-bc87-5f67fa1a7ae4] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 707.597008] env[61923]: ERROR nova.compute.manager [instance: 2939d1e5-07fe-48b3-bc87-5f67fa1a7ae4] return func(*args, **kwargs) [ 707.597008] env[61923]: ERROR nova.compute.manager [instance: 2939d1e5-07fe-48b3-bc87-5f67fa1a7ae4] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 707.597008] env[61923]: ERROR nova.compute.manager [instance: 2939d1e5-07fe-48b3-bc87-5f67fa1a7ae4] raise e [ 707.597008] env[61923]: ERROR nova.compute.manager [instance: 2939d1e5-07fe-48b3-bc87-5f67fa1a7ae4] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 707.597008] env[61923]: ERROR nova.compute.manager [instance: 2939d1e5-07fe-48b3-bc87-5f67fa1a7ae4] nwinfo = self.network_api.allocate_for_instance( [ 707.597008] env[61923]: ERROR nova.compute.manager [instance: 2939d1e5-07fe-48b3-bc87-5f67fa1a7ae4] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 707.597008] env[61923]: ERROR nova.compute.manager [instance: 2939d1e5-07fe-48b3-bc87-5f67fa1a7ae4] created_port_ids = self._update_ports_for_instance( [ 707.597328] env[61923]: ERROR nova.compute.manager [instance: 2939d1e5-07fe-48b3-bc87-5f67fa1a7ae4] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 707.597328] env[61923]: ERROR nova.compute.manager [instance: 2939d1e5-07fe-48b3-bc87-5f67fa1a7ae4] with excutils.save_and_reraise_exception(): [ 707.597328] env[61923]: ERROR nova.compute.manager [instance: 2939d1e5-07fe-48b3-bc87-5f67fa1a7ae4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 707.597328] env[61923]: ERROR nova.compute.manager [instance: 2939d1e5-07fe-48b3-bc87-5f67fa1a7ae4] self.force_reraise() [ 707.597328] env[61923]: ERROR nova.compute.manager [instance: 2939d1e5-07fe-48b3-bc87-5f67fa1a7ae4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 707.597328] env[61923]: ERROR nova.compute.manager [instance: 2939d1e5-07fe-48b3-bc87-5f67fa1a7ae4] raise self.value [ 707.597328] env[61923]: ERROR nova.compute.manager [instance: 2939d1e5-07fe-48b3-bc87-5f67fa1a7ae4] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 707.597328] env[61923]: ERROR nova.compute.manager [instance: 2939d1e5-07fe-48b3-bc87-5f67fa1a7ae4] updated_port = self._update_port( [ 707.597328] env[61923]: ERROR nova.compute.manager [instance: 2939d1e5-07fe-48b3-bc87-5f67fa1a7ae4] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 707.597328] env[61923]: ERROR nova.compute.manager [instance: 2939d1e5-07fe-48b3-bc87-5f67fa1a7ae4] _ensure_no_port_binding_failure(port) [ 707.597328] env[61923]: ERROR nova.compute.manager [instance: 2939d1e5-07fe-48b3-bc87-5f67fa1a7ae4] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 707.597328] env[61923]: ERROR nova.compute.manager [instance: 2939d1e5-07fe-48b3-bc87-5f67fa1a7ae4] raise exception.PortBindingFailed(port_id=port['id']) [ 707.597628] env[61923]: ERROR nova.compute.manager [instance: 2939d1e5-07fe-48b3-bc87-5f67fa1a7ae4] nova.exception.PortBindingFailed: Binding failed for port cb32985d-46e9-42d3-a1cd-7e7620873e5e, please check neutron logs for more information. [ 707.597628] env[61923]: ERROR nova.compute.manager [instance: 2939d1e5-07fe-48b3-bc87-5f67fa1a7ae4] [ 707.597628] env[61923]: INFO nova.compute.manager [None req-24801cc4-3974-42da-978f-a22cb8f70af0 tempest-SecurityGroupsTestJSON-509490211 tempest-SecurityGroupsTestJSON-509490211-project-member] [instance: 2939d1e5-07fe-48b3-bc87-5f67fa1a7ae4] Terminating instance [ 707.598691] env[61923]: DEBUG oslo_concurrency.lockutils [None req-24801cc4-3974-42da-978f-a22cb8f70af0 tempest-SecurityGroupsTestJSON-509490211 tempest-SecurityGroupsTestJSON-509490211-project-member] Acquiring lock "refresh_cache-2939d1e5-07fe-48b3-bc87-5f67fa1a7ae4" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 707.693440] env[61923]: DEBUG nova.network.neutron [None req-aadda1bb-f770-4742-8c34-cd25e3f7e7cb tempest-InstanceActionsNegativeTestJSON-1700332362 tempest-InstanceActionsNegativeTestJSON-1700332362-project-member] [instance: 64584976-b3f3-4da5-a76b-a05cf2ed6aa4] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 707.815858] env[61923]: DEBUG nova.network.neutron [None req-aadda1bb-f770-4742-8c34-cd25e3f7e7cb tempest-InstanceActionsNegativeTestJSON-1700332362 tempest-InstanceActionsNegativeTestJSON-1700332362-project-member] [instance: 64584976-b3f3-4da5-a76b-a05cf2ed6aa4] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 707.867064] env[61923]: DEBUG nova.network.neutron [req-7f736d39-3ccb-4328-925e-2aac55186c77 req-49221ae2-2f1d-4094-bd5f-1fdf5ce18a29 service nova] [instance: 2939d1e5-07fe-48b3-bc87-5f67fa1a7ae4] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 707.982331] env[61923]: DEBUG nova.network.neutron [req-7f736d39-3ccb-4328-925e-2aac55186c77 req-49221ae2-2f1d-4094-bd5f-1fdf5ce18a29 service nova] [instance: 2939d1e5-07fe-48b3-bc87-5f67fa1a7ae4] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 708.071005] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53c0d87b-b3b9-446d-a59f-1434c10d7c3e {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.078518] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5bebe738-dfad-4d7d-b9ce-01ad63708ada {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.108844] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be66f162-fe53-4999-acbc-57892d11ee2f {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.115768] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36a685eb-2e7a-42e2-b2a7-615415a46900 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.128438] env[61923]: DEBUG nova.compute.provider_tree [None req-5f08bfba-63c5-4161-be1b-a884c0ae71af tempest-VolumesAdminNegativeTest-1422120048 tempest-VolumesAdminNegativeTest-1422120048-project-member] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 708.319043] env[61923]: DEBUG oslo_concurrency.lockutils [None req-aadda1bb-f770-4742-8c34-cd25e3f7e7cb tempest-InstanceActionsNegativeTestJSON-1700332362 tempest-InstanceActionsNegativeTestJSON-1700332362-project-member] Releasing lock "refresh_cache-64584976-b3f3-4da5-a76b-a05cf2ed6aa4" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 708.319043] env[61923]: DEBUG nova.compute.manager [None req-aadda1bb-f770-4742-8c34-cd25e3f7e7cb tempest-InstanceActionsNegativeTestJSON-1700332362 tempest-InstanceActionsNegativeTestJSON-1700332362-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61923) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 708.319043] env[61923]: DEBUG nova.compute.manager [None req-aadda1bb-f770-4742-8c34-cd25e3f7e7cb tempest-InstanceActionsNegativeTestJSON-1700332362 tempest-InstanceActionsNegativeTestJSON-1700332362-project-member] [instance: 64584976-b3f3-4da5-a76b-a05cf2ed6aa4] Deallocating network for instance {{(pid=61923) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 708.319043] env[61923]: DEBUG nova.network.neutron [None req-aadda1bb-f770-4742-8c34-cd25e3f7e7cb tempest-InstanceActionsNegativeTestJSON-1700332362 tempest-InstanceActionsNegativeTestJSON-1700332362-project-member] [instance: 64584976-b3f3-4da5-a76b-a05cf2ed6aa4] deallocate_for_instance() {{(pid=61923) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 708.341633] env[61923]: DEBUG nova.network.neutron [None req-aadda1bb-f770-4742-8c34-cd25e3f7e7cb tempest-InstanceActionsNegativeTestJSON-1700332362 tempest-InstanceActionsNegativeTestJSON-1700332362-project-member] [instance: 64584976-b3f3-4da5-a76b-a05cf2ed6aa4] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 708.485219] env[61923]: DEBUG oslo_concurrency.lockutils [req-7f736d39-3ccb-4328-925e-2aac55186c77 req-49221ae2-2f1d-4094-bd5f-1fdf5ce18a29 service nova] Releasing lock "refresh_cache-2939d1e5-07fe-48b3-bc87-5f67fa1a7ae4" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 708.485655] env[61923]: DEBUG oslo_concurrency.lockutils [None req-24801cc4-3974-42da-978f-a22cb8f70af0 tempest-SecurityGroupsTestJSON-509490211 tempest-SecurityGroupsTestJSON-509490211-project-member] Acquired lock "refresh_cache-2939d1e5-07fe-48b3-bc87-5f67fa1a7ae4" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 708.486047] env[61923]: DEBUG nova.network.neutron [None req-24801cc4-3974-42da-978f-a22cb8f70af0 tempest-SecurityGroupsTestJSON-509490211 tempest-SecurityGroupsTestJSON-509490211-project-member] [instance: 2939d1e5-07fe-48b3-bc87-5f67fa1a7ae4] Building network info cache for instance {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 708.634019] env[61923]: DEBUG nova.scheduler.client.report [None req-5f08bfba-63c5-4161-be1b-a884c0ae71af tempest-VolumesAdminNegativeTest-1422120048 tempest-VolumesAdminNegativeTest-1422120048-project-member] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 708.845218] env[61923]: DEBUG nova.network.neutron [None req-aadda1bb-f770-4742-8c34-cd25e3f7e7cb tempest-InstanceActionsNegativeTestJSON-1700332362 tempest-InstanceActionsNegativeTestJSON-1700332362-project-member] [instance: 64584976-b3f3-4da5-a76b-a05cf2ed6aa4] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 709.007589] env[61923]: DEBUG nova.network.neutron [None req-24801cc4-3974-42da-978f-a22cb8f70af0 tempest-SecurityGroupsTestJSON-509490211 tempest-SecurityGroupsTestJSON-509490211-project-member] [instance: 2939d1e5-07fe-48b3-bc87-5f67fa1a7ae4] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 709.116566] env[61923]: DEBUG nova.network.neutron [None req-24801cc4-3974-42da-978f-a22cb8f70af0 tempest-SecurityGroupsTestJSON-509490211 tempest-SecurityGroupsTestJSON-509490211-project-member] [instance: 2939d1e5-07fe-48b3-bc87-5f67fa1a7ae4] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 709.138218] env[61923]: DEBUG oslo_concurrency.lockutils [None req-5f08bfba-63c5-4161-be1b-a884c0ae71af tempest-VolumesAdminNegativeTest-1422120048 tempest-VolumesAdminNegativeTest-1422120048-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.983s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 709.138996] env[61923]: ERROR nova.compute.manager [None req-5f08bfba-63c5-4161-be1b-a884c0ae71af tempest-VolumesAdminNegativeTest-1422120048 tempest-VolumesAdminNegativeTest-1422120048-project-member] [instance: 590907a7-b6cc-48fc-a4bf-e4a2e48b05b8] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 989fba39-a679-413f-8cca-904ac748fcbe, please check neutron logs for more information. [ 709.138996] env[61923]: ERROR nova.compute.manager [instance: 590907a7-b6cc-48fc-a4bf-e4a2e48b05b8] Traceback (most recent call last): [ 709.138996] env[61923]: ERROR nova.compute.manager [instance: 590907a7-b6cc-48fc-a4bf-e4a2e48b05b8] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 709.138996] env[61923]: ERROR nova.compute.manager [instance: 590907a7-b6cc-48fc-a4bf-e4a2e48b05b8] self.driver.spawn(context, instance, image_meta, [ 709.138996] env[61923]: ERROR nova.compute.manager [instance: 590907a7-b6cc-48fc-a4bf-e4a2e48b05b8] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 709.138996] env[61923]: ERROR nova.compute.manager [instance: 590907a7-b6cc-48fc-a4bf-e4a2e48b05b8] self._vmops.spawn(context, instance, image_meta, injected_files, [ 709.138996] env[61923]: ERROR nova.compute.manager [instance: 590907a7-b6cc-48fc-a4bf-e4a2e48b05b8] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 709.138996] env[61923]: ERROR nova.compute.manager [instance: 590907a7-b6cc-48fc-a4bf-e4a2e48b05b8] vm_ref = self.build_virtual_machine(instance, [ 709.138996] env[61923]: ERROR nova.compute.manager [instance: 590907a7-b6cc-48fc-a4bf-e4a2e48b05b8] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 709.138996] env[61923]: ERROR nova.compute.manager [instance: 590907a7-b6cc-48fc-a4bf-e4a2e48b05b8] vif_infos = vmwarevif.get_vif_info(self._session, [ 709.138996] env[61923]: ERROR nova.compute.manager [instance: 590907a7-b6cc-48fc-a4bf-e4a2e48b05b8] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 709.139398] env[61923]: ERROR nova.compute.manager [instance: 590907a7-b6cc-48fc-a4bf-e4a2e48b05b8] for vif in network_info: [ 709.139398] env[61923]: ERROR nova.compute.manager [instance: 590907a7-b6cc-48fc-a4bf-e4a2e48b05b8] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 709.139398] env[61923]: ERROR nova.compute.manager [instance: 590907a7-b6cc-48fc-a4bf-e4a2e48b05b8] return self._sync_wrapper(fn, *args, **kwargs) [ 709.139398] env[61923]: ERROR nova.compute.manager [instance: 590907a7-b6cc-48fc-a4bf-e4a2e48b05b8] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 709.139398] env[61923]: ERROR nova.compute.manager [instance: 590907a7-b6cc-48fc-a4bf-e4a2e48b05b8] self.wait() [ 709.139398] env[61923]: ERROR nova.compute.manager [instance: 590907a7-b6cc-48fc-a4bf-e4a2e48b05b8] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 709.139398] env[61923]: ERROR nova.compute.manager [instance: 590907a7-b6cc-48fc-a4bf-e4a2e48b05b8] self[:] = self._gt.wait() [ 709.139398] env[61923]: ERROR nova.compute.manager [instance: 590907a7-b6cc-48fc-a4bf-e4a2e48b05b8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 709.139398] env[61923]: ERROR nova.compute.manager [instance: 590907a7-b6cc-48fc-a4bf-e4a2e48b05b8] return self._exit_event.wait() [ 709.139398] env[61923]: ERROR nova.compute.manager [instance: 590907a7-b6cc-48fc-a4bf-e4a2e48b05b8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 709.139398] env[61923]: ERROR nova.compute.manager [instance: 590907a7-b6cc-48fc-a4bf-e4a2e48b05b8] result = hub.switch() [ 709.139398] env[61923]: ERROR nova.compute.manager [instance: 590907a7-b6cc-48fc-a4bf-e4a2e48b05b8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 709.139398] env[61923]: ERROR nova.compute.manager [instance: 590907a7-b6cc-48fc-a4bf-e4a2e48b05b8] return self.greenlet.switch() [ 709.139775] env[61923]: ERROR nova.compute.manager [instance: 590907a7-b6cc-48fc-a4bf-e4a2e48b05b8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 709.139775] env[61923]: ERROR nova.compute.manager [instance: 590907a7-b6cc-48fc-a4bf-e4a2e48b05b8] result = function(*args, **kwargs) [ 709.139775] env[61923]: ERROR nova.compute.manager [instance: 590907a7-b6cc-48fc-a4bf-e4a2e48b05b8] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 709.139775] env[61923]: ERROR nova.compute.manager [instance: 590907a7-b6cc-48fc-a4bf-e4a2e48b05b8] return func(*args, **kwargs) [ 709.139775] env[61923]: ERROR nova.compute.manager [instance: 590907a7-b6cc-48fc-a4bf-e4a2e48b05b8] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 709.139775] env[61923]: ERROR nova.compute.manager [instance: 590907a7-b6cc-48fc-a4bf-e4a2e48b05b8] raise e [ 709.139775] env[61923]: ERROR nova.compute.manager [instance: 590907a7-b6cc-48fc-a4bf-e4a2e48b05b8] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 709.139775] env[61923]: ERROR nova.compute.manager [instance: 590907a7-b6cc-48fc-a4bf-e4a2e48b05b8] nwinfo = self.network_api.allocate_for_instance( [ 709.139775] env[61923]: ERROR nova.compute.manager [instance: 590907a7-b6cc-48fc-a4bf-e4a2e48b05b8] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 709.139775] env[61923]: ERROR nova.compute.manager [instance: 590907a7-b6cc-48fc-a4bf-e4a2e48b05b8] created_port_ids = self._update_ports_for_instance( [ 709.139775] env[61923]: ERROR nova.compute.manager [instance: 590907a7-b6cc-48fc-a4bf-e4a2e48b05b8] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 709.139775] env[61923]: ERROR nova.compute.manager [instance: 590907a7-b6cc-48fc-a4bf-e4a2e48b05b8] with excutils.save_and_reraise_exception(): [ 709.139775] env[61923]: ERROR nova.compute.manager [instance: 590907a7-b6cc-48fc-a4bf-e4a2e48b05b8] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 709.140182] env[61923]: ERROR nova.compute.manager [instance: 590907a7-b6cc-48fc-a4bf-e4a2e48b05b8] self.force_reraise() [ 709.140182] env[61923]: ERROR nova.compute.manager [instance: 590907a7-b6cc-48fc-a4bf-e4a2e48b05b8] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 709.140182] env[61923]: ERROR nova.compute.manager [instance: 590907a7-b6cc-48fc-a4bf-e4a2e48b05b8] raise self.value [ 709.140182] env[61923]: ERROR nova.compute.manager [instance: 590907a7-b6cc-48fc-a4bf-e4a2e48b05b8] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 709.140182] env[61923]: ERROR nova.compute.manager [instance: 590907a7-b6cc-48fc-a4bf-e4a2e48b05b8] updated_port = self._update_port( [ 709.140182] env[61923]: ERROR nova.compute.manager [instance: 590907a7-b6cc-48fc-a4bf-e4a2e48b05b8] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 709.140182] env[61923]: ERROR nova.compute.manager [instance: 590907a7-b6cc-48fc-a4bf-e4a2e48b05b8] _ensure_no_port_binding_failure(port) [ 709.140182] env[61923]: ERROR nova.compute.manager [instance: 590907a7-b6cc-48fc-a4bf-e4a2e48b05b8] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 709.140182] env[61923]: ERROR nova.compute.manager [instance: 590907a7-b6cc-48fc-a4bf-e4a2e48b05b8] raise exception.PortBindingFailed(port_id=port['id']) [ 709.140182] env[61923]: ERROR nova.compute.manager [instance: 590907a7-b6cc-48fc-a4bf-e4a2e48b05b8] nova.exception.PortBindingFailed: Binding failed for port 989fba39-a679-413f-8cca-904ac748fcbe, please check neutron logs for more information. [ 709.140182] env[61923]: ERROR nova.compute.manager [instance: 590907a7-b6cc-48fc-a4bf-e4a2e48b05b8] [ 709.141250] env[61923]: DEBUG nova.compute.utils [None req-5f08bfba-63c5-4161-be1b-a884c0ae71af tempest-VolumesAdminNegativeTest-1422120048 tempest-VolumesAdminNegativeTest-1422120048-project-member] [instance: 590907a7-b6cc-48fc-a4bf-e4a2e48b05b8] Binding failed for port 989fba39-a679-413f-8cca-904ac748fcbe, please check neutron logs for more information. {{(pid=61923) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 709.142411] env[61923]: DEBUG oslo_concurrency.lockutils [None req-08382130-d70f-42e8-a5b5-93e86f5fa49a tempest-AttachVolumeShelveTestJSON-1252849694 tempest-AttachVolumeShelveTestJSON-1252849694-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.912s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 709.145879] env[61923]: DEBUG nova.compute.manager [None req-5f08bfba-63c5-4161-be1b-a884c0ae71af tempest-VolumesAdminNegativeTest-1422120048 tempest-VolumesAdminNegativeTest-1422120048-project-member] [instance: 590907a7-b6cc-48fc-a4bf-e4a2e48b05b8] Build of instance 590907a7-b6cc-48fc-a4bf-e4a2e48b05b8 was re-scheduled: Binding failed for port 989fba39-a679-413f-8cca-904ac748fcbe, please check neutron logs for more information. {{(pid=61923) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 709.146196] env[61923]: DEBUG nova.compute.manager [None req-5f08bfba-63c5-4161-be1b-a884c0ae71af tempest-VolumesAdminNegativeTest-1422120048 tempest-VolumesAdminNegativeTest-1422120048-project-member] [instance: 590907a7-b6cc-48fc-a4bf-e4a2e48b05b8] Unplugging VIFs for instance {{(pid=61923) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 709.146424] env[61923]: DEBUG oslo_concurrency.lockutils [None req-5f08bfba-63c5-4161-be1b-a884c0ae71af tempest-VolumesAdminNegativeTest-1422120048 tempest-VolumesAdminNegativeTest-1422120048-project-member] Acquiring lock "refresh_cache-590907a7-b6cc-48fc-a4bf-e4a2e48b05b8" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 709.146567] env[61923]: DEBUG oslo_concurrency.lockutils [None req-5f08bfba-63c5-4161-be1b-a884c0ae71af tempest-VolumesAdminNegativeTest-1422120048 tempest-VolumesAdminNegativeTest-1422120048-project-member] Acquired lock "refresh_cache-590907a7-b6cc-48fc-a4bf-e4a2e48b05b8" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 709.146744] env[61923]: DEBUG nova.network.neutron [None req-5f08bfba-63c5-4161-be1b-a884c0ae71af tempest-VolumesAdminNegativeTest-1422120048 tempest-VolumesAdminNegativeTest-1422120048-project-member] [instance: 590907a7-b6cc-48fc-a4bf-e4a2e48b05b8] Building network info cache for instance {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 709.348404] env[61923]: INFO nova.compute.manager [None req-aadda1bb-f770-4742-8c34-cd25e3f7e7cb tempest-InstanceActionsNegativeTestJSON-1700332362 tempest-InstanceActionsNegativeTestJSON-1700332362-project-member] [instance: 64584976-b3f3-4da5-a76b-a05cf2ed6aa4] Took 1.03 seconds to deallocate network for instance. [ 709.397240] env[61923]: DEBUG nova.compute.manager [req-8f22656e-1db3-4d0c-b032-7b5c24bcbb4f req-6b540d53-54d4-4824-93fd-1ded35d93105 service nova] [instance: 2939d1e5-07fe-48b3-bc87-5f67fa1a7ae4] Received event network-vif-deleted-cb32985d-46e9-42d3-a1cd-7e7620873e5e {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 709.618063] env[61923]: DEBUG oslo_concurrency.lockutils [None req-24801cc4-3974-42da-978f-a22cb8f70af0 tempest-SecurityGroupsTestJSON-509490211 tempest-SecurityGroupsTestJSON-509490211-project-member] Releasing lock "refresh_cache-2939d1e5-07fe-48b3-bc87-5f67fa1a7ae4" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 709.618656] env[61923]: DEBUG nova.compute.manager [None req-24801cc4-3974-42da-978f-a22cb8f70af0 tempest-SecurityGroupsTestJSON-509490211 tempest-SecurityGroupsTestJSON-509490211-project-member] [instance: 2939d1e5-07fe-48b3-bc87-5f67fa1a7ae4] Start destroying the instance on the hypervisor. {{(pid=61923) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 709.618846] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-24801cc4-3974-42da-978f-a22cb8f70af0 tempest-SecurityGroupsTestJSON-509490211 tempest-SecurityGroupsTestJSON-509490211-project-member] [instance: 2939d1e5-07fe-48b3-bc87-5f67fa1a7ae4] Destroying instance {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 709.619205] env[61923]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d7466fd8-c61b-4423-8eb5-abf1b096acdf {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.628200] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8746c1e2-68e0-43ad-bd9c-0c40d2d8d644 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.656720] env[61923]: WARNING nova.virt.vmwareapi.vmops [None req-24801cc4-3974-42da-978f-a22cb8f70af0 tempest-SecurityGroupsTestJSON-509490211 tempest-SecurityGroupsTestJSON-509490211-project-member] [instance: 2939d1e5-07fe-48b3-bc87-5f67fa1a7ae4] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 2939d1e5-07fe-48b3-bc87-5f67fa1a7ae4 could not be found. [ 709.657089] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-24801cc4-3974-42da-978f-a22cb8f70af0 tempest-SecurityGroupsTestJSON-509490211 tempest-SecurityGroupsTestJSON-509490211-project-member] [instance: 2939d1e5-07fe-48b3-bc87-5f67fa1a7ae4] Instance destroyed {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 709.657380] env[61923]: INFO nova.compute.manager [None req-24801cc4-3974-42da-978f-a22cb8f70af0 tempest-SecurityGroupsTestJSON-509490211 tempest-SecurityGroupsTestJSON-509490211-project-member] [instance: 2939d1e5-07fe-48b3-bc87-5f67fa1a7ae4] Took 0.04 seconds to destroy the instance on the hypervisor. [ 709.657842] env[61923]: DEBUG oslo.service.loopingcall [None req-24801cc4-3974-42da-978f-a22cb8f70af0 tempest-SecurityGroupsTestJSON-509490211 tempest-SecurityGroupsTestJSON-509490211-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61923) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 709.659191] env[61923]: DEBUG nova.compute.manager [-] [instance: 2939d1e5-07fe-48b3-bc87-5f67fa1a7ae4] Deallocating network for instance {{(pid=61923) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 709.659404] env[61923]: DEBUG nova.network.neutron [-] [instance: 2939d1e5-07fe-48b3-bc87-5f67fa1a7ae4] deallocate_for_instance() {{(pid=61923) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 709.681760] env[61923]: DEBUG nova.network.neutron [-] [instance: 2939d1e5-07fe-48b3-bc87-5f67fa1a7ae4] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 709.683763] env[61923]: DEBUG nova.network.neutron [None req-5f08bfba-63c5-4161-be1b-a884c0ae71af tempest-VolumesAdminNegativeTest-1422120048 tempest-VolumesAdminNegativeTest-1422120048-project-member] [instance: 590907a7-b6cc-48fc-a4bf-e4a2e48b05b8] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 709.784176] env[61923]: DEBUG nova.network.neutron [None req-5f08bfba-63c5-4161-be1b-a884c0ae71af tempest-VolumesAdminNegativeTest-1422120048 tempest-VolumesAdminNegativeTest-1422120048-project-member] [instance: 590907a7-b6cc-48fc-a4bf-e4a2e48b05b8] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 710.047035] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05e42c54-7420-49bc-8644-3e938266ccd3 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.055374] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf1f5d0e-b980-40ba-b08f-48f2afc08625 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.087369] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-834b2ef8-f37d-450b-a523-a7ff12e9ae03 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.094834] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-191619c4-828f-4867-9bff-735d77a628b5 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.111054] env[61923]: DEBUG nova.compute.provider_tree [None req-08382130-d70f-42e8-a5b5-93e86f5fa49a tempest-AttachVolumeShelveTestJSON-1252849694 tempest-AttachVolumeShelveTestJSON-1252849694-project-member] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 710.188308] env[61923]: DEBUG nova.network.neutron [-] [instance: 2939d1e5-07fe-48b3-bc87-5f67fa1a7ae4] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 710.287585] env[61923]: DEBUG oslo_concurrency.lockutils [None req-5f08bfba-63c5-4161-be1b-a884c0ae71af tempest-VolumesAdminNegativeTest-1422120048 tempest-VolumesAdminNegativeTest-1422120048-project-member] Releasing lock "refresh_cache-590907a7-b6cc-48fc-a4bf-e4a2e48b05b8" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 710.287863] env[61923]: DEBUG nova.compute.manager [None req-5f08bfba-63c5-4161-be1b-a884c0ae71af tempest-VolumesAdminNegativeTest-1422120048 tempest-VolumesAdminNegativeTest-1422120048-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61923) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 710.288063] env[61923]: DEBUG nova.compute.manager [None req-5f08bfba-63c5-4161-be1b-a884c0ae71af tempest-VolumesAdminNegativeTest-1422120048 tempest-VolumesAdminNegativeTest-1422120048-project-member] [instance: 590907a7-b6cc-48fc-a4bf-e4a2e48b05b8] Deallocating network for instance {{(pid=61923) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 710.288238] env[61923]: DEBUG nova.network.neutron [None req-5f08bfba-63c5-4161-be1b-a884c0ae71af tempest-VolumesAdminNegativeTest-1422120048 tempest-VolumesAdminNegativeTest-1422120048-project-member] [instance: 590907a7-b6cc-48fc-a4bf-e4a2e48b05b8] deallocate_for_instance() {{(pid=61923) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 710.303452] env[61923]: DEBUG nova.network.neutron [None req-5f08bfba-63c5-4161-be1b-a884c0ae71af tempest-VolumesAdminNegativeTest-1422120048 tempest-VolumesAdminNegativeTest-1422120048-project-member] [instance: 590907a7-b6cc-48fc-a4bf-e4a2e48b05b8] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 710.375408] env[61923]: INFO nova.scheduler.client.report [None req-aadda1bb-f770-4742-8c34-cd25e3f7e7cb tempest-InstanceActionsNegativeTestJSON-1700332362 tempest-InstanceActionsNegativeTestJSON-1700332362-project-member] Deleted allocations for instance 64584976-b3f3-4da5-a76b-a05cf2ed6aa4 [ 710.614147] env[61923]: DEBUG nova.scheduler.client.report [None req-08382130-d70f-42e8-a5b5-93e86f5fa49a tempest-AttachVolumeShelveTestJSON-1252849694 tempest-AttachVolumeShelveTestJSON-1252849694-project-member] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 710.690644] env[61923]: INFO nova.compute.manager [-] [instance: 2939d1e5-07fe-48b3-bc87-5f67fa1a7ae4] Took 1.03 seconds to deallocate network for instance. [ 710.693313] env[61923]: DEBUG nova.compute.claims [None req-24801cc4-3974-42da-978f-a22cb8f70af0 tempest-SecurityGroupsTestJSON-509490211 tempest-SecurityGroupsTestJSON-509490211-project-member] [instance: 2939d1e5-07fe-48b3-bc87-5f67fa1a7ae4] Aborting claim: {{(pid=61923) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 710.693502] env[61923]: DEBUG oslo_concurrency.lockutils [None req-24801cc4-3974-42da-978f-a22cb8f70af0 tempest-SecurityGroupsTestJSON-509490211 tempest-SecurityGroupsTestJSON-509490211-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 710.806867] env[61923]: DEBUG nova.network.neutron [None req-5f08bfba-63c5-4161-be1b-a884c0ae71af tempest-VolumesAdminNegativeTest-1422120048 tempest-VolumesAdminNegativeTest-1422120048-project-member] [instance: 590907a7-b6cc-48fc-a4bf-e4a2e48b05b8] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 710.883264] env[61923]: DEBUG oslo_concurrency.lockutils [None req-aadda1bb-f770-4742-8c34-cd25e3f7e7cb tempest-InstanceActionsNegativeTestJSON-1700332362 tempest-InstanceActionsNegativeTestJSON-1700332362-project-member] Lock "64584976-b3f3-4da5-a76b-a05cf2ed6aa4" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 134.019s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 711.119445] env[61923]: DEBUG oslo_concurrency.lockutils [None req-08382130-d70f-42e8-a5b5-93e86f5fa49a tempest-AttachVolumeShelveTestJSON-1252849694 tempest-AttachVolumeShelveTestJSON-1252849694-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.977s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 711.120103] env[61923]: ERROR nova.compute.manager [None req-08382130-d70f-42e8-a5b5-93e86f5fa49a tempest-AttachVolumeShelveTestJSON-1252849694 tempest-AttachVolumeShelveTestJSON-1252849694-project-member] [instance: 60ffc8b0-4a7b-4e2a-8774-d16e502d8a06] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port c88b194c-774a-448e-8f31-5a89cfdbc187, please check neutron logs for more information. [ 711.120103] env[61923]: ERROR nova.compute.manager [instance: 60ffc8b0-4a7b-4e2a-8774-d16e502d8a06] Traceback (most recent call last): [ 711.120103] env[61923]: ERROR nova.compute.manager [instance: 60ffc8b0-4a7b-4e2a-8774-d16e502d8a06] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 711.120103] env[61923]: ERROR nova.compute.manager [instance: 60ffc8b0-4a7b-4e2a-8774-d16e502d8a06] self.driver.spawn(context, instance, image_meta, [ 711.120103] env[61923]: ERROR nova.compute.manager [instance: 60ffc8b0-4a7b-4e2a-8774-d16e502d8a06] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 711.120103] env[61923]: ERROR nova.compute.manager [instance: 60ffc8b0-4a7b-4e2a-8774-d16e502d8a06] self._vmops.spawn(context, instance, image_meta, injected_files, [ 711.120103] env[61923]: ERROR nova.compute.manager [instance: 60ffc8b0-4a7b-4e2a-8774-d16e502d8a06] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 711.120103] env[61923]: ERROR nova.compute.manager [instance: 60ffc8b0-4a7b-4e2a-8774-d16e502d8a06] vm_ref = self.build_virtual_machine(instance, [ 711.120103] env[61923]: ERROR nova.compute.manager [instance: 60ffc8b0-4a7b-4e2a-8774-d16e502d8a06] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 711.120103] env[61923]: ERROR nova.compute.manager [instance: 60ffc8b0-4a7b-4e2a-8774-d16e502d8a06] vif_infos = vmwarevif.get_vif_info(self._session, [ 711.120103] env[61923]: ERROR nova.compute.manager [instance: 60ffc8b0-4a7b-4e2a-8774-d16e502d8a06] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 711.120488] env[61923]: ERROR nova.compute.manager [instance: 60ffc8b0-4a7b-4e2a-8774-d16e502d8a06] for vif in network_info: [ 711.120488] env[61923]: ERROR nova.compute.manager [instance: 60ffc8b0-4a7b-4e2a-8774-d16e502d8a06] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 711.120488] env[61923]: ERROR nova.compute.manager [instance: 60ffc8b0-4a7b-4e2a-8774-d16e502d8a06] return self._sync_wrapper(fn, *args, **kwargs) [ 711.120488] env[61923]: ERROR nova.compute.manager [instance: 60ffc8b0-4a7b-4e2a-8774-d16e502d8a06] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 711.120488] env[61923]: ERROR nova.compute.manager [instance: 60ffc8b0-4a7b-4e2a-8774-d16e502d8a06] self.wait() [ 711.120488] env[61923]: ERROR nova.compute.manager [instance: 60ffc8b0-4a7b-4e2a-8774-d16e502d8a06] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 711.120488] env[61923]: ERROR nova.compute.manager [instance: 60ffc8b0-4a7b-4e2a-8774-d16e502d8a06] self[:] = self._gt.wait() [ 711.120488] env[61923]: ERROR nova.compute.manager [instance: 60ffc8b0-4a7b-4e2a-8774-d16e502d8a06] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 711.120488] env[61923]: ERROR nova.compute.manager [instance: 60ffc8b0-4a7b-4e2a-8774-d16e502d8a06] return self._exit_event.wait() [ 711.120488] env[61923]: ERROR nova.compute.manager [instance: 60ffc8b0-4a7b-4e2a-8774-d16e502d8a06] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 711.120488] env[61923]: ERROR nova.compute.manager [instance: 60ffc8b0-4a7b-4e2a-8774-d16e502d8a06] result = hub.switch() [ 711.120488] env[61923]: ERROR nova.compute.manager [instance: 60ffc8b0-4a7b-4e2a-8774-d16e502d8a06] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 711.120488] env[61923]: ERROR nova.compute.manager [instance: 60ffc8b0-4a7b-4e2a-8774-d16e502d8a06] return self.greenlet.switch() [ 711.120862] env[61923]: ERROR nova.compute.manager [instance: 60ffc8b0-4a7b-4e2a-8774-d16e502d8a06] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 711.120862] env[61923]: ERROR nova.compute.manager [instance: 60ffc8b0-4a7b-4e2a-8774-d16e502d8a06] result = function(*args, **kwargs) [ 711.120862] env[61923]: ERROR nova.compute.manager [instance: 60ffc8b0-4a7b-4e2a-8774-d16e502d8a06] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 711.120862] env[61923]: ERROR nova.compute.manager [instance: 60ffc8b0-4a7b-4e2a-8774-d16e502d8a06] return func(*args, **kwargs) [ 711.120862] env[61923]: ERROR nova.compute.manager [instance: 60ffc8b0-4a7b-4e2a-8774-d16e502d8a06] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 711.120862] env[61923]: ERROR nova.compute.manager [instance: 60ffc8b0-4a7b-4e2a-8774-d16e502d8a06] raise e [ 711.120862] env[61923]: ERROR nova.compute.manager [instance: 60ffc8b0-4a7b-4e2a-8774-d16e502d8a06] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 711.120862] env[61923]: ERROR nova.compute.manager [instance: 60ffc8b0-4a7b-4e2a-8774-d16e502d8a06] nwinfo = self.network_api.allocate_for_instance( [ 711.120862] env[61923]: ERROR nova.compute.manager [instance: 60ffc8b0-4a7b-4e2a-8774-d16e502d8a06] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 711.120862] env[61923]: ERROR nova.compute.manager [instance: 60ffc8b0-4a7b-4e2a-8774-d16e502d8a06] created_port_ids = self._update_ports_for_instance( [ 711.120862] env[61923]: ERROR nova.compute.manager [instance: 60ffc8b0-4a7b-4e2a-8774-d16e502d8a06] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 711.120862] env[61923]: ERROR nova.compute.manager [instance: 60ffc8b0-4a7b-4e2a-8774-d16e502d8a06] with excutils.save_and_reraise_exception(): [ 711.120862] env[61923]: ERROR nova.compute.manager [instance: 60ffc8b0-4a7b-4e2a-8774-d16e502d8a06] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 711.121361] env[61923]: ERROR nova.compute.manager [instance: 60ffc8b0-4a7b-4e2a-8774-d16e502d8a06] self.force_reraise() [ 711.121361] env[61923]: ERROR nova.compute.manager [instance: 60ffc8b0-4a7b-4e2a-8774-d16e502d8a06] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 711.121361] env[61923]: ERROR nova.compute.manager [instance: 60ffc8b0-4a7b-4e2a-8774-d16e502d8a06] raise self.value [ 711.121361] env[61923]: ERROR nova.compute.manager [instance: 60ffc8b0-4a7b-4e2a-8774-d16e502d8a06] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 711.121361] env[61923]: ERROR nova.compute.manager [instance: 60ffc8b0-4a7b-4e2a-8774-d16e502d8a06] updated_port = self._update_port( [ 711.121361] env[61923]: ERROR nova.compute.manager [instance: 60ffc8b0-4a7b-4e2a-8774-d16e502d8a06] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 711.121361] env[61923]: ERROR nova.compute.manager [instance: 60ffc8b0-4a7b-4e2a-8774-d16e502d8a06] _ensure_no_port_binding_failure(port) [ 711.121361] env[61923]: ERROR nova.compute.manager [instance: 60ffc8b0-4a7b-4e2a-8774-d16e502d8a06] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 711.121361] env[61923]: ERROR nova.compute.manager [instance: 60ffc8b0-4a7b-4e2a-8774-d16e502d8a06] raise exception.PortBindingFailed(port_id=port['id']) [ 711.121361] env[61923]: ERROR nova.compute.manager [instance: 60ffc8b0-4a7b-4e2a-8774-d16e502d8a06] nova.exception.PortBindingFailed: Binding failed for port c88b194c-774a-448e-8f31-5a89cfdbc187, please check neutron logs for more information. [ 711.121361] env[61923]: ERROR nova.compute.manager [instance: 60ffc8b0-4a7b-4e2a-8774-d16e502d8a06] [ 711.121926] env[61923]: DEBUG nova.compute.utils [None req-08382130-d70f-42e8-a5b5-93e86f5fa49a tempest-AttachVolumeShelveTestJSON-1252849694 tempest-AttachVolumeShelveTestJSON-1252849694-project-member] [instance: 60ffc8b0-4a7b-4e2a-8774-d16e502d8a06] Binding failed for port c88b194c-774a-448e-8f31-5a89cfdbc187, please check neutron logs for more information. {{(pid=61923) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 711.128142] env[61923]: DEBUG nova.compute.manager [None req-08382130-d70f-42e8-a5b5-93e86f5fa49a tempest-AttachVolumeShelveTestJSON-1252849694 tempest-AttachVolumeShelveTestJSON-1252849694-project-member] [instance: 60ffc8b0-4a7b-4e2a-8774-d16e502d8a06] Build of instance 60ffc8b0-4a7b-4e2a-8774-d16e502d8a06 was re-scheduled: Binding failed for port c88b194c-774a-448e-8f31-5a89cfdbc187, please check neutron logs for more information. {{(pid=61923) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 711.128599] env[61923]: DEBUG nova.compute.manager [None req-08382130-d70f-42e8-a5b5-93e86f5fa49a tempest-AttachVolumeShelveTestJSON-1252849694 tempest-AttachVolumeShelveTestJSON-1252849694-project-member] [instance: 60ffc8b0-4a7b-4e2a-8774-d16e502d8a06] Unplugging VIFs for instance {{(pid=61923) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 711.128831] env[61923]: DEBUG oslo_concurrency.lockutils [None req-08382130-d70f-42e8-a5b5-93e86f5fa49a tempest-AttachVolumeShelveTestJSON-1252849694 tempest-AttachVolumeShelveTestJSON-1252849694-project-member] Acquiring lock "refresh_cache-60ffc8b0-4a7b-4e2a-8774-d16e502d8a06" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 711.129007] env[61923]: DEBUG oslo_concurrency.lockutils [None req-08382130-d70f-42e8-a5b5-93e86f5fa49a tempest-AttachVolumeShelveTestJSON-1252849694 tempest-AttachVolumeShelveTestJSON-1252849694-project-member] Acquired lock "refresh_cache-60ffc8b0-4a7b-4e2a-8774-d16e502d8a06" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 711.129186] env[61923]: DEBUG nova.network.neutron [None req-08382130-d70f-42e8-a5b5-93e86f5fa49a tempest-AttachVolumeShelveTestJSON-1252849694 tempest-AttachVolumeShelveTestJSON-1252849694-project-member] [instance: 60ffc8b0-4a7b-4e2a-8774-d16e502d8a06] Building network info cache for instance {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 711.130230] env[61923]: DEBUG oslo_concurrency.lockutils [None req-6833b601-845b-439d-ad97-e2f9c0927775 tempest-ServerRescueNegativeTestJSON-1435265126 tempest-ServerRescueNegativeTestJSON-1435265126-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.919s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 711.312663] env[61923]: INFO nova.compute.manager [None req-5f08bfba-63c5-4161-be1b-a884c0ae71af tempest-VolumesAdminNegativeTest-1422120048 tempest-VolumesAdminNegativeTest-1422120048-project-member] [instance: 590907a7-b6cc-48fc-a4bf-e4a2e48b05b8] Took 1.02 seconds to deallocate network for instance. [ 711.386201] env[61923]: DEBUG nova.compute.manager [None req-13202777-73e8-40d3-a6ad-7bb49e61267a tempest-ImagesNegativeTestJSON-1202128281 tempest-ImagesNegativeTestJSON-1202128281-project-member] [instance: f7e1d517-3e39-4381-83be-0152368b8120] Starting instance... {{(pid=61923) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 711.666886] env[61923]: DEBUG nova.network.neutron [None req-08382130-d70f-42e8-a5b5-93e86f5fa49a tempest-AttachVolumeShelveTestJSON-1252849694 tempest-AttachVolumeShelveTestJSON-1252849694-project-member] [instance: 60ffc8b0-4a7b-4e2a-8774-d16e502d8a06] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 711.914640] env[61923]: DEBUG oslo_concurrency.lockutils [None req-13202777-73e8-40d3-a6ad-7bb49e61267a tempest-ImagesNegativeTestJSON-1202128281 tempest-ImagesNegativeTestJSON-1202128281-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 712.021483] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e90ce55-65af-47b8-9eb4-b9ac3a1f92a7 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.029412] env[61923]: DEBUG nova.network.neutron [None req-08382130-d70f-42e8-a5b5-93e86f5fa49a tempest-AttachVolumeShelveTestJSON-1252849694 tempest-AttachVolumeShelveTestJSON-1252849694-project-member] [instance: 60ffc8b0-4a7b-4e2a-8774-d16e502d8a06] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 712.034255] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd2aafde-da29-49f7-864a-9b9098075529 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.063269] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be9d027d-6869-4d09-9c17-dbc01835874c {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.072165] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03506dfb-139d-425c-96d3-bed4b508ff9b {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.090372] env[61923]: DEBUG nova.compute.provider_tree [None req-6833b601-845b-439d-ad97-e2f9c0927775 tempest-ServerRescueNegativeTestJSON-1435265126 tempest-ServerRescueNegativeTestJSON-1435265126-project-member] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 712.351793] env[61923]: INFO nova.scheduler.client.report [None req-5f08bfba-63c5-4161-be1b-a884c0ae71af tempest-VolumesAdminNegativeTest-1422120048 tempest-VolumesAdminNegativeTest-1422120048-project-member] Deleted allocations for instance 590907a7-b6cc-48fc-a4bf-e4a2e48b05b8 [ 712.537299] env[61923]: DEBUG oslo_concurrency.lockutils [None req-08382130-d70f-42e8-a5b5-93e86f5fa49a tempest-AttachVolumeShelveTestJSON-1252849694 tempest-AttachVolumeShelveTestJSON-1252849694-project-member] Releasing lock "refresh_cache-60ffc8b0-4a7b-4e2a-8774-d16e502d8a06" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 712.537299] env[61923]: DEBUG nova.compute.manager [None req-08382130-d70f-42e8-a5b5-93e86f5fa49a tempest-AttachVolumeShelveTestJSON-1252849694 tempest-AttachVolumeShelveTestJSON-1252849694-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61923) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 712.537299] env[61923]: DEBUG nova.compute.manager [None req-08382130-d70f-42e8-a5b5-93e86f5fa49a tempest-AttachVolumeShelveTestJSON-1252849694 tempest-AttachVolumeShelveTestJSON-1252849694-project-member] [instance: 60ffc8b0-4a7b-4e2a-8774-d16e502d8a06] Deallocating network for instance {{(pid=61923) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 712.537299] env[61923]: DEBUG nova.network.neutron [None req-08382130-d70f-42e8-a5b5-93e86f5fa49a tempest-AttachVolumeShelveTestJSON-1252849694 tempest-AttachVolumeShelveTestJSON-1252849694-project-member] [instance: 60ffc8b0-4a7b-4e2a-8774-d16e502d8a06] deallocate_for_instance() {{(pid=61923) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 712.566118] env[61923]: DEBUG nova.network.neutron [None req-08382130-d70f-42e8-a5b5-93e86f5fa49a tempest-AttachVolumeShelveTestJSON-1252849694 tempest-AttachVolumeShelveTestJSON-1252849694-project-member] [instance: 60ffc8b0-4a7b-4e2a-8774-d16e502d8a06] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 712.593974] env[61923]: DEBUG nova.scheduler.client.report [None req-6833b601-845b-439d-ad97-e2f9c0927775 tempest-ServerRescueNegativeTestJSON-1435265126 tempest-ServerRescueNegativeTestJSON-1435265126-project-member] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 712.859422] env[61923]: DEBUG oslo_concurrency.lockutils [None req-5f08bfba-63c5-4161-be1b-a884c0ae71af tempest-VolumesAdminNegativeTest-1422120048 tempest-VolumesAdminNegativeTest-1422120048-project-member] Lock "590907a7-b6cc-48fc-a4bf-e4a2e48b05b8" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 134.644s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 713.070330] env[61923]: DEBUG nova.network.neutron [None req-08382130-d70f-42e8-a5b5-93e86f5fa49a tempest-AttachVolumeShelveTestJSON-1252849694 tempest-AttachVolumeShelveTestJSON-1252849694-project-member] [instance: 60ffc8b0-4a7b-4e2a-8774-d16e502d8a06] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 713.103109] env[61923]: DEBUG oslo_concurrency.lockutils [None req-6833b601-845b-439d-ad97-e2f9c0927775 tempest-ServerRescueNegativeTestJSON-1435265126 tempest-ServerRescueNegativeTestJSON-1435265126-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.972s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 713.103216] env[61923]: ERROR nova.compute.manager [None req-6833b601-845b-439d-ad97-e2f9c0927775 tempest-ServerRescueNegativeTestJSON-1435265126 tempest-ServerRescueNegativeTestJSON-1435265126-project-member] [instance: 4f5348ec-f0b6-45e2-be04-31dbf1d49a3a] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port a616fa04-1fe8-475c-9237-a801d7430538, please check neutron logs for more information. [ 713.103216] env[61923]: ERROR nova.compute.manager [instance: 4f5348ec-f0b6-45e2-be04-31dbf1d49a3a] Traceback (most recent call last): [ 713.103216] env[61923]: ERROR nova.compute.manager [instance: 4f5348ec-f0b6-45e2-be04-31dbf1d49a3a] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 713.103216] env[61923]: ERROR nova.compute.manager [instance: 4f5348ec-f0b6-45e2-be04-31dbf1d49a3a] self.driver.spawn(context, instance, image_meta, [ 713.103216] env[61923]: ERROR nova.compute.manager [instance: 4f5348ec-f0b6-45e2-be04-31dbf1d49a3a] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 713.103216] env[61923]: ERROR nova.compute.manager [instance: 4f5348ec-f0b6-45e2-be04-31dbf1d49a3a] self._vmops.spawn(context, instance, image_meta, injected_files, [ 713.103216] env[61923]: ERROR nova.compute.manager [instance: 4f5348ec-f0b6-45e2-be04-31dbf1d49a3a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 713.103216] env[61923]: ERROR nova.compute.manager [instance: 4f5348ec-f0b6-45e2-be04-31dbf1d49a3a] vm_ref = self.build_virtual_machine(instance, [ 713.103216] env[61923]: ERROR nova.compute.manager [instance: 4f5348ec-f0b6-45e2-be04-31dbf1d49a3a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 713.103216] env[61923]: ERROR nova.compute.manager [instance: 4f5348ec-f0b6-45e2-be04-31dbf1d49a3a] vif_infos = vmwarevif.get_vif_info(self._session, [ 713.103216] env[61923]: ERROR nova.compute.manager [instance: 4f5348ec-f0b6-45e2-be04-31dbf1d49a3a] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 713.103998] env[61923]: ERROR nova.compute.manager [instance: 4f5348ec-f0b6-45e2-be04-31dbf1d49a3a] for vif in network_info: [ 713.103998] env[61923]: ERROR nova.compute.manager [instance: 4f5348ec-f0b6-45e2-be04-31dbf1d49a3a] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 713.103998] env[61923]: ERROR nova.compute.manager [instance: 4f5348ec-f0b6-45e2-be04-31dbf1d49a3a] return self._sync_wrapper(fn, *args, **kwargs) [ 713.103998] env[61923]: ERROR nova.compute.manager [instance: 4f5348ec-f0b6-45e2-be04-31dbf1d49a3a] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 713.103998] env[61923]: ERROR nova.compute.manager [instance: 4f5348ec-f0b6-45e2-be04-31dbf1d49a3a] self.wait() [ 713.103998] env[61923]: ERROR nova.compute.manager [instance: 4f5348ec-f0b6-45e2-be04-31dbf1d49a3a] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 713.103998] env[61923]: ERROR nova.compute.manager [instance: 4f5348ec-f0b6-45e2-be04-31dbf1d49a3a] self[:] = self._gt.wait() [ 713.103998] env[61923]: ERROR nova.compute.manager [instance: 4f5348ec-f0b6-45e2-be04-31dbf1d49a3a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 713.103998] env[61923]: ERROR nova.compute.manager [instance: 4f5348ec-f0b6-45e2-be04-31dbf1d49a3a] return self._exit_event.wait() [ 713.103998] env[61923]: ERROR nova.compute.manager [instance: 4f5348ec-f0b6-45e2-be04-31dbf1d49a3a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 713.103998] env[61923]: ERROR nova.compute.manager [instance: 4f5348ec-f0b6-45e2-be04-31dbf1d49a3a] result = hub.switch() [ 713.103998] env[61923]: ERROR nova.compute.manager [instance: 4f5348ec-f0b6-45e2-be04-31dbf1d49a3a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 713.103998] env[61923]: ERROR nova.compute.manager [instance: 4f5348ec-f0b6-45e2-be04-31dbf1d49a3a] return self.greenlet.switch() [ 713.104379] env[61923]: ERROR nova.compute.manager [instance: 4f5348ec-f0b6-45e2-be04-31dbf1d49a3a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 713.104379] env[61923]: ERROR nova.compute.manager [instance: 4f5348ec-f0b6-45e2-be04-31dbf1d49a3a] result = function(*args, **kwargs) [ 713.104379] env[61923]: ERROR nova.compute.manager [instance: 4f5348ec-f0b6-45e2-be04-31dbf1d49a3a] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 713.104379] env[61923]: ERROR nova.compute.manager [instance: 4f5348ec-f0b6-45e2-be04-31dbf1d49a3a] return func(*args, **kwargs) [ 713.104379] env[61923]: ERROR nova.compute.manager [instance: 4f5348ec-f0b6-45e2-be04-31dbf1d49a3a] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 713.104379] env[61923]: ERROR nova.compute.manager [instance: 4f5348ec-f0b6-45e2-be04-31dbf1d49a3a] raise e [ 713.104379] env[61923]: ERROR nova.compute.manager [instance: 4f5348ec-f0b6-45e2-be04-31dbf1d49a3a] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 713.104379] env[61923]: ERROR nova.compute.manager [instance: 4f5348ec-f0b6-45e2-be04-31dbf1d49a3a] nwinfo = self.network_api.allocate_for_instance( [ 713.104379] env[61923]: ERROR nova.compute.manager [instance: 4f5348ec-f0b6-45e2-be04-31dbf1d49a3a] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 713.104379] env[61923]: ERROR nova.compute.manager [instance: 4f5348ec-f0b6-45e2-be04-31dbf1d49a3a] created_port_ids = self._update_ports_for_instance( [ 713.104379] env[61923]: ERROR nova.compute.manager [instance: 4f5348ec-f0b6-45e2-be04-31dbf1d49a3a] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 713.104379] env[61923]: ERROR nova.compute.manager [instance: 4f5348ec-f0b6-45e2-be04-31dbf1d49a3a] with excutils.save_and_reraise_exception(): [ 713.104379] env[61923]: ERROR nova.compute.manager [instance: 4f5348ec-f0b6-45e2-be04-31dbf1d49a3a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 713.104752] env[61923]: ERROR nova.compute.manager [instance: 4f5348ec-f0b6-45e2-be04-31dbf1d49a3a] self.force_reraise() [ 713.104752] env[61923]: ERROR nova.compute.manager [instance: 4f5348ec-f0b6-45e2-be04-31dbf1d49a3a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 713.104752] env[61923]: ERROR nova.compute.manager [instance: 4f5348ec-f0b6-45e2-be04-31dbf1d49a3a] raise self.value [ 713.104752] env[61923]: ERROR nova.compute.manager [instance: 4f5348ec-f0b6-45e2-be04-31dbf1d49a3a] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 713.104752] env[61923]: ERROR nova.compute.manager [instance: 4f5348ec-f0b6-45e2-be04-31dbf1d49a3a] updated_port = self._update_port( [ 713.104752] env[61923]: ERROR nova.compute.manager [instance: 4f5348ec-f0b6-45e2-be04-31dbf1d49a3a] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 713.104752] env[61923]: ERROR nova.compute.manager [instance: 4f5348ec-f0b6-45e2-be04-31dbf1d49a3a] _ensure_no_port_binding_failure(port) [ 713.104752] env[61923]: ERROR nova.compute.manager [instance: 4f5348ec-f0b6-45e2-be04-31dbf1d49a3a] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 713.104752] env[61923]: ERROR nova.compute.manager [instance: 4f5348ec-f0b6-45e2-be04-31dbf1d49a3a] raise exception.PortBindingFailed(port_id=port['id']) [ 713.104752] env[61923]: ERROR nova.compute.manager [instance: 4f5348ec-f0b6-45e2-be04-31dbf1d49a3a] nova.exception.PortBindingFailed: Binding failed for port a616fa04-1fe8-475c-9237-a801d7430538, please check neutron logs for more information. [ 713.104752] env[61923]: ERROR nova.compute.manager [instance: 4f5348ec-f0b6-45e2-be04-31dbf1d49a3a] [ 713.105097] env[61923]: DEBUG nova.compute.utils [None req-6833b601-845b-439d-ad97-e2f9c0927775 tempest-ServerRescueNegativeTestJSON-1435265126 tempest-ServerRescueNegativeTestJSON-1435265126-project-member] [instance: 4f5348ec-f0b6-45e2-be04-31dbf1d49a3a] Binding failed for port a616fa04-1fe8-475c-9237-a801d7430538, please check neutron logs for more information. {{(pid=61923) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 713.105351] env[61923]: DEBUG oslo_concurrency.lockutils [None req-425a3606-915c-4f9b-8507-50f1c8d0d8d9 tempest-ServerRescueTestJSON-521767381 tempest-ServerRescueTestJSON-521767381-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.867s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 713.107200] env[61923]: INFO nova.compute.claims [None req-425a3606-915c-4f9b-8507-50f1c8d0d8d9 tempest-ServerRescueTestJSON-521767381 tempest-ServerRescueTestJSON-521767381-project-member] [instance: 74b32299-983f-41f7-b3f4-3a12815a83e6] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 713.109949] env[61923]: DEBUG nova.compute.manager [None req-6833b601-845b-439d-ad97-e2f9c0927775 tempest-ServerRescueNegativeTestJSON-1435265126 tempest-ServerRescueNegativeTestJSON-1435265126-project-member] [instance: 4f5348ec-f0b6-45e2-be04-31dbf1d49a3a] Build of instance 4f5348ec-f0b6-45e2-be04-31dbf1d49a3a was re-scheduled: Binding failed for port a616fa04-1fe8-475c-9237-a801d7430538, please check neutron logs for more information. {{(pid=61923) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 713.110438] env[61923]: DEBUG nova.compute.manager [None req-6833b601-845b-439d-ad97-e2f9c0927775 tempest-ServerRescueNegativeTestJSON-1435265126 tempest-ServerRescueNegativeTestJSON-1435265126-project-member] [instance: 4f5348ec-f0b6-45e2-be04-31dbf1d49a3a] Unplugging VIFs for instance {{(pid=61923) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 713.110661] env[61923]: DEBUG oslo_concurrency.lockutils [None req-6833b601-845b-439d-ad97-e2f9c0927775 tempest-ServerRescueNegativeTestJSON-1435265126 tempest-ServerRescueNegativeTestJSON-1435265126-project-member] Acquiring lock "refresh_cache-4f5348ec-f0b6-45e2-be04-31dbf1d49a3a" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 713.110805] env[61923]: DEBUG oslo_concurrency.lockutils [None req-6833b601-845b-439d-ad97-e2f9c0927775 tempest-ServerRescueNegativeTestJSON-1435265126 tempest-ServerRescueNegativeTestJSON-1435265126-project-member] Acquired lock "refresh_cache-4f5348ec-f0b6-45e2-be04-31dbf1d49a3a" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 713.110958] env[61923]: DEBUG nova.network.neutron [None req-6833b601-845b-439d-ad97-e2f9c0927775 tempest-ServerRescueNegativeTestJSON-1435265126 tempest-ServerRescueNegativeTestJSON-1435265126-project-member] [instance: 4f5348ec-f0b6-45e2-be04-31dbf1d49a3a] Building network info cache for instance {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 713.362537] env[61923]: DEBUG nova.compute.manager [None req-f7b52f4c-b2e3-49f8-b056-b7254f72bb20 tempest-AttachInterfacesUnderV243Test-498638506 tempest-AttachInterfacesUnderV243Test-498638506-project-member] [instance: 0eab2200-2528-4a89-821a-cdd6a3c99297] Starting instance... {{(pid=61923) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 713.573464] env[61923]: INFO nova.compute.manager [None req-08382130-d70f-42e8-a5b5-93e86f5fa49a tempest-AttachVolumeShelveTestJSON-1252849694 tempest-AttachVolumeShelveTestJSON-1252849694-project-member] [instance: 60ffc8b0-4a7b-4e2a-8774-d16e502d8a06] Took 1.04 seconds to deallocate network for instance. [ 713.635465] env[61923]: DEBUG nova.network.neutron [None req-6833b601-845b-439d-ad97-e2f9c0927775 tempest-ServerRescueNegativeTestJSON-1435265126 tempest-ServerRescueNegativeTestJSON-1435265126-project-member] [instance: 4f5348ec-f0b6-45e2-be04-31dbf1d49a3a] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 713.792471] env[61923]: DEBUG nova.network.neutron [None req-6833b601-845b-439d-ad97-e2f9c0927775 tempest-ServerRescueNegativeTestJSON-1435265126 tempest-ServerRescueNegativeTestJSON-1435265126-project-member] [instance: 4f5348ec-f0b6-45e2-be04-31dbf1d49a3a] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 713.888192] env[61923]: DEBUG oslo_concurrency.lockutils [None req-f7b52f4c-b2e3-49f8-b056-b7254f72bb20 tempest-AttachInterfacesUnderV243Test-498638506 tempest-AttachInterfacesUnderV243Test-498638506-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 714.295376] env[61923]: DEBUG oslo_concurrency.lockutils [None req-6833b601-845b-439d-ad97-e2f9c0927775 tempest-ServerRescueNegativeTestJSON-1435265126 tempest-ServerRescueNegativeTestJSON-1435265126-project-member] Releasing lock "refresh_cache-4f5348ec-f0b6-45e2-be04-31dbf1d49a3a" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 714.295604] env[61923]: DEBUG nova.compute.manager [None req-6833b601-845b-439d-ad97-e2f9c0927775 tempest-ServerRescueNegativeTestJSON-1435265126 tempest-ServerRescueNegativeTestJSON-1435265126-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61923) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 714.295781] env[61923]: DEBUG nova.compute.manager [None req-6833b601-845b-439d-ad97-e2f9c0927775 tempest-ServerRescueNegativeTestJSON-1435265126 tempest-ServerRescueNegativeTestJSON-1435265126-project-member] [instance: 4f5348ec-f0b6-45e2-be04-31dbf1d49a3a] Deallocating network for instance {{(pid=61923) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 714.295938] env[61923]: DEBUG nova.network.neutron [None req-6833b601-845b-439d-ad97-e2f9c0927775 tempest-ServerRescueNegativeTestJSON-1435265126 tempest-ServerRescueNegativeTestJSON-1435265126-project-member] [instance: 4f5348ec-f0b6-45e2-be04-31dbf1d49a3a] deallocate_for_instance() {{(pid=61923) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 714.319140] env[61923]: DEBUG nova.network.neutron [None req-6833b601-845b-439d-ad97-e2f9c0927775 tempest-ServerRescueNegativeTestJSON-1435265126 tempest-ServerRescueNegativeTestJSON-1435265126-project-member] [instance: 4f5348ec-f0b6-45e2-be04-31dbf1d49a3a] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 714.471413] env[61923]: DEBUG oslo_concurrency.lockutils [None req-0adbd41a-bb08-4b5a-bc10-ab9778329dd8 tempest-VolumesAdminNegativeTest-1422120048 tempest-VolumesAdminNegativeTest-1422120048-project-member] Acquiring lock "7bd3204b-5754-4ff2-b0ff-f30556dd72a1" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 714.471823] env[61923]: DEBUG oslo_concurrency.lockutils [None req-0adbd41a-bb08-4b5a-bc10-ab9778329dd8 tempest-VolumesAdminNegativeTest-1422120048 tempest-VolumesAdminNegativeTest-1422120048-project-member] Lock "7bd3204b-5754-4ff2-b0ff-f30556dd72a1" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 714.477901] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-752e2508-f6df-476d-851a-5ace18c3c7c7 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.491611] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-160a76c5-a3fe-4c54-8f45-367d65a52a25 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.524036] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7e1d825-3706-4575-9929-7235419c5f19 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.531560] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e31d8e3-50b3-426a-a2d9-415ff5e627c6 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 714.545939] env[61923]: DEBUG nova.compute.provider_tree [None req-425a3606-915c-4f9b-8507-50f1c8d0d8d9 tempest-ServerRescueTestJSON-521767381 tempest-ServerRescueTestJSON-521767381-project-member] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 714.603111] env[61923]: INFO nova.scheduler.client.report [None req-08382130-d70f-42e8-a5b5-93e86f5fa49a tempest-AttachVolumeShelveTestJSON-1252849694 tempest-AttachVolumeShelveTestJSON-1252849694-project-member] Deleted allocations for instance 60ffc8b0-4a7b-4e2a-8774-d16e502d8a06 [ 714.822586] env[61923]: DEBUG nova.network.neutron [None req-6833b601-845b-439d-ad97-e2f9c0927775 tempest-ServerRescueNegativeTestJSON-1435265126 tempest-ServerRescueNegativeTestJSON-1435265126-project-member] [instance: 4f5348ec-f0b6-45e2-be04-31dbf1d49a3a] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 715.049637] env[61923]: DEBUG nova.scheduler.client.report [None req-425a3606-915c-4f9b-8507-50f1c8d0d8d9 tempest-ServerRescueTestJSON-521767381 tempest-ServerRescueTestJSON-521767381-project-member] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 715.112191] env[61923]: DEBUG oslo_concurrency.lockutils [None req-08382130-d70f-42e8-a5b5-93e86f5fa49a tempest-AttachVolumeShelveTestJSON-1252849694 tempest-AttachVolumeShelveTestJSON-1252849694-project-member] Lock "60ffc8b0-4a7b-4e2a-8774-d16e502d8a06" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 136.076s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 715.328100] env[61923]: INFO nova.compute.manager [None req-6833b601-845b-439d-ad97-e2f9c0927775 tempest-ServerRescueNegativeTestJSON-1435265126 tempest-ServerRescueNegativeTestJSON-1435265126-project-member] [instance: 4f5348ec-f0b6-45e2-be04-31dbf1d49a3a] Took 1.03 seconds to deallocate network for instance. [ 715.554796] env[61923]: DEBUG oslo_concurrency.lockutils [None req-425a3606-915c-4f9b-8507-50f1c8d0d8d9 tempest-ServerRescueTestJSON-521767381 tempest-ServerRescueTestJSON-521767381-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.449s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 715.555278] env[61923]: DEBUG nova.compute.manager [None req-425a3606-915c-4f9b-8507-50f1c8d0d8d9 tempest-ServerRescueTestJSON-521767381 tempest-ServerRescueTestJSON-521767381-project-member] [instance: 74b32299-983f-41f7-b3f4-3a12815a83e6] Start building networks asynchronously for instance. {{(pid=61923) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 715.559032] env[61923]: DEBUG oslo_concurrency.lockutils [None req-df4eb7d8-5b64-4db4-b042-683b16e0b991 tempest-ServersTestBootFromVolume-2104901628 tempest-ServersTestBootFromVolume-2104901628-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 12.811s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 715.616581] env[61923]: DEBUG nova.compute.manager [None req-4a9f7dc2-7221-4911-9bc8-b39b7a22ddda tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] [instance: 22d1e8ed-75c5-42e9-ae55-12d59e565914] Starting instance... {{(pid=61923) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 716.063280] env[61923]: DEBUG nova.compute.utils [None req-425a3606-915c-4f9b-8507-50f1c8d0d8d9 tempest-ServerRescueTestJSON-521767381 tempest-ServerRescueTestJSON-521767381-project-member] Using /dev/sd instead of None {{(pid=61923) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 716.068862] env[61923]: DEBUG nova.compute.manager [None req-425a3606-915c-4f9b-8507-50f1c8d0d8d9 tempest-ServerRescueTestJSON-521767381 tempest-ServerRescueTestJSON-521767381-project-member] [instance: 74b32299-983f-41f7-b3f4-3a12815a83e6] Allocating IP information in the background. {{(pid=61923) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 716.069276] env[61923]: DEBUG nova.network.neutron [None req-425a3606-915c-4f9b-8507-50f1c8d0d8d9 tempest-ServerRescueTestJSON-521767381 tempest-ServerRescueTestJSON-521767381-project-member] [instance: 74b32299-983f-41f7-b3f4-3a12815a83e6] allocate_for_instance() {{(pid=61923) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 716.129630] env[61923]: DEBUG nova.policy [None req-425a3606-915c-4f9b-8507-50f1c8d0d8d9 tempest-ServerRescueTestJSON-521767381 tempest-ServerRescueTestJSON-521767381-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '794e975f0c2646e08ba0d579ea0f3eb6', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f6623f952ddc4fa18da81ea3b1cd84da', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61923) authorize /opt/stack/nova/nova/policy.py:201}} [ 716.144982] env[61923]: DEBUG oslo_concurrency.lockutils [None req-4a9f7dc2-7221-4911-9bc8-b39b7a22ddda tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 716.364303] env[61923]: INFO nova.scheduler.client.report [None req-6833b601-845b-439d-ad97-e2f9c0927775 tempest-ServerRescueNegativeTestJSON-1435265126 tempest-ServerRescueNegativeTestJSON-1435265126-project-member] Deleted allocations for instance 4f5348ec-f0b6-45e2-be04-31dbf1d49a3a [ 716.464876] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aee6d938-eb8c-4447-9180-76c46ee342bd {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.472878] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c71e168f-5140-458a-b7ab-7e779cd0e5fd {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.509674] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f399c64e-d349-43cb-a76c-d887d5ddb661 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.517276] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36024db3-bae7-4798-b306-82ce46db76b3 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.531143] env[61923]: DEBUG nova.compute.provider_tree [None req-df4eb7d8-5b64-4db4-b042-683b16e0b991 tempest-ServersTestBootFromVolume-2104901628 tempest-ServersTestBootFromVolume-2104901628-project-member] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 716.536960] env[61923]: DEBUG nova.network.neutron [None req-425a3606-915c-4f9b-8507-50f1c8d0d8d9 tempest-ServerRescueTestJSON-521767381 tempest-ServerRescueTestJSON-521767381-project-member] [instance: 74b32299-983f-41f7-b3f4-3a12815a83e6] Successfully created port: 8065e5e4-32c9-4969-8aff-a6bc9953f4bc {{(pid=61923) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 716.569983] env[61923]: DEBUG nova.compute.manager [None req-425a3606-915c-4f9b-8507-50f1c8d0d8d9 tempest-ServerRescueTestJSON-521767381 tempest-ServerRescueTestJSON-521767381-project-member] [instance: 74b32299-983f-41f7-b3f4-3a12815a83e6] Start building block device mappings for instance. {{(pid=61923) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 716.770551] env[61923]: DEBUG oslo_concurrency.lockutils [None req-28282894-8344-49f7-8696-a5bcafc57c69 tempest-ServerRescueTestJSONUnderV235-1784422495 tempest-ServerRescueTestJSONUnderV235-1784422495-project-member] Acquiring lock "b2d2b34d-827f-4f96-af8e-936135ba8700" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 716.771160] env[61923]: DEBUG oslo_concurrency.lockutils [None req-28282894-8344-49f7-8696-a5bcafc57c69 tempest-ServerRescueTestJSONUnderV235-1784422495 tempest-ServerRescueTestJSONUnderV235-1784422495-project-member] Lock "b2d2b34d-827f-4f96-af8e-936135ba8700" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 716.880089] env[61923]: DEBUG oslo_concurrency.lockutils [None req-6833b601-845b-439d-ad97-e2f9c0927775 tempest-ServerRescueNegativeTestJSON-1435265126 tempest-ServerRescueNegativeTestJSON-1435265126-project-member] Lock "4f5348ec-f0b6-45e2-be04-31dbf1d49a3a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 137.157s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 717.035217] env[61923]: DEBUG nova.scheduler.client.report [None req-df4eb7d8-5b64-4db4-b042-683b16e0b991 tempest-ServersTestBootFromVolume-2104901628 tempest-ServersTestBootFromVolume-2104901628-project-member] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 717.381831] env[61923]: DEBUG nova.compute.manager [None req-5bb1ca17-5442-4564-86ab-53f866a97e1d tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: 15eaaa7c-77be-4b8b-89d0-0b2b11a0c49e] Starting instance... {{(pid=61923) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 717.543460] env[61923]: DEBUG oslo_concurrency.lockutils [None req-df4eb7d8-5b64-4db4-b042-683b16e0b991 tempest-ServersTestBootFromVolume-2104901628 tempest-ServersTestBootFromVolume-2104901628-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.985s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 717.544100] env[61923]: ERROR nova.compute.manager [None req-df4eb7d8-5b64-4db4-b042-683b16e0b991 tempest-ServersTestBootFromVolume-2104901628 tempest-ServersTestBootFromVolume-2104901628-project-member] [instance: fb548893-2df5-4629-bf3a-3b0f807f6357] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 02c3270e-c070-4203-bf7e-d9f1db17ad7e, please check neutron logs for more information. [ 717.544100] env[61923]: ERROR nova.compute.manager [instance: fb548893-2df5-4629-bf3a-3b0f807f6357] Traceback (most recent call last): [ 717.544100] env[61923]: ERROR nova.compute.manager [instance: fb548893-2df5-4629-bf3a-3b0f807f6357] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 717.544100] env[61923]: ERROR nova.compute.manager [instance: fb548893-2df5-4629-bf3a-3b0f807f6357] self.driver.spawn(context, instance, image_meta, [ 717.544100] env[61923]: ERROR nova.compute.manager [instance: fb548893-2df5-4629-bf3a-3b0f807f6357] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 717.544100] env[61923]: ERROR nova.compute.manager [instance: fb548893-2df5-4629-bf3a-3b0f807f6357] self._vmops.spawn(context, instance, image_meta, injected_files, [ 717.544100] env[61923]: ERROR nova.compute.manager [instance: fb548893-2df5-4629-bf3a-3b0f807f6357] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 717.544100] env[61923]: ERROR nova.compute.manager [instance: fb548893-2df5-4629-bf3a-3b0f807f6357] vm_ref = self.build_virtual_machine(instance, [ 717.544100] env[61923]: ERROR nova.compute.manager [instance: fb548893-2df5-4629-bf3a-3b0f807f6357] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 717.544100] env[61923]: ERROR nova.compute.manager [instance: fb548893-2df5-4629-bf3a-3b0f807f6357] vif_infos = vmwarevif.get_vif_info(self._session, [ 717.544100] env[61923]: ERROR nova.compute.manager [instance: fb548893-2df5-4629-bf3a-3b0f807f6357] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 717.544607] env[61923]: ERROR nova.compute.manager [instance: fb548893-2df5-4629-bf3a-3b0f807f6357] for vif in network_info: [ 717.544607] env[61923]: ERROR nova.compute.manager [instance: fb548893-2df5-4629-bf3a-3b0f807f6357] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 717.544607] env[61923]: ERROR nova.compute.manager [instance: fb548893-2df5-4629-bf3a-3b0f807f6357] return self._sync_wrapper(fn, *args, **kwargs) [ 717.544607] env[61923]: ERROR nova.compute.manager [instance: fb548893-2df5-4629-bf3a-3b0f807f6357] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 717.544607] env[61923]: ERROR nova.compute.manager [instance: fb548893-2df5-4629-bf3a-3b0f807f6357] self.wait() [ 717.544607] env[61923]: ERROR nova.compute.manager [instance: fb548893-2df5-4629-bf3a-3b0f807f6357] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 717.544607] env[61923]: ERROR nova.compute.manager [instance: fb548893-2df5-4629-bf3a-3b0f807f6357] self[:] = self._gt.wait() [ 717.544607] env[61923]: ERROR nova.compute.manager [instance: fb548893-2df5-4629-bf3a-3b0f807f6357] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 717.544607] env[61923]: ERROR nova.compute.manager [instance: fb548893-2df5-4629-bf3a-3b0f807f6357] return self._exit_event.wait() [ 717.544607] env[61923]: ERROR nova.compute.manager [instance: fb548893-2df5-4629-bf3a-3b0f807f6357] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 717.544607] env[61923]: ERROR nova.compute.manager [instance: fb548893-2df5-4629-bf3a-3b0f807f6357] current.throw(*self._exc) [ 717.544607] env[61923]: ERROR nova.compute.manager [instance: fb548893-2df5-4629-bf3a-3b0f807f6357] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 717.544607] env[61923]: ERROR nova.compute.manager [instance: fb548893-2df5-4629-bf3a-3b0f807f6357] result = function(*args, **kwargs) [ 717.545703] env[61923]: ERROR nova.compute.manager [instance: fb548893-2df5-4629-bf3a-3b0f807f6357] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 717.545703] env[61923]: ERROR nova.compute.manager [instance: fb548893-2df5-4629-bf3a-3b0f807f6357] return func(*args, **kwargs) [ 717.545703] env[61923]: ERROR nova.compute.manager [instance: fb548893-2df5-4629-bf3a-3b0f807f6357] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 717.545703] env[61923]: ERROR nova.compute.manager [instance: fb548893-2df5-4629-bf3a-3b0f807f6357] raise e [ 717.545703] env[61923]: ERROR nova.compute.manager [instance: fb548893-2df5-4629-bf3a-3b0f807f6357] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 717.545703] env[61923]: ERROR nova.compute.manager [instance: fb548893-2df5-4629-bf3a-3b0f807f6357] nwinfo = self.network_api.allocate_for_instance( [ 717.545703] env[61923]: ERROR nova.compute.manager [instance: fb548893-2df5-4629-bf3a-3b0f807f6357] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 717.545703] env[61923]: ERROR nova.compute.manager [instance: fb548893-2df5-4629-bf3a-3b0f807f6357] created_port_ids = self._update_ports_for_instance( [ 717.545703] env[61923]: ERROR nova.compute.manager [instance: fb548893-2df5-4629-bf3a-3b0f807f6357] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 717.545703] env[61923]: ERROR nova.compute.manager [instance: fb548893-2df5-4629-bf3a-3b0f807f6357] with excutils.save_and_reraise_exception(): [ 717.545703] env[61923]: ERROR nova.compute.manager [instance: fb548893-2df5-4629-bf3a-3b0f807f6357] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 717.545703] env[61923]: ERROR nova.compute.manager [instance: fb548893-2df5-4629-bf3a-3b0f807f6357] self.force_reraise() [ 717.545703] env[61923]: ERROR nova.compute.manager [instance: fb548893-2df5-4629-bf3a-3b0f807f6357] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 717.546213] env[61923]: ERROR nova.compute.manager [instance: fb548893-2df5-4629-bf3a-3b0f807f6357] raise self.value [ 717.546213] env[61923]: ERROR nova.compute.manager [instance: fb548893-2df5-4629-bf3a-3b0f807f6357] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 717.546213] env[61923]: ERROR nova.compute.manager [instance: fb548893-2df5-4629-bf3a-3b0f807f6357] updated_port = self._update_port( [ 717.546213] env[61923]: ERROR nova.compute.manager [instance: fb548893-2df5-4629-bf3a-3b0f807f6357] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 717.546213] env[61923]: ERROR nova.compute.manager [instance: fb548893-2df5-4629-bf3a-3b0f807f6357] _ensure_no_port_binding_failure(port) [ 717.546213] env[61923]: ERROR nova.compute.manager [instance: fb548893-2df5-4629-bf3a-3b0f807f6357] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 717.546213] env[61923]: ERROR nova.compute.manager [instance: fb548893-2df5-4629-bf3a-3b0f807f6357] raise exception.PortBindingFailed(port_id=port['id']) [ 717.546213] env[61923]: ERROR nova.compute.manager [instance: fb548893-2df5-4629-bf3a-3b0f807f6357] nova.exception.PortBindingFailed: Binding failed for port 02c3270e-c070-4203-bf7e-d9f1db17ad7e, please check neutron logs for more information. [ 717.546213] env[61923]: ERROR nova.compute.manager [instance: fb548893-2df5-4629-bf3a-3b0f807f6357] [ 717.546213] env[61923]: DEBUG nova.compute.utils [None req-df4eb7d8-5b64-4db4-b042-683b16e0b991 tempest-ServersTestBootFromVolume-2104901628 tempest-ServersTestBootFromVolume-2104901628-project-member] [instance: fb548893-2df5-4629-bf3a-3b0f807f6357] Binding failed for port 02c3270e-c070-4203-bf7e-d9f1db17ad7e, please check neutron logs for more information. {{(pid=61923) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 717.546500] env[61923]: DEBUG oslo_concurrency.lockutils [None req-44a76586-0f76-448f-af41-a5d846bfda81 tempest-ServerActionsV293TestJSON-849444864 tempest-ServerActionsV293TestJSON-849444864-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 12.479s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 717.550638] env[61923]: DEBUG nova.compute.manager [None req-df4eb7d8-5b64-4db4-b042-683b16e0b991 tempest-ServersTestBootFromVolume-2104901628 tempest-ServersTestBootFromVolume-2104901628-project-member] [instance: fb548893-2df5-4629-bf3a-3b0f807f6357] Build of instance fb548893-2df5-4629-bf3a-3b0f807f6357 was re-scheduled: Binding failed for port 02c3270e-c070-4203-bf7e-d9f1db17ad7e, please check neutron logs for more information. {{(pid=61923) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 717.550638] env[61923]: DEBUG nova.compute.manager [None req-df4eb7d8-5b64-4db4-b042-683b16e0b991 tempest-ServersTestBootFromVolume-2104901628 tempest-ServersTestBootFromVolume-2104901628-project-member] [instance: fb548893-2df5-4629-bf3a-3b0f807f6357] Unplugging VIFs for instance {{(pid=61923) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 717.550638] env[61923]: DEBUG oslo_concurrency.lockutils [None req-df4eb7d8-5b64-4db4-b042-683b16e0b991 tempest-ServersTestBootFromVolume-2104901628 tempest-ServersTestBootFromVolume-2104901628-project-member] Acquiring lock "refresh_cache-fb548893-2df5-4629-bf3a-3b0f807f6357" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 717.550638] env[61923]: DEBUG oslo_concurrency.lockutils [None req-df4eb7d8-5b64-4db4-b042-683b16e0b991 tempest-ServersTestBootFromVolume-2104901628 tempest-ServersTestBootFromVolume-2104901628-project-member] Acquired lock "refresh_cache-fb548893-2df5-4629-bf3a-3b0f807f6357" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 717.550821] env[61923]: DEBUG nova.network.neutron [None req-df4eb7d8-5b64-4db4-b042-683b16e0b991 tempest-ServersTestBootFromVolume-2104901628 tempest-ServersTestBootFromVolume-2104901628-project-member] [instance: fb548893-2df5-4629-bf3a-3b0f807f6357] Building network info cache for instance {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 717.578017] env[61923]: DEBUG nova.compute.manager [None req-425a3606-915c-4f9b-8507-50f1c8d0d8d9 tempest-ServerRescueTestJSON-521767381 tempest-ServerRescueTestJSON-521767381-project-member] [instance: 74b32299-983f-41f7-b3f4-3a12815a83e6] Start spawning the instance on the hypervisor. {{(pid=61923) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 717.602326] env[61923]: DEBUG nova.virt.hardware [None req-425a3606-915c-4f9b-8507-50f1c8d0d8d9 tempest-ServerRescueTestJSON-521767381 tempest-ServerRescueTestJSON-521767381-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-29T20:07:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-29T20:07:35Z,direct_url=,disk_format='vmdk',id=0f153f63-ae0a-45b1-b7f5-7f9b673c947f,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='4e7b5e3b3c3443c8bd5c70f8a15739f5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-29T20:07:36Z,virtual_size=,visibility=), allow threads: False {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 717.602579] env[61923]: DEBUG nova.virt.hardware [None req-425a3606-915c-4f9b-8507-50f1c8d0d8d9 tempest-ServerRescueTestJSON-521767381 tempest-ServerRescueTestJSON-521767381-project-member] Flavor limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 717.602731] env[61923]: DEBUG nova.virt.hardware [None req-425a3606-915c-4f9b-8507-50f1c8d0d8d9 tempest-ServerRescueTestJSON-521767381 tempest-ServerRescueTestJSON-521767381-project-member] Image limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 717.602905] env[61923]: DEBUG nova.virt.hardware [None req-425a3606-915c-4f9b-8507-50f1c8d0d8d9 tempest-ServerRescueTestJSON-521767381 tempest-ServerRescueTestJSON-521767381-project-member] Flavor pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 717.603090] env[61923]: DEBUG nova.virt.hardware [None req-425a3606-915c-4f9b-8507-50f1c8d0d8d9 tempest-ServerRescueTestJSON-521767381 tempest-ServerRescueTestJSON-521767381-project-member] Image pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 717.603324] env[61923]: DEBUG nova.virt.hardware [None req-425a3606-915c-4f9b-8507-50f1c8d0d8d9 tempest-ServerRescueTestJSON-521767381 tempest-ServerRescueTestJSON-521767381-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 717.603552] env[61923]: DEBUG nova.virt.hardware [None req-425a3606-915c-4f9b-8507-50f1c8d0d8d9 tempest-ServerRescueTestJSON-521767381 tempest-ServerRescueTestJSON-521767381-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 717.603710] env[61923]: DEBUG nova.virt.hardware [None req-425a3606-915c-4f9b-8507-50f1c8d0d8d9 tempest-ServerRescueTestJSON-521767381 tempest-ServerRescueTestJSON-521767381-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 717.603873] env[61923]: DEBUG nova.virt.hardware [None req-425a3606-915c-4f9b-8507-50f1c8d0d8d9 tempest-ServerRescueTestJSON-521767381 tempest-ServerRescueTestJSON-521767381-project-member] Got 1 possible topologies {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 717.604060] env[61923]: DEBUG nova.virt.hardware [None req-425a3606-915c-4f9b-8507-50f1c8d0d8d9 tempest-ServerRescueTestJSON-521767381 tempest-ServerRescueTestJSON-521767381-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 717.604245] env[61923]: DEBUG nova.virt.hardware [None req-425a3606-915c-4f9b-8507-50f1c8d0d8d9 tempest-ServerRescueTestJSON-521767381 tempest-ServerRescueTestJSON-521767381-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 717.605382] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d482b404-19c5-4e10-aa83-b11f09ceab8c {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.613624] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eebccbdd-3328-4b88-b11b-c4772c483dbf {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.903893] env[61923]: DEBUG oslo_concurrency.lockutils [None req-5bb1ca17-5442-4564-86ab-53f866a97e1d tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 717.972175] env[61923]: DEBUG nova.compute.manager [req-b35eecb4-7346-406a-969d-8e6844be522f req-c971b888-2935-4c09-9ff3-9f92debad42e service nova] [instance: 74b32299-983f-41f7-b3f4-3a12815a83e6] Received event network-changed-8065e5e4-32c9-4969-8aff-a6bc9953f4bc {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 717.972379] env[61923]: DEBUG nova.compute.manager [req-b35eecb4-7346-406a-969d-8e6844be522f req-c971b888-2935-4c09-9ff3-9f92debad42e service nova] [instance: 74b32299-983f-41f7-b3f4-3a12815a83e6] Refreshing instance network info cache due to event network-changed-8065e5e4-32c9-4969-8aff-a6bc9953f4bc. {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 717.972587] env[61923]: DEBUG oslo_concurrency.lockutils [req-b35eecb4-7346-406a-969d-8e6844be522f req-c971b888-2935-4c09-9ff3-9f92debad42e service nova] Acquiring lock "refresh_cache-74b32299-983f-41f7-b3f4-3a12815a83e6" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 717.972727] env[61923]: DEBUG oslo_concurrency.lockutils [req-b35eecb4-7346-406a-969d-8e6844be522f req-c971b888-2935-4c09-9ff3-9f92debad42e service nova] Acquired lock "refresh_cache-74b32299-983f-41f7-b3f4-3a12815a83e6" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 717.972882] env[61923]: DEBUG nova.network.neutron [req-b35eecb4-7346-406a-969d-8e6844be522f req-c971b888-2935-4c09-9ff3-9f92debad42e service nova] [instance: 74b32299-983f-41f7-b3f4-3a12815a83e6] Refreshing network info cache for port 8065e5e4-32c9-4969-8aff-a6bc9953f4bc {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 718.020971] env[61923]: ERROR nova.compute.manager [None req-425a3606-915c-4f9b-8507-50f1c8d0d8d9 tempest-ServerRescueTestJSON-521767381 tempest-ServerRescueTestJSON-521767381-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 8065e5e4-32c9-4969-8aff-a6bc9953f4bc, please check neutron logs for more information. [ 718.020971] env[61923]: ERROR nova.compute.manager Traceback (most recent call last): [ 718.020971] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 718.020971] env[61923]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 718.020971] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 718.020971] env[61923]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 718.020971] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 718.020971] env[61923]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 718.020971] env[61923]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 718.020971] env[61923]: ERROR nova.compute.manager self.force_reraise() [ 718.020971] env[61923]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 718.020971] env[61923]: ERROR nova.compute.manager raise self.value [ 718.020971] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 718.020971] env[61923]: ERROR nova.compute.manager updated_port = self._update_port( [ 718.020971] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 718.020971] env[61923]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 718.021506] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 718.021506] env[61923]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 718.021506] env[61923]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 8065e5e4-32c9-4969-8aff-a6bc9953f4bc, please check neutron logs for more information. [ 718.021506] env[61923]: ERROR nova.compute.manager [ 718.021506] env[61923]: Traceback (most recent call last): [ 718.021506] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 718.021506] env[61923]: listener.cb(fileno) [ 718.021506] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 718.021506] env[61923]: result = function(*args, **kwargs) [ 718.021506] env[61923]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 718.021506] env[61923]: return func(*args, **kwargs) [ 718.021506] env[61923]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 718.021506] env[61923]: raise e [ 718.021506] env[61923]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 718.021506] env[61923]: nwinfo = self.network_api.allocate_for_instance( [ 718.021506] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 718.021506] env[61923]: created_port_ids = self._update_ports_for_instance( [ 718.021506] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 718.021506] env[61923]: with excutils.save_and_reraise_exception(): [ 718.021506] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 718.021506] env[61923]: self.force_reraise() [ 718.021506] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 718.021506] env[61923]: raise self.value [ 718.021506] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 718.021506] env[61923]: updated_port = self._update_port( [ 718.021506] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 718.021506] env[61923]: _ensure_no_port_binding_failure(port) [ 718.021506] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 718.021506] env[61923]: raise exception.PortBindingFailed(port_id=port['id']) [ 718.022554] env[61923]: nova.exception.PortBindingFailed: Binding failed for port 8065e5e4-32c9-4969-8aff-a6bc9953f4bc, please check neutron logs for more information. [ 718.022554] env[61923]: Removing descriptor: 17 [ 718.022554] env[61923]: ERROR nova.compute.manager [None req-425a3606-915c-4f9b-8507-50f1c8d0d8d9 tempest-ServerRescueTestJSON-521767381 tempest-ServerRescueTestJSON-521767381-project-member] [instance: 74b32299-983f-41f7-b3f4-3a12815a83e6] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 8065e5e4-32c9-4969-8aff-a6bc9953f4bc, please check neutron logs for more information. [ 718.022554] env[61923]: ERROR nova.compute.manager [instance: 74b32299-983f-41f7-b3f4-3a12815a83e6] Traceback (most recent call last): [ 718.022554] env[61923]: ERROR nova.compute.manager [instance: 74b32299-983f-41f7-b3f4-3a12815a83e6] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 718.022554] env[61923]: ERROR nova.compute.manager [instance: 74b32299-983f-41f7-b3f4-3a12815a83e6] yield resources [ 718.022554] env[61923]: ERROR nova.compute.manager [instance: 74b32299-983f-41f7-b3f4-3a12815a83e6] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 718.022554] env[61923]: ERROR nova.compute.manager [instance: 74b32299-983f-41f7-b3f4-3a12815a83e6] self.driver.spawn(context, instance, image_meta, [ 718.022554] env[61923]: ERROR nova.compute.manager [instance: 74b32299-983f-41f7-b3f4-3a12815a83e6] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 718.022554] env[61923]: ERROR nova.compute.manager [instance: 74b32299-983f-41f7-b3f4-3a12815a83e6] self._vmops.spawn(context, instance, image_meta, injected_files, [ 718.022554] env[61923]: ERROR nova.compute.manager [instance: 74b32299-983f-41f7-b3f4-3a12815a83e6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 718.022554] env[61923]: ERROR nova.compute.manager [instance: 74b32299-983f-41f7-b3f4-3a12815a83e6] vm_ref = self.build_virtual_machine(instance, [ 718.023337] env[61923]: ERROR nova.compute.manager [instance: 74b32299-983f-41f7-b3f4-3a12815a83e6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 718.023337] env[61923]: ERROR nova.compute.manager [instance: 74b32299-983f-41f7-b3f4-3a12815a83e6] vif_infos = vmwarevif.get_vif_info(self._session, [ 718.023337] env[61923]: ERROR nova.compute.manager [instance: 74b32299-983f-41f7-b3f4-3a12815a83e6] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 718.023337] env[61923]: ERROR nova.compute.manager [instance: 74b32299-983f-41f7-b3f4-3a12815a83e6] for vif in network_info: [ 718.023337] env[61923]: ERROR nova.compute.manager [instance: 74b32299-983f-41f7-b3f4-3a12815a83e6] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 718.023337] env[61923]: ERROR nova.compute.manager [instance: 74b32299-983f-41f7-b3f4-3a12815a83e6] return self._sync_wrapper(fn, *args, **kwargs) [ 718.023337] env[61923]: ERROR nova.compute.manager [instance: 74b32299-983f-41f7-b3f4-3a12815a83e6] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 718.023337] env[61923]: ERROR nova.compute.manager [instance: 74b32299-983f-41f7-b3f4-3a12815a83e6] self.wait() [ 718.023337] env[61923]: ERROR nova.compute.manager [instance: 74b32299-983f-41f7-b3f4-3a12815a83e6] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 718.023337] env[61923]: ERROR nova.compute.manager [instance: 74b32299-983f-41f7-b3f4-3a12815a83e6] self[:] = self._gt.wait() [ 718.023337] env[61923]: ERROR nova.compute.manager [instance: 74b32299-983f-41f7-b3f4-3a12815a83e6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 718.023337] env[61923]: ERROR nova.compute.manager [instance: 74b32299-983f-41f7-b3f4-3a12815a83e6] return self._exit_event.wait() [ 718.023337] env[61923]: ERROR nova.compute.manager [instance: 74b32299-983f-41f7-b3f4-3a12815a83e6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 718.023766] env[61923]: ERROR nova.compute.manager [instance: 74b32299-983f-41f7-b3f4-3a12815a83e6] result = hub.switch() [ 718.023766] env[61923]: ERROR nova.compute.manager [instance: 74b32299-983f-41f7-b3f4-3a12815a83e6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 718.023766] env[61923]: ERROR nova.compute.manager [instance: 74b32299-983f-41f7-b3f4-3a12815a83e6] return self.greenlet.switch() [ 718.023766] env[61923]: ERROR nova.compute.manager [instance: 74b32299-983f-41f7-b3f4-3a12815a83e6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 718.023766] env[61923]: ERROR nova.compute.manager [instance: 74b32299-983f-41f7-b3f4-3a12815a83e6] result = function(*args, **kwargs) [ 718.023766] env[61923]: ERROR nova.compute.manager [instance: 74b32299-983f-41f7-b3f4-3a12815a83e6] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 718.023766] env[61923]: ERROR nova.compute.manager [instance: 74b32299-983f-41f7-b3f4-3a12815a83e6] return func(*args, **kwargs) [ 718.023766] env[61923]: ERROR nova.compute.manager [instance: 74b32299-983f-41f7-b3f4-3a12815a83e6] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 718.023766] env[61923]: ERROR nova.compute.manager [instance: 74b32299-983f-41f7-b3f4-3a12815a83e6] raise e [ 718.023766] env[61923]: ERROR nova.compute.manager [instance: 74b32299-983f-41f7-b3f4-3a12815a83e6] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 718.023766] env[61923]: ERROR nova.compute.manager [instance: 74b32299-983f-41f7-b3f4-3a12815a83e6] nwinfo = self.network_api.allocate_for_instance( [ 718.023766] env[61923]: ERROR nova.compute.manager [instance: 74b32299-983f-41f7-b3f4-3a12815a83e6] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 718.023766] env[61923]: ERROR nova.compute.manager [instance: 74b32299-983f-41f7-b3f4-3a12815a83e6] created_port_ids = self._update_ports_for_instance( [ 718.024199] env[61923]: ERROR nova.compute.manager [instance: 74b32299-983f-41f7-b3f4-3a12815a83e6] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 718.024199] env[61923]: ERROR nova.compute.manager [instance: 74b32299-983f-41f7-b3f4-3a12815a83e6] with excutils.save_and_reraise_exception(): [ 718.024199] env[61923]: ERROR nova.compute.manager [instance: 74b32299-983f-41f7-b3f4-3a12815a83e6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 718.024199] env[61923]: ERROR nova.compute.manager [instance: 74b32299-983f-41f7-b3f4-3a12815a83e6] self.force_reraise() [ 718.024199] env[61923]: ERROR nova.compute.manager [instance: 74b32299-983f-41f7-b3f4-3a12815a83e6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 718.024199] env[61923]: ERROR nova.compute.manager [instance: 74b32299-983f-41f7-b3f4-3a12815a83e6] raise self.value [ 718.024199] env[61923]: ERROR nova.compute.manager [instance: 74b32299-983f-41f7-b3f4-3a12815a83e6] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 718.024199] env[61923]: ERROR nova.compute.manager [instance: 74b32299-983f-41f7-b3f4-3a12815a83e6] updated_port = self._update_port( [ 718.024199] env[61923]: ERROR nova.compute.manager [instance: 74b32299-983f-41f7-b3f4-3a12815a83e6] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 718.024199] env[61923]: ERROR nova.compute.manager [instance: 74b32299-983f-41f7-b3f4-3a12815a83e6] _ensure_no_port_binding_failure(port) [ 718.024199] env[61923]: ERROR nova.compute.manager [instance: 74b32299-983f-41f7-b3f4-3a12815a83e6] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 718.024199] env[61923]: ERROR nova.compute.manager [instance: 74b32299-983f-41f7-b3f4-3a12815a83e6] raise exception.PortBindingFailed(port_id=port['id']) [ 718.024579] env[61923]: ERROR nova.compute.manager [instance: 74b32299-983f-41f7-b3f4-3a12815a83e6] nova.exception.PortBindingFailed: Binding failed for port 8065e5e4-32c9-4969-8aff-a6bc9953f4bc, please check neutron logs for more information. [ 718.024579] env[61923]: ERROR nova.compute.manager [instance: 74b32299-983f-41f7-b3f4-3a12815a83e6] [ 718.024579] env[61923]: INFO nova.compute.manager [None req-425a3606-915c-4f9b-8507-50f1c8d0d8d9 tempest-ServerRescueTestJSON-521767381 tempest-ServerRescueTestJSON-521767381-project-member] [instance: 74b32299-983f-41f7-b3f4-3a12815a83e6] Terminating instance [ 718.025327] env[61923]: DEBUG oslo_concurrency.lockutils [None req-425a3606-915c-4f9b-8507-50f1c8d0d8d9 tempest-ServerRescueTestJSON-521767381 tempest-ServerRescueTestJSON-521767381-project-member] Acquiring lock "refresh_cache-74b32299-983f-41f7-b3f4-3a12815a83e6" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 718.074604] env[61923]: DEBUG nova.network.neutron [None req-df4eb7d8-5b64-4db4-b042-683b16e0b991 tempest-ServersTestBootFromVolume-2104901628 tempest-ServersTestBootFromVolume-2104901628-project-member] [instance: fb548893-2df5-4629-bf3a-3b0f807f6357] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 718.198833] env[61923]: DEBUG oslo_concurrency.lockutils [None req-998fcb75-b821-4586-ae48-2fe4d02a9864 tempest-AttachVolumeShelveTestJSON-1252849694 tempest-AttachVolumeShelveTestJSON-1252849694-project-member] Acquiring lock "870602dd-1b85-4c97-9506-f08e9f1bade6" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 718.199143] env[61923]: DEBUG oslo_concurrency.lockutils [None req-998fcb75-b821-4586-ae48-2fe4d02a9864 tempest-AttachVolumeShelveTestJSON-1252849694 tempest-AttachVolumeShelveTestJSON-1252849694-project-member] Lock "870602dd-1b85-4c97-9506-f08e9f1bade6" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 718.202327] env[61923]: DEBUG nova.network.neutron [None req-df4eb7d8-5b64-4db4-b042-683b16e0b991 tempest-ServersTestBootFromVolume-2104901628 tempest-ServersTestBootFromVolume-2104901628-project-member] [instance: fb548893-2df5-4629-bf3a-3b0f807f6357] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 718.353084] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d85ccfcc-a737-44b8-bf0c-ddbaa7d3c06d {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.358784] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-895b45fd-9d42-4542-8f8b-c86992aaea9f {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.390858] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3077971b-1b7e-4ef0-a8b0-6db957dd7fdd {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.399069] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42d31fff-3639-4774-92eb-12542920c6e0 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.413053] env[61923]: DEBUG nova.compute.provider_tree [None req-44a76586-0f76-448f-af41-a5d846bfda81 tempest-ServerActionsV293TestJSON-849444864 tempest-ServerActionsV293TestJSON-849444864-project-member] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 718.492680] env[61923]: DEBUG nova.network.neutron [req-b35eecb4-7346-406a-969d-8e6844be522f req-c971b888-2935-4c09-9ff3-9f92debad42e service nova] [instance: 74b32299-983f-41f7-b3f4-3a12815a83e6] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 718.591562] env[61923]: DEBUG nova.network.neutron [req-b35eecb4-7346-406a-969d-8e6844be522f req-c971b888-2935-4c09-9ff3-9f92debad42e service nova] [instance: 74b32299-983f-41f7-b3f4-3a12815a83e6] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 718.704276] env[61923]: DEBUG oslo_concurrency.lockutils [None req-df4eb7d8-5b64-4db4-b042-683b16e0b991 tempest-ServersTestBootFromVolume-2104901628 tempest-ServersTestBootFromVolume-2104901628-project-member] Releasing lock "refresh_cache-fb548893-2df5-4629-bf3a-3b0f807f6357" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 718.704527] env[61923]: DEBUG nova.compute.manager [None req-df4eb7d8-5b64-4db4-b042-683b16e0b991 tempest-ServersTestBootFromVolume-2104901628 tempest-ServersTestBootFromVolume-2104901628-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61923) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 718.704710] env[61923]: DEBUG nova.compute.manager [None req-df4eb7d8-5b64-4db4-b042-683b16e0b991 tempest-ServersTestBootFromVolume-2104901628 tempest-ServersTestBootFromVolume-2104901628-project-member] [instance: fb548893-2df5-4629-bf3a-3b0f807f6357] Deallocating network for instance {{(pid=61923) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 718.704873] env[61923]: DEBUG nova.network.neutron [None req-df4eb7d8-5b64-4db4-b042-683b16e0b991 tempest-ServersTestBootFromVolume-2104901628 tempest-ServersTestBootFromVolume-2104901628-project-member] [instance: fb548893-2df5-4629-bf3a-3b0f807f6357] deallocate_for_instance() {{(pid=61923) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 718.722514] env[61923]: DEBUG nova.network.neutron [None req-df4eb7d8-5b64-4db4-b042-683b16e0b991 tempest-ServersTestBootFromVolume-2104901628 tempest-ServersTestBootFromVolume-2104901628-project-member] [instance: fb548893-2df5-4629-bf3a-3b0f807f6357] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 718.916437] env[61923]: DEBUG nova.scheduler.client.report [None req-44a76586-0f76-448f-af41-a5d846bfda81 tempest-ServerActionsV293TestJSON-849444864 tempest-ServerActionsV293TestJSON-849444864-project-member] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 719.095663] env[61923]: DEBUG oslo_concurrency.lockutils [req-b35eecb4-7346-406a-969d-8e6844be522f req-c971b888-2935-4c09-9ff3-9f92debad42e service nova] Releasing lock "refresh_cache-74b32299-983f-41f7-b3f4-3a12815a83e6" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 719.095663] env[61923]: DEBUG oslo_concurrency.lockutils [None req-425a3606-915c-4f9b-8507-50f1c8d0d8d9 tempest-ServerRescueTestJSON-521767381 tempest-ServerRescueTestJSON-521767381-project-member] Acquired lock "refresh_cache-74b32299-983f-41f7-b3f4-3a12815a83e6" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 719.095663] env[61923]: DEBUG nova.network.neutron [None req-425a3606-915c-4f9b-8507-50f1c8d0d8d9 tempest-ServerRescueTestJSON-521767381 tempest-ServerRescueTestJSON-521767381-project-member] [instance: 74b32299-983f-41f7-b3f4-3a12815a83e6] Building network info cache for instance {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 719.227117] env[61923]: DEBUG nova.network.neutron [None req-df4eb7d8-5b64-4db4-b042-683b16e0b991 tempest-ServersTestBootFromVolume-2104901628 tempest-ServersTestBootFromVolume-2104901628-project-member] [instance: fb548893-2df5-4629-bf3a-3b0f807f6357] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 719.423928] env[61923]: DEBUG oslo_concurrency.lockutils [None req-44a76586-0f76-448f-af41-a5d846bfda81 tempest-ServerActionsV293TestJSON-849444864 tempest-ServerActionsV293TestJSON-849444864-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.878s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 719.424589] env[61923]: ERROR nova.compute.manager [None req-44a76586-0f76-448f-af41-a5d846bfda81 tempest-ServerActionsV293TestJSON-849444864 tempest-ServerActionsV293TestJSON-849444864-project-member] [instance: 62144317-0354-4751-8f2b-4eae588c9da7] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 6059f47a-1eef-4b32-9c0c-25cdd6ab2609, please check neutron logs for more information. [ 719.424589] env[61923]: ERROR nova.compute.manager [instance: 62144317-0354-4751-8f2b-4eae588c9da7] Traceback (most recent call last): [ 719.424589] env[61923]: ERROR nova.compute.manager [instance: 62144317-0354-4751-8f2b-4eae588c9da7] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 719.424589] env[61923]: ERROR nova.compute.manager [instance: 62144317-0354-4751-8f2b-4eae588c9da7] self.driver.spawn(context, instance, image_meta, [ 719.424589] env[61923]: ERROR nova.compute.manager [instance: 62144317-0354-4751-8f2b-4eae588c9da7] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 719.424589] env[61923]: ERROR nova.compute.manager [instance: 62144317-0354-4751-8f2b-4eae588c9da7] self._vmops.spawn(context, instance, image_meta, injected_files, [ 719.424589] env[61923]: ERROR nova.compute.manager [instance: 62144317-0354-4751-8f2b-4eae588c9da7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 719.424589] env[61923]: ERROR nova.compute.manager [instance: 62144317-0354-4751-8f2b-4eae588c9da7] vm_ref = self.build_virtual_machine(instance, [ 719.424589] env[61923]: ERROR nova.compute.manager [instance: 62144317-0354-4751-8f2b-4eae588c9da7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 719.424589] env[61923]: ERROR nova.compute.manager [instance: 62144317-0354-4751-8f2b-4eae588c9da7] vif_infos = vmwarevif.get_vif_info(self._session, [ 719.424589] env[61923]: ERROR nova.compute.manager [instance: 62144317-0354-4751-8f2b-4eae588c9da7] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 719.424910] env[61923]: ERROR nova.compute.manager [instance: 62144317-0354-4751-8f2b-4eae588c9da7] for vif in network_info: [ 719.424910] env[61923]: ERROR nova.compute.manager [instance: 62144317-0354-4751-8f2b-4eae588c9da7] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 719.424910] env[61923]: ERROR nova.compute.manager [instance: 62144317-0354-4751-8f2b-4eae588c9da7] return self._sync_wrapper(fn, *args, **kwargs) [ 719.424910] env[61923]: ERROR nova.compute.manager [instance: 62144317-0354-4751-8f2b-4eae588c9da7] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 719.424910] env[61923]: ERROR nova.compute.manager [instance: 62144317-0354-4751-8f2b-4eae588c9da7] self.wait() [ 719.424910] env[61923]: ERROR nova.compute.manager [instance: 62144317-0354-4751-8f2b-4eae588c9da7] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 719.424910] env[61923]: ERROR nova.compute.manager [instance: 62144317-0354-4751-8f2b-4eae588c9da7] self[:] = self._gt.wait() [ 719.424910] env[61923]: ERROR nova.compute.manager [instance: 62144317-0354-4751-8f2b-4eae588c9da7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 719.424910] env[61923]: ERROR nova.compute.manager [instance: 62144317-0354-4751-8f2b-4eae588c9da7] return self._exit_event.wait() [ 719.424910] env[61923]: ERROR nova.compute.manager [instance: 62144317-0354-4751-8f2b-4eae588c9da7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 719.424910] env[61923]: ERROR nova.compute.manager [instance: 62144317-0354-4751-8f2b-4eae588c9da7] current.throw(*self._exc) [ 719.424910] env[61923]: ERROR nova.compute.manager [instance: 62144317-0354-4751-8f2b-4eae588c9da7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 719.424910] env[61923]: ERROR nova.compute.manager [instance: 62144317-0354-4751-8f2b-4eae588c9da7] result = function(*args, **kwargs) [ 719.425266] env[61923]: ERROR nova.compute.manager [instance: 62144317-0354-4751-8f2b-4eae588c9da7] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 719.425266] env[61923]: ERROR nova.compute.manager [instance: 62144317-0354-4751-8f2b-4eae588c9da7] return func(*args, **kwargs) [ 719.425266] env[61923]: ERROR nova.compute.manager [instance: 62144317-0354-4751-8f2b-4eae588c9da7] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 719.425266] env[61923]: ERROR nova.compute.manager [instance: 62144317-0354-4751-8f2b-4eae588c9da7] raise e [ 719.425266] env[61923]: ERROR nova.compute.manager [instance: 62144317-0354-4751-8f2b-4eae588c9da7] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 719.425266] env[61923]: ERROR nova.compute.manager [instance: 62144317-0354-4751-8f2b-4eae588c9da7] nwinfo = self.network_api.allocate_for_instance( [ 719.425266] env[61923]: ERROR nova.compute.manager [instance: 62144317-0354-4751-8f2b-4eae588c9da7] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 719.425266] env[61923]: ERROR nova.compute.manager [instance: 62144317-0354-4751-8f2b-4eae588c9da7] created_port_ids = self._update_ports_for_instance( [ 719.425266] env[61923]: ERROR nova.compute.manager [instance: 62144317-0354-4751-8f2b-4eae588c9da7] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 719.425266] env[61923]: ERROR nova.compute.manager [instance: 62144317-0354-4751-8f2b-4eae588c9da7] with excutils.save_and_reraise_exception(): [ 719.425266] env[61923]: ERROR nova.compute.manager [instance: 62144317-0354-4751-8f2b-4eae588c9da7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 719.425266] env[61923]: ERROR nova.compute.manager [instance: 62144317-0354-4751-8f2b-4eae588c9da7] self.force_reraise() [ 719.425266] env[61923]: ERROR nova.compute.manager [instance: 62144317-0354-4751-8f2b-4eae588c9da7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 719.425599] env[61923]: ERROR nova.compute.manager [instance: 62144317-0354-4751-8f2b-4eae588c9da7] raise self.value [ 719.425599] env[61923]: ERROR nova.compute.manager [instance: 62144317-0354-4751-8f2b-4eae588c9da7] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 719.425599] env[61923]: ERROR nova.compute.manager [instance: 62144317-0354-4751-8f2b-4eae588c9da7] updated_port = self._update_port( [ 719.425599] env[61923]: ERROR nova.compute.manager [instance: 62144317-0354-4751-8f2b-4eae588c9da7] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 719.425599] env[61923]: ERROR nova.compute.manager [instance: 62144317-0354-4751-8f2b-4eae588c9da7] _ensure_no_port_binding_failure(port) [ 719.425599] env[61923]: ERROR nova.compute.manager [instance: 62144317-0354-4751-8f2b-4eae588c9da7] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 719.425599] env[61923]: ERROR nova.compute.manager [instance: 62144317-0354-4751-8f2b-4eae588c9da7] raise exception.PortBindingFailed(port_id=port['id']) [ 719.425599] env[61923]: ERROR nova.compute.manager [instance: 62144317-0354-4751-8f2b-4eae588c9da7] nova.exception.PortBindingFailed: Binding failed for port 6059f47a-1eef-4b32-9c0c-25cdd6ab2609, please check neutron logs for more information. [ 719.425599] env[61923]: ERROR nova.compute.manager [instance: 62144317-0354-4751-8f2b-4eae588c9da7] [ 719.425599] env[61923]: DEBUG nova.compute.utils [None req-44a76586-0f76-448f-af41-a5d846bfda81 tempest-ServerActionsV293TestJSON-849444864 tempest-ServerActionsV293TestJSON-849444864-project-member] [instance: 62144317-0354-4751-8f2b-4eae588c9da7] Binding failed for port 6059f47a-1eef-4b32-9c0c-25cdd6ab2609, please check neutron logs for more information. {{(pid=61923) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 719.426646] env[61923]: DEBUG oslo_concurrency.lockutils [None req-c09a27e9-8e33-411f-b17b-7cd8d66a175c tempest-ServerMetadataNegativeTestJSON-1976245157 tempest-ServerMetadataNegativeTestJSON-1976245157-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.164s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 719.429830] env[61923]: DEBUG nova.compute.manager [None req-44a76586-0f76-448f-af41-a5d846bfda81 tempest-ServerActionsV293TestJSON-849444864 tempest-ServerActionsV293TestJSON-849444864-project-member] [instance: 62144317-0354-4751-8f2b-4eae588c9da7] Build of instance 62144317-0354-4751-8f2b-4eae588c9da7 was re-scheduled: Binding failed for port 6059f47a-1eef-4b32-9c0c-25cdd6ab2609, please check neutron logs for more information. {{(pid=61923) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 719.430347] env[61923]: DEBUG nova.compute.manager [None req-44a76586-0f76-448f-af41-a5d846bfda81 tempest-ServerActionsV293TestJSON-849444864 tempest-ServerActionsV293TestJSON-849444864-project-member] [instance: 62144317-0354-4751-8f2b-4eae588c9da7] Unplugging VIFs for instance {{(pid=61923) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 719.430610] env[61923]: DEBUG oslo_concurrency.lockutils [None req-44a76586-0f76-448f-af41-a5d846bfda81 tempest-ServerActionsV293TestJSON-849444864 tempest-ServerActionsV293TestJSON-849444864-project-member] Acquiring lock "refresh_cache-62144317-0354-4751-8f2b-4eae588c9da7" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 719.430785] env[61923]: DEBUG oslo_concurrency.lockutils [None req-44a76586-0f76-448f-af41-a5d846bfda81 tempest-ServerActionsV293TestJSON-849444864 tempest-ServerActionsV293TestJSON-849444864-project-member] Acquired lock "refresh_cache-62144317-0354-4751-8f2b-4eae588c9da7" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 719.430969] env[61923]: DEBUG nova.network.neutron [None req-44a76586-0f76-448f-af41-a5d846bfda81 tempest-ServerActionsV293TestJSON-849444864 tempest-ServerActionsV293TestJSON-849444864-project-member] [instance: 62144317-0354-4751-8f2b-4eae588c9da7] Building network info cache for instance {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 719.618630] env[61923]: DEBUG nova.network.neutron [None req-425a3606-915c-4f9b-8507-50f1c8d0d8d9 tempest-ServerRescueTestJSON-521767381 tempest-ServerRescueTestJSON-521767381-project-member] [instance: 74b32299-983f-41f7-b3f4-3a12815a83e6] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 719.729871] env[61923]: INFO nova.compute.manager [None req-df4eb7d8-5b64-4db4-b042-683b16e0b991 tempest-ServersTestBootFromVolume-2104901628 tempest-ServersTestBootFromVolume-2104901628-project-member] [instance: fb548893-2df5-4629-bf3a-3b0f807f6357] Took 1.02 seconds to deallocate network for instance. [ 719.734143] env[61923]: DEBUG nova.network.neutron [None req-425a3606-915c-4f9b-8507-50f1c8d0d8d9 tempest-ServerRescueTestJSON-521767381 tempest-ServerRescueTestJSON-521767381-project-member] [instance: 74b32299-983f-41f7-b3f4-3a12815a83e6] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 719.966110] env[61923]: DEBUG nova.network.neutron [None req-44a76586-0f76-448f-af41-a5d846bfda81 tempest-ServerActionsV293TestJSON-849444864 tempest-ServerActionsV293TestJSON-849444864-project-member] [instance: 62144317-0354-4751-8f2b-4eae588c9da7] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 720.040415] env[61923]: DEBUG nova.compute.manager [req-034dc83c-81ec-41d7-bc0e-26932be64931 req-e81a734d-78ea-40d3-b128-31a25b423bdb service nova] [instance: 74b32299-983f-41f7-b3f4-3a12815a83e6] Received event network-vif-deleted-8065e5e4-32c9-4969-8aff-a6bc9953f4bc {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 720.098493] env[61923]: DEBUG nova.network.neutron [None req-44a76586-0f76-448f-af41-a5d846bfda81 tempest-ServerActionsV293TestJSON-849444864 tempest-ServerActionsV293TestJSON-849444864-project-member] [instance: 62144317-0354-4751-8f2b-4eae588c9da7] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 720.239792] env[61923]: DEBUG oslo_concurrency.lockutils [None req-425a3606-915c-4f9b-8507-50f1c8d0d8d9 tempest-ServerRescueTestJSON-521767381 tempest-ServerRescueTestJSON-521767381-project-member] Releasing lock "refresh_cache-74b32299-983f-41f7-b3f4-3a12815a83e6" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 720.239792] env[61923]: DEBUG nova.compute.manager [None req-425a3606-915c-4f9b-8507-50f1c8d0d8d9 tempest-ServerRescueTestJSON-521767381 tempest-ServerRescueTestJSON-521767381-project-member] [instance: 74b32299-983f-41f7-b3f4-3a12815a83e6] Start destroying the instance on the hypervisor. {{(pid=61923) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 720.239792] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-425a3606-915c-4f9b-8507-50f1c8d0d8d9 tempest-ServerRescueTestJSON-521767381 tempest-ServerRescueTestJSON-521767381-project-member] [instance: 74b32299-983f-41f7-b3f4-3a12815a83e6] Destroying instance {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 720.243026] env[61923]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-6ee1152d-1759-45f4-9d1f-dbcdd90d2835 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.251259] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-248af6de-2b85-40c9-b89e-1d068efba172 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.277220] env[61923]: WARNING nova.virt.vmwareapi.vmops [None req-425a3606-915c-4f9b-8507-50f1c8d0d8d9 tempest-ServerRescueTestJSON-521767381 tempest-ServerRescueTestJSON-521767381-project-member] [instance: 74b32299-983f-41f7-b3f4-3a12815a83e6] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 74b32299-983f-41f7-b3f4-3a12815a83e6 could not be found. [ 720.277380] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-425a3606-915c-4f9b-8507-50f1c8d0d8d9 tempest-ServerRescueTestJSON-521767381 tempest-ServerRescueTestJSON-521767381-project-member] [instance: 74b32299-983f-41f7-b3f4-3a12815a83e6] Instance destroyed {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 720.277558] env[61923]: INFO nova.compute.manager [None req-425a3606-915c-4f9b-8507-50f1c8d0d8d9 tempest-ServerRescueTestJSON-521767381 tempest-ServerRescueTestJSON-521767381-project-member] [instance: 74b32299-983f-41f7-b3f4-3a12815a83e6] Took 0.04 seconds to destroy the instance on the hypervisor. [ 720.278210] env[61923]: DEBUG oslo.service.loopingcall [None req-425a3606-915c-4f9b-8507-50f1c8d0d8d9 tempest-ServerRescueTestJSON-521767381 tempest-ServerRescueTestJSON-521767381-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61923) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 720.280575] env[61923]: DEBUG nova.compute.manager [-] [instance: 74b32299-983f-41f7-b3f4-3a12815a83e6] Deallocating network for instance {{(pid=61923) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 720.280676] env[61923]: DEBUG nova.network.neutron [-] [instance: 74b32299-983f-41f7-b3f4-3a12815a83e6] deallocate_for_instance() {{(pid=61923) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 720.306206] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22afded8-9f78-4cf7-ac7e-163e8cc5d1da {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.311349] env[61923]: DEBUG nova.network.neutron [-] [instance: 74b32299-983f-41f7-b3f4-3a12815a83e6] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 720.315738] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cea63f98-6c9f-429d-9fa9-e12bec0ef81f {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.347457] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a82ca166-bd84-4b6f-a788-236ed30460bd {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.355309] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2ef9e64-2d7b-4203-9432-60f99508069f {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.369170] env[61923]: DEBUG nova.compute.provider_tree [None req-c09a27e9-8e33-411f-b17b-7cd8d66a175c tempest-ServerMetadataNegativeTestJSON-1976245157 tempest-ServerMetadataNegativeTestJSON-1976245157-project-member] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 720.601441] env[61923]: DEBUG oslo_concurrency.lockutils [None req-44a76586-0f76-448f-af41-a5d846bfda81 tempest-ServerActionsV293TestJSON-849444864 tempest-ServerActionsV293TestJSON-849444864-project-member] Releasing lock "refresh_cache-62144317-0354-4751-8f2b-4eae588c9da7" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 720.601853] env[61923]: DEBUG nova.compute.manager [None req-44a76586-0f76-448f-af41-a5d846bfda81 tempest-ServerActionsV293TestJSON-849444864 tempest-ServerActionsV293TestJSON-849444864-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61923) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 720.602172] env[61923]: DEBUG nova.compute.manager [None req-44a76586-0f76-448f-af41-a5d846bfda81 tempest-ServerActionsV293TestJSON-849444864 tempest-ServerActionsV293TestJSON-849444864-project-member] [instance: 62144317-0354-4751-8f2b-4eae588c9da7] Deallocating network for instance {{(pid=61923) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 720.602468] env[61923]: DEBUG nova.network.neutron [None req-44a76586-0f76-448f-af41-a5d846bfda81 tempest-ServerActionsV293TestJSON-849444864 tempest-ServerActionsV293TestJSON-849444864-project-member] [instance: 62144317-0354-4751-8f2b-4eae588c9da7] deallocate_for_instance() {{(pid=61923) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 720.621021] env[61923]: DEBUG nova.network.neutron [None req-44a76586-0f76-448f-af41-a5d846bfda81 tempest-ServerActionsV293TestJSON-849444864 tempest-ServerActionsV293TestJSON-849444864-project-member] [instance: 62144317-0354-4751-8f2b-4eae588c9da7] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 720.762523] env[61923]: INFO nova.scheduler.client.report [None req-df4eb7d8-5b64-4db4-b042-683b16e0b991 tempest-ServersTestBootFromVolume-2104901628 tempest-ServersTestBootFromVolume-2104901628-project-member] Deleted allocations for instance fb548893-2df5-4629-bf3a-3b0f807f6357 [ 720.812515] env[61923]: DEBUG nova.network.neutron [-] [instance: 74b32299-983f-41f7-b3f4-3a12815a83e6] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 720.872665] env[61923]: DEBUG nova.scheduler.client.report [None req-c09a27e9-8e33-411f-b17b-7cd8d66a175c tempest-ServerMetadataNegativeTestJSON-1976245157 tempest-ServerMetadataNegativeTestJSON-1976245157-project-member] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 721.123616] env[61923]: DEBUG nova.network.neutron [None req-44a76586-0f76-448f-af41-a5d846bfda81 tempest-ServerActionsV293TestJSON-849444864 tempest-ServerActionsV293TestJSON-849444864-project-member] [instance: 62144317-0354-4751-8f2b-4eae588c9da7] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 721.246613] env[61923]: DEBUG oslo_concurrency.lockutils [None req-c06a566a-3828-47dc-91c4-25d15d717c22 tempest-ServerShowV257Test-986496257 tempest-ServerShowV257Test-986496257-project-member] Acquiring lock "b8c60af7-3561-468c-a9a5-1c645080fc69" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 721.246865] env[61923]: DEBUG oslo_concurrency.lockutils [None req-c06a566a-3828-47dc-91c4-25d15d717c22 tempest-ServerShowV257Test-986496257 tempest-ServerShowV257Test-986496257-project-member] Lock "b8c60af7-3561-468c-a9a5-1c645080fc69" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 721.272782] env[61923]: DEBUG oslo_concurrency.lockutils [None req-df4eb7d8-5b64-4db4-b042-683b16e0b991 tempest-ServersTestBootFromVolume-2104901628 tempest-ServersTestBootFromVolume-2104901628-project-member] Lock "fb548893-2df5-4629-bf3a-3b0f807f6357" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 137.316s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 721.316046] env[61923]: INFO nova.compute.manager [-] [instance: 74b32299-983f-41f7-b3f4-3a12815a83e6] Took 1.03 seconds to deallocate network for instance. [ 721.317496] env[61923]: DEBUG nova.compute.claims [None req-425a3606-915c-4f9b-8507-50f1c8d0d8d9 tempest-ServerRescueTestJSON-521767381 tempest-ServerRescueTestJSON-521767381-project-member] [instance: 74b32299-983f-41f7-b3f4-3a12815a83e6] Aborting claim: {{(pid=61923) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 721.317871] env[61923]: DEBUG oslo_concurrency.lockutils [None req-425a3606-915c-4f9b-8507-50f1c8d0d8d9 tempest-ServerRescueTestJSON-521767381 tempest-ServerRescueTestJSON-521767381-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 721.378385] env[61923]: DEBUG oslo_concurrency.lockutils [None req-c09a27e9-8e33-411f-b17b-7cd8d66a175c tempest-ServerMetadataNegativeTestJSON-1976245157 tempest-ServerMetadataNegativeTestJSON-1976245157-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.952s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 721.379077] env[61923]: ERROR nova.compute.manager [None req-c09a27e9-8e33-411f-b17b-7cd8d66a175c tempest-ServerMetadataNegativeTestJSON-1976245157 tempest-ServerMetadataNegativeTestJSON-1976245157-project-member] [instance: 25b62db6-6e64-4e38-bfd3-82a2e9a4883a] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port bf236ea4-19e2-487c-ad15-61be750ff9c5, please check neutron logs for more information. [ 721.379077] env[61923]: ERROR nova.compute.manager [instance: 25b62db6-6e64-4e38-bfd3-82a2e9a4883a] Traceback (most recent call last): [ 721.379077] env[61923]: ERROR nova.compute.manager [instance: 25b62db6-6e64-4e38-bfd3-82a2e9a4883a] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 721.379077] env[61923]: ERROR nova.compute.manager [instance: 25b62db6-6e64-4e38-bfd3-82a2e9a4883a] self.driver.spawn(context, instance, image_meta, [ 721.379077] env[61923]: ERROR nova.compute.manager [instance: 25b62db6-6e64-4e38-bfd3-82a2e9a4883a] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 721.379077] env[61923]: ERROR nova.compute.manager [instance: 25b62db6-6e64-4e38-bfd3-82a2e9a4883a] self._vmops.spawn(context, instance, image_meta, injected_files, [ 721.379077] env[61923]: ERROR nova.compute.manager [instance: 25b62db6-6e64-4e38-bfd3-82a2e9a4883a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 721.379077] env[61923]: ERROR nova.compute.manager [instance: 25b62db6-6e64-4e38-bfd3-82a2e9a4883a] vm_ref = self.build_virtual_machine(instance, [ 721.379077] env[61923]: ERROR nova.compute.manager [instance: 25b62db6-6e64-4e38-bfd3-82a2e9a4883a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 721.379077] env[61923]: ERROR nova.compute.manager [instance: 25b62db6-6e64-4e38-bfd3-82a2e9a4883a] vif_infos = vmwarevif.get_vif_info(self._session, [ 721.379077] env[61923]: ERROR nova.compute.manager [instance: 25b62db6-6e64-4e38-bfd3-82a2e9a4883a] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 721.379405] env[61923]: ERROR nova.compute.manager [instance: 25b62db6-6e64-4e38-bfd3-82a2e9a4883a] for vif in network_info: [ 721.379405] env[61923]: ERROR nova.compute.manager [instance: 25b62db6-6e64-4e38-bfd3-82a2e9a4883a] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 721.379405] env[61923]: ERROR nova.compute.manager [instance: 25b62db6-6e64-4e38-bfd3-82a2e9a4883a] return self._sync_wrapper(fn, *args, **kwargs) [ 721.379405] env[61923]: ERROR nova.compute.manager [instance: 25b62db6-6e64-4e38-bfd3-82a2e9a4883a] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 721.379405] env[61923]: ERROR nova.compute.manager [instance: 25b62db6-6e64-4e38-bfd3-82a2e9a4883a] self.wait() [ 721.379405] env[61923]: ERROR nova.compute.manager [instance: 25b62db6-6e64-4e38-bfd3-82a2e9a4883a] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 721.379405] env[61923]: ERROR nova.compute.manager [instance: 25b62db6-6e64-4e38-bfd3-82a2e9a4883a] self[:] = self._gt.wait() [ 721.379405] env[61923]: ERROR nova.compute.manager [instance: 25b62db6-6e64-4e38-bfd3-82a2e9a4883a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 721.379405] env[61923]: ERROR nova.compute.manager [instance: 25b62db6-6e64-4e38-bfd3-82a2e9a4883a] return self._exit_event.wait() [ 721.379405] env[61923]: ERROR nova.compute.manager [instance: 25b62db6-6e64-4e38-bfd3-82a2e9a4883a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 721.379405] env[61923]: ERROR nova.compute.manager [instance: 25b62db6-6e64-4e38-bfd3-82a2e9a4883a] result = hub.switch() [ 721.379405] env[61923]: ERROR nova.compute.manager [instance: 25b62db6-6e64-4e38-bfd3-82a2e9a4883a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 721.379405] env[61923]: ERROR nova.compute.manager [instance: 25b62db6-6e64-4e38-bfd3-82a2e9a4883a] return self.greenlet.switch() [ 721.379735] env[61923]: ERROR nova.compute.manager [instance: 25b62db6-6e64-4e38-bfd3-82a2e9a4883a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 721.379735] env[61923]: ERROR nova.compute.manager [instance: 25b62db6-6e64-4e38-bfd3-82a2e9a4883a] result = function(*args, **kwargs) [ 721.379735] env[61923]: ERROR nova.compute.manager [instance: 25b62db6-6e64-4e38-bfd3-82a2e9a4883a] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 721.379735] env[61923]: ERROR nova.compute.manager [instance: 25b62db6-6e64-4e38-bfd3-82a2e9a4883a] return func(*args, **kwargs) [ 721.379735] env[61923]: ERROR nova.compute.manager [instance: 25b62db6-6e64-4e38-bfd3-82a2e9a4883a] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 721.379735] env[61923]: ERROR nova.compute.manager [instance: 25b62db6-6e64-4e38-bfd3-82a2e9a4883a] raise e [ 721.379735] env[61923]: ERROR nova.compute.manager [instance: 25b62db6-6e64-4e38-bfd3-82a2e9a4883a] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 721.379735] env[61923]: ERROR nova.compute.manager [instance: 25b62db6-6e64-4e38-bfd3-82a2e9a4883a] nwinfo = self.network_api.allocate_for_instance( [ 721.379735] env[61923]: ERROR nova.compute.manager [instance: 25b62db6-6e64-4e38-bfd3-82a2e9a4883a] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 721.379735] env[61923]: ERROR nova.compute.manager [instance: 25b62db6-6e64-4e38-bfd3-82a2e9a4883a] created_port_ids = self._update_ports_for_instance( [ 721.379735] env[61923]: ERROR nova.compute.manager [instance: 25b62db6-6e64-4e38-bfd3-82a2e9a4883a] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 721.379735] env[61923]: ERROR nova.compute.manager [instance: 25b62db6-6e64-4e38-bfd3-82a2e9a4883a] with excutils.save_and_reraise_exception(): [ 721.379735] env[61923]: ERROR nova.compute.manager [instance: 25b62db6-6e64-4e38-bfd3-82a2e9a4883a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 721.380084] env[61923]: ERROR nova.compute.manager [instance: 25b62db6-6e64-4e38-bfd3-82a2e9a4883a] self.force_reraise() [ 721.380084] env[61923]: ERROR nova.compute.manager [instance: 25b62db6-6e64-4e38-bfd3-82a2e9a4883a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 721.380084] env[61923]: ERROR nova.compute.manager [instance: 25b62db6-6e64-4e38-bfd3-82a2e9a4883a] raise self.value [ 721.380084] env[61923]: ERROR nova.compute.manager [instance: 25b62db6-6e64-4e38-bfd3-82a2e9a4883a] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 721.380084] env[61923]: ERROR nova.compute.manager [instance: 25b62db6-6e64-4e38-bfd3-82a2e9a4883a] updated_port = self._update_port( [ 721.380084] env[61923]: ERROR nova.compute.manager [instance: 25b62db6-6e64-4e38-bfd3-82a2e9a4883a] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 721.380084] env[61923]: ERROR nova.compute.manager [instance: 25b62db6-6e64-4e38-bfd3-82a2e9a4883a] _ensure_no_port_binding_failure(port) [ 721.380084] env[61923]: ERROR nova.compute.manager [instance: 25b62db6-6e64-4e38-bfd3-82a2e9a4883a] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 721.380084] env[61923]: ERROR nova.compute.manager [instance: 25b62db6-6e64-4e38-bfd3-82a2e9a4883a] raise exception.PortBindingFailed(port_id=port['id']) [ 721.380084] env[61923]: ERROR nova.compute.manager [instance: 25b62db6-6e64-4e38-bfd3-82a2e9a4883a] nova.exception.PortBindingFailed: Binding failed for port bf236ea4-19e2-487c-ad15-61be750ff9c5, please check neutron logs for more information. [ 721.380084] env[61923]: ERROR nova.compute.manager [instance: 25b62db6-6e64-4e38-bfd3-82a2e9a4883a] [ 721.380412] env[61923]: DEBUG nova.compute.utils [None req-c09a27e9-8e33-411f-b17b-7cd8d66a175c tempest-ServerMetadataNegativeTestJSON-1976245157 tempest-ServerMetadataNegativeTestJSON-1976245157-project-member] [instance: 25b62db6-6e64-4e38-bfd3-82a2e9a4883a] Binding failed for port bf236ea4-19e2-487c-ad15-61be750ff9c5, please check neutron logs for more information. {{(pid=61923) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 721.381704] env[61923]: DEBUG oslo_concurrency.lockutils [None req-40b011d4-39ab-4684-a124-b3ba87d489df tempest-ServerPasswordTestJSON-595643156 tempest-ServerPasswordTestJSON-595643156-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.880s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 721.383238] env[61923]: INFO nova.compute.claims [None req-40b011d4-39ab-4684-a124-b3ba87d489df tempest-ServerPasswordTestJSON-595643156 tempest-ServerPasswordTestJSON-595643156-project-member] [instance: 7c0341f9-7401-42ee-9512-afd832c76940] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 721.385974] env[61923]: DEBUG nova.compute.manager [None req-c09a27e9-8e33-411f-b17b-7cd8d66a175c tempest-ServerMetadataNegativeTestJSON-1976245157 tempest-ServerMetadataNegativeTestJSON-1976245157-project-member] [instance: 25b62db6-6e64-4e38-bfd3-82a2e9a4883a] Build of instance 25b62db6-6e64-4e38-bfd3-82a2e9a4883a was re-scheduled: Binding failed for port bf236ea4-19e2-487c-ad15-61be750ff9c5, please check neutron logs for more information. {{(pid=61923) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 721.386524] env[61923]: DEBUG nova.compute.manager [None req-c09a27e9-8e33-411f-b17b-7cd8d66a175c tempest-ServerMetadataNegativeTestJSON-1976245157 tempest-ServerMetadataNegativeTestJSON-1976245157-project-member] [instance: 25b62db6-6e64-4e38-bfd3-82a2e9a4883a] Unplugging VIFs for instance {{(pid=61923) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 721.386889] env[61923]: DEBUG oslo_concurrency.lockutils [None req-c09a27e9-8e33-411f-b17b-7cd8d66a175c tempest-ServerMetadataNegativeTestJSON-1976245157 tempest-ServerMetadataNegativeTestJSON-1976245157-project-member] Acquiring lock "refresh_cache-25b62db6-6e64-4e38-bfd3-82a2e9a4883a" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 721.387109] env[61923]: DEBUG oslo_concurrency.lockutils [None req-c09a27e9-8e33-411f-b17b-7cd8d66a175c tempest-ServerMetadataNegativeTestJSON-1976245157 tempest-ServerMetadataNegativeTestJSON-1976245157-project-member] Acquired lock "refresh_cache-25b62db6-6e64-4e38-bfd3-82a2e9a4883a" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 721.387360] env[61923]: DEBUG nova.network.neutron [None req-c09a27e9-8e33-411f-b17b-7cd8d66a175c tempest-ServerMetadataNegativeTestJSON-1976245157 tempest-ServerMetadataNegativeTestJSON-1976245157-project-member] [instance: 25b62db6-6e64-4e38-bfd3-82a2e9a4883a] Building network info cache for instance {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 721.626346] env[61923]: INFO nova.compute.manager [None req-44a76586-0f76-448f-af41-a5d846bfda81 tempest-ServerActionsV293TestJSON-849444864 tempest-ServerActionsV293TestJSON-849444864-project-member] [instance: 62144317-0354-4751-8f2b-4eae588c9da7] Took 1.02 seconds to deallocate network for instance. [ 721.774417] env[61923]: DEBUG nova.compute.manager [None req-0708225f-80bb-41ec-b3ed-7dd45d589b95 tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] [instance: 62d8a170-e7ad-4d14-90d8-6f6ce32b7cf8] Starting instance... {{(pid=61923) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 721.914746] env[61923]: DEBUG nova.network.neutron [None req-c09a27e9-8e33-411f-b17b-7cd8d66a175c tempest-ServerMetadataNegativeTestJSON-1976245157 tempest-ServerMetadataNegativeTestJSON-1976245157-project-member] [instance: 25b62db6-6e64-4e38-bfd3-82a2e9a4883a] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 722.004858] env[61923]: DEBUG nova.network.neutron [None req-c09a27e9-8e33-411f-b17b-7cd8d66a175c tempest-ServerMetadataNegativeTestJSON-1976245157 tempest-ServerMetadataNegativeTestJSON-1976245157-project-member] [instance: 25b62db6-6e64-4e38-bfd3-82a2e9a4883a] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 722.299974] env[61923]: DEBUG oslo_concurrency.lockutils [None req-0708225f-80bb-41ec-b3ed-7dd45d589b95 tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 722.508281] env[61923]: DEBUG oslo_concurrency.lockutils [None req-c09a27e9-8e33-411f-b17b-7cd8d66a175c tempest-ServerMetadataNegativeTestJSON-1976245157 tempest-ServerMetadataNegativeTestJSON-1976245157-project-member] Releasing lock "refresh_cache-25b62db6-6e64-4e38-bfd3-82a2e9a4883a" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 722.508281] env[61923]: DEBUG nova.compute.manager [None req-c09a27e9-8e33-411f-b17b-7cd8d66a175c tempest-ServerMetadataNegativeTestJSON-1976245157 tempest-ServerMetadataNegativeTestJSON-1976245157-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61923) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 722.508611] env[61923]: DEBUG nova.compute.manager [None req-c09a27e9-8e33-411f-b17b-7cd8d66a175c tempest-ServerMetadataNegativeTestJSON-1976245157 tempest-ServerMetadataNegativeTestJSON-1976245157-project-member] [instance: 25b62db6-6e64-4e38-bfd3-82a2e9a4883a] Deallocating network for instance {{(pid=61923) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 722.509249] env[61923]: DEBUG nova.network.neutron [None req-c09a27e9-8e33-411f-b17b-7cd8d66a175c tempest-ServerMetadataNegativeTestJSON-1976245157 tempest-ServerMetadataNegativeTestJSON-1976245157-project-member] [instance: 25b62db6-6e64-4e38-bfd3-82a2e9a4883a] deallocate_for_instance() {{(pid=61923) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 722.544718] env[61923]: DEBUG nova.network.neutron [None req-c09a27e9-8e33-411f-b17b-7cd8d66a175c tempest-ServerMetadataNegativeTestJSON-1976245157 tempest-ServerMetadataNegativeTestJSON-1976245157-project-member] [instance: 25b62db6-6e64-4e38-bfd3-82a2e9a4883a] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 722.661035] env[61923]: INFO nova.scheduler.client.report [None req-44a76586-0f76-448f-af41-a5d846bfda81 tempest-ServerActionsV293TestJSON-849444864 tempest-ServerActionsV293TestJSON-849444864-project-member] Deleted allocations for instance 62144317-0354-4751-8f2b-4eae588c9da7 [ 722.768454] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31c69831-a7d7-4316-83d2-528a852c30b0 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.776471] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33facbcb-ae23-424e-ba79-e997c6fc912e {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.810269] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-517726e6-83af-4887-99ec-c3074c6a4d0e {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.817951] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e7a9446-d094-4962-83a4-1b0c255df700 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.832190] env[61923]: DEBUG nova.compute.provider_tree [None req-40b011d4-39ab-4684-a124-b3ba87d489df tempest-ServerPasswordTestJSON-595643156 tempest-ServerPasswordTestJSON-595643156-project-member] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 723.047055] env[61923]: DEBUG nova.network.neutron [None req-c09a27e9-8e33-411f-b17b-7cd8d66a175c tempest-ServerMetadataNegativeTestJSON-1976245157 tempest-ServerMetadataNegativeTestJSON-1976245157-project-member] [instance: 25b62db6-6e64-4e38-bfd3-82a2e9a4883a] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 723.176668] env[61923]: DEBUG oslo_concurrency.lockutils [None req-44a76586-0f76-448f-af41-a5d846bfda81 tempest-ServerActionsV293TestJSON-849444864 tempest-ServerActionsV293TestJSON-849444864-project-member] Lock "62144317-0354-4751-8f2b-4eae588c9da7" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 130.836s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 723.336127] env[61923]: DEBUG nova.scheduler.client.report [None req-40b011d4-39ab-4684-a124-b3ba87d489df tempest-ServerPasswordTestJSON-595643156 tempest-ServerPasswordTestJSON-595643156-project-member] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 723.549719] env[61923]: INFO nova.compute.manager [None req-c09a27e9-8e33-411f-b17b-7cd8d66a175c tempest-ServerMetadataNegativeTestJSON-1976245157 tempest-ServerMetadataNegativeTestJSON-1976245157-project-member] [instance: 25b62db6-6e64-4e38-bfd3-82a2e9a4883a] Took 1.04 seconds to deallocate network for instance. [ 723.679217] env[61923]: DEBUG nova.compute.manager [None req-ab7b3da0-3a01-44d4-9dfd-3e5d2ad27432 tempest-TenantUsagesTestJSON-871253541 tempest-TenantUsagesTestJSON-871253541-project-member] [instance: b6712878-3270-43f4-9518-d41eb77124b7] Starting instance... {{(pid=61923) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 723.841198] env[61923]: DEBUG oslo_concurrency.lockutils [None req-40b011d4-39ab-4684-a124-b3ba87d489df tempest-ServerPasswordTestJSON-595643156 tempest-ServerPasswordTestJSON-595643156-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.459s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 723.841709] env[61923]: DEBUG nova.compute.manager [None req-40b011d4-39ab-4684-a124-b3ba87d489df tempest-ServerPasswordTestJSON-595643156 tempest-ServerPasswordTestJSON-595643156-project-member] [instance: 7c0341f9-7401-42ee-9512-afd832c76940] Start building networks asynchronously for instance. {{(pid=61923) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 723.844604] env[61923]: DEBUG oslo_concurrency.lockutils [None req-24801cc4-3974-42da-978f-a22cb8f70af0 tempest-SecurityGroupsTestJSON-509490211 tempest-SecurityGroupsTestJSON-509490211-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 13.151s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 724.204898] env[61923]: DEBUG oslo_concurrency.lockutils [None req-ab7b3da0-3a01-44d4-9dfd-3e5d2ad27432 tempest-TenantUsagesTestJSON-871253541 tempest-TenantUsagesTestJSON-871253541-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 724.350878] env[61923]: DEBUG nova.compute.utils [None req-40b011d4-39ab-4684-a124-b3ba87d489df tempest-ServerPasswordTestJSON-595643156 tempest-ServerPasswordTestJSON-595643156-project-member] Using /dev/sd instead of None {{(pid=61923) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 724.355512] env[61923]: DEBUG nova.compute.manager [None req-40b011d4-39ab-4684-a124-b3ba87d489df tempest-ServerPasswordTestJSON-595643156 tempest-ServerPasswordTestJSON-595643156-project-member] [instance: 7c0341f9-7401-42ee-9512-afd832c76940] Allocating IP information in the background. {{(pid=61923) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 724.355687] env[61923]: DEBUG nova.network.neutron [None req-40b011d4-39ab-4684-a124-b3ba87d489df tempest-ServerPasswordTestJSON-595643156 tempest-ServerPasswordTestJSON-595643156-project-member] [instance: 7c0341f9-7401-42ee-9512-afd832c76940] allocate_for_instance() {{(pid=61923) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 724.589205] env[61923]: INFO nova.scheduler.client.report [None req-c09a27e9-8e33-411f-b17b-7cd8d66a175c tempest-ServerMetadataNegativeTestJSON-1976245157 tempest-ServerMetadataNegativeTestJSON-1976245157-project-member] Deleted allocations for instance 25b62db6-6e64-4e38-bfd3-82a2e9a4883a [ 724.609886] env[61923]: DEBUG nova.policy [None req-40b011d4-39ab-4684-a124-b3ba87d489df tempest-ServerPasswordTestJSON-595643156 tempest-ServerPasswordTestJSON-595643156-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '43bcbd57bb024fe4bfe0d88539775808', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '6f2e284cda8b488794fb282d996e9d49', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61923) authorize /opt/stack/nova/nova/policy.py:201}} [ 724.768828] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-165f1c1a-aba3-499e-90c3-dc8d432b1f4c {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.778809] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d4674ac-8040-402c-b3d1-75748219d675 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.809077] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5fbbd809-db1d-4062-872c-e34f9b170944 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.816722] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-333da1ee-a4e3-418b-a2d1-88cbdc40d847 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 724.832371] env[61923]: DEBUG nova.compute.provider_tree [None req-24801cc4-3974-42da-978f-a22cb8f70af0 tempest-SecurityGroupsTestJSON-509490211 tempest-SecurityGroupsTestJSON-509490211-project-member] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 724.858955] env[61923]: DEBUG nova.compute.manager [None req-40b011d4-39ab-4684-a124-b3ba87d489df tempest-ServerPasswordTestJSON-595643156 tempest-ServerPasswordTestJSON-595643156-project-member] [instance: 7c0341f9-7401-42ee-9512-afd832c76940] Start building block device mappings for instance. {{(pid=61923) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 725.045593] env[61923]: DEBUG nova.network.neutron [None req-40b011d4-39ab-4684-a124-b3ba87d489df tempest-ServerPasswordTestJSON-595643156 tempest-ServerPasswordTestJSON-595643156-project-member] [instance: 7c0341f9-7401-42ee-9512-afd832c76940] Successfully created port: e84bf904-db99-4abc-90aa-4be30a8f32e8 {{(pid=61923) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 725.098569] env[61923]: DEBUG oslo_concurrency.lockutils [None req-c09a27e9-8e33-411f-b17b-7cd8d66a175c tempest-ServerMetadataNegativeTestJSON-1976245157 tempest-ServerMetadataNegativeTestJSON-1976245157-project-member] Lock "25b62db6-6e64-4e38-bfd3-82a2e9a4883a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 130.756s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 725.336264] env[61923]: DEBUG nova.scheduler.client.report [None req-24801cc4-3974-42da-978f-a22cb8f70af0 tempest-SecurityGroupsTestJSON-509490211 tempest-SecurityGroupsTestJSON-509490211-project-member] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 725.601119] env[61923]: DEBUG nova.compute.manager [None req-5138192d-6349-432b-b7de-64ec5d391f8d tempest-ServerActionsTestOtherB-987045973 tempest-ServerActionsTestOtherB-987045973-project-member] [instance: b0bc08c1-38d5-45ee-b521-1cffda41c77c] Starting instance... {{(pid=61923) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 725.843116] env[61923]: DEBUG oslo_concurrency.lockutils [None req-24801cc4-3974-42da-978f-a22cb8f70af0 tempest-SecurityGroupsTestJSON-509490211 tempest-SecurityGroupsTestJSON-509490211-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.997s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 725.843116] env[61923]: ERROR nova.compute.manager [None req-24801cc4-3974-42da-978f-a22cb8f70af0 tempest-SecurityGroupsTestJSON-509490211 tempest-SecurityGroupsTestJSON-509490211-project-member] [instance: 2939d1e5-07fe-48b3-bc87-5f67fa1a7ae4] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port cb32985d-46e9-42d3-a1cd-7e7620873e5e, please check neutron logs for more information. [ 725.843116] env[61923]: ERROR nova.compute.manager [instance: 2939d1e5-07fe-48b3-bc87-5f67fa1a7ae4] Traceback (most recent call last): [ 725.843116] env[61923]: ERROR nova.compute.manager [instance: 2939d1e5-07fe-48b3-bc87-5f67fa1a7ae4] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 725.843116] env[61923]: ERROR nova.compute.manager [instance: 2939d1e5-07fe-48b3-bc87-5f67fa1a7ae4] self.driver.spawn(context, instance, image_meta, [ 725.843116] env[61923]: ERROR nova.compute.manager [instance: 2939d1e5-07fe-48b3-bc87-5f67fa1a7ae4] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 725.843116] env[61923]: ERROR nova.compute.manager [instance: 2939d1e5-07fe-48b3-bc87-5f67fa1a7ae4] self._vmops.spawn(context, instance, image_meta, injected_files, [ 725.843116] env[61923]: ERROR nova.compute.manager [instance: 2939d1e5-07fe-48b3-bc87-5f67fa1a7ae4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 725.843116] env[61923]: ERROR nova.compute.manager [instance: 2939d1e5-07fe-48b3-bc87-5f67fa1a7ae4] vm_ref = self.build_virtual_machine(instance, [ 725.843797] env[61923]: ERROR nova.compute.manager [instance: 2939d1e5-07fe-48b3-bc87-5f67fa1a7ae4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 725.843797] env[61923]: ERROR nova.compute.manager [instance: 2939d1e5-07fe-48b3-bc87-5f67fa1a7ae4] vif_infos = vmwarevif.get_vif_info(self._session, [ 725.843797] env[61923]: ERROR nova.compute.manager [instance: 2939d1e5-07fe-48b3-bc87-5f67fa1a7ae4] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 725.843797] env[61923]: ERROR nova.compute.manager [instance: 2939d1e5-07fe-48b3-bc87-5f67fa1a7ae4] for vif in network_info: [ 725.843797] env[61923]: ERROR nova.compute.manager [instance: 2939d1e5-07fe-48b3-bc87-5f67fa1a7ae4] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 725.843797] env[61923]: ERROR nova.compute.manager [instance: 2939d1e5-07fe-48b3-bc87-5f67fa1a7ae4] return self._sync_wrapper(fn, *args, **kwargs) [ 725.843797] env[61923]: ERROR nova.compute.manager [instance: 2939d1e5-07fe-48b3-bc87-5f67fa1a7ae4] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 725.843797] env[61923]: ERROR nova.compute.manager [instance: 2939d1e5-07fe-48b3-bc87-5f67fa1a7ae4] self.wait() [ 725.843797] env[61923]: ERROR nova.compute.manager [instance: 2939d1e5-07fe-48b3-bc87-5f67fa1a7ae4] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 725.843797] env[61923]: ERROR nova.compute.manager [instance: 2939d1e5-07fe-48b3-bc87-5f67fa1a7ae4] self[:] = self._gt.wait() [ 725.843797] env[61923]: ERROR nova.compute.manager [instance: 2939d1e5-07fe-48b3-bc87-5f67fa1a7ae4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 725.843797] env[61923]: ERROR nova.compute.manager [instance: 2939d1e5-07fe-48b3-bc87-5f67fa1a7ae4] return self._exit_event.wait() [ 725.843797] env[61923]: ERROR nova.compute.manager [instance: 2939d1e5-07fe-48b3-bc87-5f67fa1a7ae4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 725.844355] env[61923]: ERROR nova.compute.manager [instance: 2939d1e5-07fe-48b3-bc87-5f67fa1a7ae4] result = hub.switch() [ 725.844355] env[61923]: ERROR nova.compute.manager [instance: 2939d1e5-07fe-48b3-bc87-5f67fa1a7ae4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 725.844355] env[61923]: ERROR nova.compute.manager [instance: 2939d1e5-07fe-48b3-bc87-5f67fa1a7ae4] return self.greenlet.switch() [ 725.844355] env[61923]: ERROR nova.compute.manager [instance: 2939d1e5-07fe-48b3-bc87-5f67fa1a7ae4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 725.844355] env[61923]: ERROR nova.compute.manager [instance: 2939d1e5-07fe-48b3-bc87-5f67fa1a7ae4] result = function(*args, **kwargs) [ 725.844355] env[61923]: ERROR nova.compute.manager [instance: 2939d1e5-07fe-48b3-bc87-5f67fa1a7ae4] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 725.844355] env[61923]: ERROR nova.compute.manager [instance: 2939d1e5-07fe-48b3-bc87-5f67fa1a7ae4] return func(*args, **kwargs) [ 725.844355] env[61923]: ERROR nova.compute.manager [instance: 2939d1e5-07fe-48b3-bc87-5f67fa1a7ae4] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 725.844355] env[61923]: ERROR nova.compute.manager [instance: 2939d1e5-07fe-48b3-bc87-5f67fa1a7ae4] raise e [ 725.844355] env[61923]: ERROR nova.compute.manager [instance: 2939d1e5-07fe-48b3-bc87-5f67fa1a7ae4] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 725.844355] env[61923]: ERROR nova.compute.manager [instance: 2939d1e5-07fe-48b3-bc87-5f67fa1a7ae4] nwinfo = self.network_api.allocate_for_instance( [ 725.844355] env[61923]: ERROR nova.compute.manager [instance: 2939d1e5-07fe-48b3-bc87-5f67fa1a7ae4] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 725.844355] env[61923]: ERROR nova.compute.manager [instance: 2939d1e5-07fe-48b3-bc87-5f67fa1a7ae4] created_port_ids = self._update_ports_for_instance( [ 725.844717] env[61923]: ERROR nova.compute.manager [instance: 2939d1e5-07fe-48b3-bc87-5f67fa1a7ae4] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 725.844717] env[61923]: ERROR nova.compute.manager [instance: 2939d1e5-07fe-48b3-bc87-5f67fa1a7ae4] with excutils.save_and_reraise_exception(): [ 725.844717] env[61923]: ERROR nova.compute.manager [instance: 2939d1e5-07fe-48b3-bc87-5f67fa1a7ae4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 725.844717] env[61923]: ERROR nova.compute.manager [instance: 2939d1e5-07fe-48b3-bc87-5f67fa1a7ae4] self.force_reraise() [ 725.844717] env[61923]: ERROR nova.compute.manager [instance: 2939d1e5-07fe-48b3-bc87-5f67fa1a7ae4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 725.844717] env[61923]: ERROR nova.compute.manager [instance: 2939d1e5-07fe-48b3-bc87-5f67fa1a7ae4] raise self.value [ 725.844717] env[61923]: ERROR nova.compute.manager [instance: 2939d1e5-07fe-48b3-bc87-5f67fa1a7ae4] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 725.844717] env[61923]: ERROR nova.compute.manager [instance: 2939d1e5-07fe-48b3-bc87-5f67fa1a7ae4] updated_port = self._update_port( [ 725.844717] env[61923]: ERROR nova.compute.manager [instance: 2939d1e5-07fe-48b3-bc87-5f67fa1a7ae4] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 725.844717] env[61923]: ERROR nova.compute.manager [instance: 2939d1e5-07fe-48b3-bc87-5f67fa1a7ae4] _ensure_no_port_binding_failure(port) [ 725.844717] env[61923]: ERROR nova.compute.manager [instance: 2939d1e5-07fe-48b3-bc87-5f67fa1a7ae4] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 725.844717] env[61923]: ERROR nova.compute.manager [instance: 2939d1e5-07fe-48b3-bc87-5f67fa1a7ae4] raise exception.PortBindingFailed(port_id=port['id']) [ 725.845112] env[61923]: ERROR nova.compute.manager [instance: 2939d1e5-07fe-48b3-bc87-5f67fa1a7ae4] nova.exception.PortBindingFailed: Binding failed for port cb32985d-46e9-42d3-a1cd-7e7620873e5e, please check neutron logs for more information. [ 725.845112] env[61923]: ERROR nova.compute.manager [instance: 2939d1e5-07fe-48b3-bc87-5f67fa1a7ae4] [ 725.845112] env[61923]: DEBUG nova.compute.utils [None req-24801cc4-3974-42da-978f-a22cb8f70af0 tempest-SecurityGroupsTestJSON-509490211 tempest-SecurityGroupsTestJSON-509490211-project-member] [instance: 2939d1e5-07fe-48b3-bc87-5f67fa1a7ae4] Binding failed for port cb32985d-46e9-42d3-a1cd-7e7620873e5e, please check neutron logs for more information. {{(pid=61923) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 725.845112] env[61923]: DEBUG oslo_concurrency.lockutils [None req-13202777-73e8-40d3-a6ad-7bb49e61267a tempest-ImagesNegativeTestJSON-1202128281 tempest-ImagesNegativeTestJSON-1202128281-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.930s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 725.846821] env[61923]: INFO nova.compute.claims [None req-13202777-73e8-40d3-a6ad-7bb49e61267a tempest-ImagesNegativeTestJSON-1202128281 tempest-ImagesNegativeTestJSON-1202128281-project-member] [instance: f7e1d517-3e39-4381-83be-0152368b8120] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 725.850143] env[61923]: DEBUG nova.compute.manager [None req-24801cc4-3974-42da-978f-a22cb8f70af0 tempest-SecurityGroupsTestJSON-509490211 tempest-SecurityGroupsTestJSON-509490211-project-member] [instance: 2939d1e5-07fe-48b3-bc87-5f67fa1a7ae4] Build of instance 2939d1e5-07fe-48b3-bc87-5f67fa1a7ae4 was re-scheduled: Binding failed for port cb32985d-46e9-42d3-a1cd-7e7620873e5e, please check neutron logs for more information. {{(pid=61923) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 725.850685] env[61923]: DEBUG nova.compute.manager [None req-24801cc4-3974-42da-978f-a22cb8f70af0 tempest-SecurityGroupsTestJSON-509490211 tempest-SecurityGroupsTestJSON-509490211-project-member] [instance: 2939d1e5-07fe-48b3-bc87-5f67fa1a7ae4] Unplugging VIFs for instance {{(pid=61923) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 725.851009] env[61923]: DEBUG oslo_concurrency.lockutils [None req-24801cc4-3974-42da-978f-a22cb8f70af0 tempest-SecurityGroupsTestJSON-509490211 tempest-SecurityGroupsTestJSON-509490211-project-member] Acquiring lock "refresh_cache-2939d1e5-07fe-48b3-bc87-5f67fa1a7ae4" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 725.851178] env[61923]: DEBUG oslo_concurrency.lockutils [None req-24801cc4-3974-42da-978f-a22cb8f70af0 tempest-SecurityGroupsTestJSON-509490211 tempest-SecurityGroupsTestJSON-509490211-project-member] Acquired lock "refresh_cache-2939d1e5-07fe-48b3-bc87-5f67fa1a7ae4" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 725.851371] env[61923]: DEBUG nova.network.neutron [None req-24801cc4-3974-42da-978f-a22cb8f70af0 tempest-SecurityGroupsTestJSON-509490211 tempest-SecurityGroupsTestJSON-509490211-project-member] [instance: 2939d1e5-07fe-48b3-bc87-5f67fa1a7ae4] Building network info cache for instance {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 725.869315] env[61923]: DEBUG nova.compute.manager [None req-40b011d4-39ab-4684-a124-b3ba87d489df tempest-ServerPasswordTestJSON-595643156 tempest-ServerPasswordTestJSON-595643156-project-member] [instance: 7c0341f9-7401-42ee-9512-afd832c76940] Start spawning the instance on the hypervisor. {{(pid=61923) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 725.914387] env[61923]: DEBUG nova.virt.hardware [None req-40b011d4-39ab-4684-a124-b3ba87d489df tempest-ServerPasswordTestJSON-595643156 tempest-ServerPasswordTestJSON-595643156-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-29T20:07:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-29T20:07:35Z,direct_url=,disk_format='vmdk',id=0f153f63-ae0a-45b1-b7f5-7f9b673c947f,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='4e7b5e3b3c3443c8bd5c70f8a15739f5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-29T20:07:36Z,virtual_size=,visibility=), allow threads: False {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 725.914387] env[61923]: DEBUG nova.virt.hardware [None req-40b011d4-39ab-4684-a124-b3ba87d489df tempest-ServerPasswordTestJSON-595643156 tempest-ServerPasswordTestJSON-595643156-project-member] Flavor limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 725.914387] env[61923]: DEBUG nova.virt.hardware [None req-40b011d4-39ab-4684-a124-b3ba87d489df tempest-ServerPasswordTestJSON-595643156 tempest-ServerPasswordTestJSON-595643156-project-member] Image limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 725.914551] env[61923]: DEBUG nova.virt.hardware [None req-40b011d4-39ab-4684-a124-b3ba87d489df tempest-ServerPasswordTestJSON-595643156 tempest-ServerPasswordTestJSON-595643156-project-member] Flavor pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 725.914551] env[61923]: DEBUG nova.virt.hardware [None req-40b011d4-39ab-4684-a124-b3ba87d489df tempest-ServerPasswordTestJSON-595643156 tempest-ServerPasswordTestJSON-595643156-project-member] Image pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 725.914551] env[61923]: DEBUG nova.virt.hardware [None req-40b011d4-39ab-4684-a124-b3ba87d489df tempest-ServerPasswordTestJSON-595643156 tempest-ServerPasswordTestJSON-595643156-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 725.914645] env[61923]: DEBUG nova.virt.hardware [None req-40b011d4-39ab-4684-a124-b3ba87d489df tempest-ServerPasswordTestJSON-595643156 tempest-ServerPasswordTestJSON-595643156-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 725.914720] env[61923]: DEBUG nova.virt.hardware [None req-40b011d4-39ab-4684-a124-b3ba87d489df tempest-ServerPasswordTestJSON-595643156 tempest-ServerPasswordTestJSON-595643156-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 725.914876] env[61923]: DEBUG nova.virt.hardware [None req-40b011d4-39ab-4684-a124-b3ba87d489df tempest-ServerPasswordTestJSON-595643156 tempest-ServerPasswordTestJSON-595643156-project-member] Got 1 possible topologies {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 725.915180] env[61923]: DEBUG nova.virt.hardware [None req-40b011d4-39ab-4684-a124-b3ba87d489df tempest-ServerPasswordTestJSON-595643156 tempest-ServerPasswordTestJSON-595643156-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 725.915253] env[61923]: DEBUG nova.virt.hardware [None req-40b011d4-39ab-4684-a124-b3ba87d489df tempest-ServerPasswordTestJSON-595643156 tempest-ServerPasswordTestJSON-595643156-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 725.918687] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4d291c8-f5a3-4752-8b75-0a06e36ad4d6 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.925331] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ab90bab-5991-4909-8a06-7e89cbbc6d22 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.135456] env[61923]: DEBUG oslo_concurrency.lockutils [None req-5138192d-6349-432b-b7de-64ec5d391f8d tempest-ServerActionsTestOtherB-987045973 tempest-ServerActionsTestOtherB-987045973-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 726.204693] env[61923]: DEBUG nova.compute.manager [req-38284c6a-a394-489f-90cf-5522e145f16b req-3e65c3e6-f7b4-4217-a2bc-5aaddadfb933 service nova] [instance: 7c0341f9-7401-42ee-9512-afd832c76940] Received event network-changed-e84bf904-db99-4abc-90aa-4be30a8f32e8 {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 726.205041] env[61923]: DEBUG nova.compute.manager [req-38284c6a-a394-489f-90cf-5522e145f16b req-3e65c3e6-f7b4-4217-a2bc-5aaddadfb933 service nova] [instance: 7c0341f9-7401-42ee-9512-afd832c76940] Refreshing instance network info cache due to event network-changed-e84bf904-db99-4abc-90aa-4be30a8f32e8. {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 726.205117] env[61923]: DEBUG oslo_concurrency.lockutils [req-38284c6a-a394-489f-90cf-5522e145f16b req-3e65c3e6-f7b4-4217-a2bc-5aaddadfb933 service nova] Acquiring lock "refresh_cache-7c0341f9-7401-42ee-9512-afd832c76940" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 726.205258] env[61923]: DEBUG oslo_concurrency.lockutils [req-38284c6a-a394-489f-90cf-5522e145f16b req-3e65c3e6-f7b4-4217-a2bc-5aaddadfb933 service nova] Acquired lock "refresh_cache-7c0341f9-7401-42ee-9512-afd832c76940" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 726.205393] env[61923]: DEBUG nova.network.neutron [req-38284c6a-a394-489f-90cf-5522e145f16b req-3e65c3e6-f7b4-4217-a2bc-5aaddadfb933 service nova] [instance: 7c0341f9-7401-42ee-9512-afd832c76940] Refreshing network info cache for port e84bf904-db99-4abc-90aa-4be30a8f32e8 {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 726.379910] env[61923]: DEBUG nova.network.neutron [None req-24801cc4-3974-42da-978f-a22cb8f70af0 tempest-SecurityGroupsTestJSON-509490211 tempest-SecurityGroupsTestJSON-509490211-project-member] [instance: 2939d1e5-07fe-48b3-bc87-5f67fa1a7ae4] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 726.431787] env[61923]: ERROR nova.compute.manager [None req-40b011d4-39ab-4684-a124-b3ba87d489df tempest-ServerPasswordTestJSON-595643156 tempest-ServerPasswordTestJSON-595643156-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port e84bf904-db99-4abc-90aa-4be30a8f32e8, please check neutron logs for more information. [ 726.431787] env[61923]: ERROR nova.compute.manager Traceback (most recent call last): [ 726.431787] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 726.431787] env[61923]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 726.431787] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 726.431787] env[61923]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 726.431787] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 726.431787] env[61923]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 726.431787] env[61923]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 726.431787] env[61923]: ERROR nova.compute.manager self.force_reraise() [ 726.431787] env[61923]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 726.431787] env[61923]: ERROR nova.compute.manager raise self.value [ 726.431787] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 726.431787] env[61923]: ERROR nova.compute.manager updated_port = self._update_port( [ 726.431787] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 726.431787] env[61923]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 726.432377] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 726.432377] env[61923]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 726.432377] env[61923]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port e84bf904-db99-4abc-90aa-4be30a8f32e8, please check neutron logs for more information. [ 726.432377] env[61923]: ERROR nova.compute.manager [ 726.434889] env[61923]: Traceback (most recent call last): [ 726.434889] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 726.434889] env[61923]: listener.cb(fileno) [ 726.434889] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 726.434889] env[61923]: result = function(*args, **kwargs) [ 726.434889] env[61923]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 726.434889] env[61923]: return func(*args, **kwargs) [ 726.434889] env[61923]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 726.434889] env[61923]: raise e [ 726.434889] env[61923]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 726.434889] env[61923]: nwinfo = self.network_api.allocate_for_instance( [ 726.434889] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 726.434889] env[61923]: created_port_ids = self._update_ports_for_instance( [ 726.434889] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 726.434889] env[61923]: with excutils.save_and_reraise_exception(): [ 726.434889] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 726.434889] env[61923]: self.force_reraise() [ 726.434889] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 726.434889] env[61923]: raise self.value [ 726.434889] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 726.434889] env[61923]: updated_port = self._update_port( [ 726.434889] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 726.434889] env[61923]: _ensure_no_port_binding_failure(port) [ 726.434889] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 726.434889] env[61923]: raise exception.PortBindingFailed(port_id=port['id']) [ 726.434889] env[61923]: nova.exception.PortBindingFailed: Binding failed for port e84bf904-db99-4abc-90aa-4be30a8f32e8, please check neutron logs for more information. [ 726.434889] env[61923]: Removing descriptor: 17 [ 726.435731] env[61923]: ERROR nova.compute.manager [None req-40b011d4-39ab-4684-a124-b3ba87d489df tempest-ServerPasswordTestJSON-595643156 tempest-ServerPasswordTestJSON-595643156-project-member] [instance: 7c0341f9-7401-42ee-9512-afd832c76940] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port e84bf904-db99-4abc-90aa-4be30a8f32e8, please check neutron logs for more information. [ 726.435731] env[61923]: ERROR nova.compute.manager [instance: 7c0341f9-7401-42ee-9512-afd832c76940] Traceback (most recent call last): [ 726.435731] env[61923]: ERROR nova.compute.manager [instance: 7c0341f9-7401-42ee-9512-afd832c76940] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 726.435731] env[61923]: ERROR nova.compute.manager [instance: 7c0341f9-7401-42ee-9512-afd832c76940] yield resources [ 726.435731] env[61923]: ERROR nova.compute.manager [instance: 7c0341f9-7401-42ee-9512-afd832c76940] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 726.435731] env[61923]: ERROR nova.compute.manager [instance: 7c0341f9-7401-42ee-9512-afd832c76940] self.driver.spawn(context, instance, image_meta, [ 726.435731] env[61923]: ERROR nova.compute.manager [instance: 7c0341f9-7401-42ee-9512-afd832c76940] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 726.435731] env[61923]: ERROR nova.compute.manager [instance: 7c0341f9-7401-42ee-9512-afd832c76940] self._vmops.spawn(context, instance, image_meta, injected_files, [ 726.435731] env[61923]: ERROR nova.compute.manager [instance: 7c0341f9-7401-42ee-9512-afd832c76940] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 726.435731] env[61923]: ERROR nova.compute.manager [instance: 7c0341f9-7401-42ee-9512-afd832c76940] vm_ref = self.build_virtual_machine(instance, [ 726.435731] env[61923]: ERROR nova.compute.manager [instance: 7c0341f9-7401-42ee-9512-afd832c76940] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 726.436124] env[61923]: ERROR nova.compute.manager [instance: 7c0341f9-7401-42ee-9512-afd832c76940] vif_infos = vmwarevif.get_vif_info(self._session, [ 726.436124] env[61923]: ERROR nova.compute.manager [instance: 7c0341f9-7401-42ee-9512-afd832c76940] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 726.436124] env[61923]: ERROR nova.compute.manager [instance: 7c0341f9-7401-42ee-9512-afd832c76940] for vif in network_info: [ 726.436124] env[61923]: ERROR nova.compute.manager [instance: 7c0341f9-7401-42ee-9512-afd832c76940] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 726.436124] env[61923]: ERROR nova.compute.manager [instance: 7c0341f9-7401-42ee-9512-afd832c76940] return self._sync_wrapper(fn, *args, **kwargs) [ 726.436124] env[61923]: ERROR nova.compute.manager [instance: 7c0341f9-7401-42ee-9512-afd832c76940] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 726.436124] env[61923]: ERROR nova.compute.manager [instance: 7c0341f9-7401-42ee-9512-afd832c76940] self.wait() [ 726.436124] env[61923]: ERROR nova.compute.manager [instance: 7c0341f9-7401-42ee-9512-afd832c76940] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 726.436124] env[61923]: ERROR nova.compute.manager [instance: 7c0341f9-7401-42ee-9512-afd832c76940] self[:] = self._gt.wait() [ 726.436124] env[61923]: ERROR nova.compute.manager [instance: 7c0341f9-7401-42ee-9512-afd832c76940] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 726.436124] env[61923]: ERROR nova.compute.manager [instance: 7c0341f9-7401-42ee-9512-afd832c76940] return self._exit_event.wait() [ 726.436124] env[61923]: ERROR nova.compute.manager [instance: 7c0341f9-7401-42ee-9512-afd832c76940] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 726.436124] env[61923]: ERROR nova.compute.manager [instance: 7c0341f9-7401-42ee-9512-afd832c76940] result = hub.switch() [ 726.436526] env[61923]: ERROR nova.compute.manager [instance: 7c0341f9-7401-42ee-9512-afd832c76940] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 726.436526] env[61923]: ERROR nova.compute.manager [instance: 7c0341f9-7401-42ee-9512-afd832c76940] return self.greenlet.switch() [ 726.436526] env[61923]: ERROR nova.compute.manager [instance: 7c0341f9-7401-42ee-9512-afd832c76940] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 726.436526] env[61923]: ERROR nova.compute.manager [instance: 7c0341f9-7401-42ee-9512-afd832c76940] result = function(*args, **kwargs) [ 726.436526] env[61923]: ERROR nova.compute.manager [instance: 7c0341f9-7401-42ee-9512-afd832c76940] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 726.436526] env[61923]: ERROR nova.compute.manager [instance: 7c0341f9-7401-42ee-9512-afd832c76940] return func(*args, **kwargs) [ 726.436526] env[61923]: ERROR nova.compute.manager [instance: 7c0341f9-7401-42ee-9512-afd832c76940] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 726.436526] env[61923]: ERROR nova.compute.manager [instance: 7c0341f9-7401-42ee-9512-afd832c76940] raise e [ 726.436526] env[61923]: ERROR nova.compute.manager [instance: 7c0341f9-7401-42ee-9512-afd832c76940] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 726.436526] env[61923]: ERROR nova.compute.manager [instance: 7c0341f9-7401-42ee-9512-afd832c76940] nwinfo = self.network_api.allocate_for_instance( [ 726.436526] env[61923]: ERROR nova.compute.manager [instance: 7c0341f9-7401-42ee-9512-afd832c76940] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 726.436526] env[61923]: ERROR nova.compute.manager [instance: 7c0341f9-7401-42ee-9512-afd832c76940] created_port_ids = self._update_ports_for_instance( [ 726.436526] env[61923]: ERROR nova.compute.manager [instance: 7c0341f9-7401-42ee-9512-afd832c76940] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 726.437013] env[61923]: ERROR nova.compute.manager [instance: 7c0341f9-7401-42ee-9512-afd832c76940] with excutils.save_and_reraise_exception(): [ 726.437013] env[61923]: ERROR nova.compute.manager [instance: 7c0341f9-7401-42ee-9512-afd832c76940] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 726.437013] env[61923]: ERROR nova.compute.manager [instance: 7c0341f9-7401-42ee-9512-afd832c76940] self.force_reraise() [ 726.437013] env[61923]: ERROR nova.compute.manager [instance: 7c0341f9-7401-42ee-9512-afd832c76940] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 726.437013] env[61923]: ERROR nova.compute.manager [instance: 7c0341f9-7401-42ee-9512-afd832c76940] raise self.value [ 726.437013] env[61923]: ERROR nova.compute.manager [instance: 7c0341f9-7401-42ee-9512-afd832c76940] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 726.437013] env[61923]: ERROR nova.compute.manager [instance: 7c0341f9-7401-42ee-9512-afd832c76940] updated_port = self._update_port( [ 726.437013] env[61923]: ERROR nova.compute.manager [instance: 7c0341f9-7401-42ee-9512-afd832c76940] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 726.437013] env[61923]: ERROR nova.compute.manager [instance: 7c0341f9-7401-42ee-9512-afd832c76940] _ensure_no_port_binding_failure(port) [ 726.437013] env[61923]: ERROR nova.compute.manager [instance: 7c0341f9-7401-42ee-9512-afd832c76940] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 726.437013] env[61923]: ERROR nova.compute.manager [instance: 7c0341f9-7401-42ee-9512-afd832c76940] raise exception.PortBindingFailed(port_id=port['id']) [ 726.437013] env[61923]: ERROR nova.compute.manager [instance: 7c0341f9-7401-42ee-9512-afd832c76940] nova.exception.PortBindingFailed: Binding failed for port e84bf904-db99-4abc-90aa-4be30a8f32e8, please check neutron logs for more information. [ 726.437013] env[61923]: ERROR nova.compute.manager [instance: 7c0341f9-7401-42ee-9512-afd832c76940] [ 726.438139] env[61923]: INFO nova.compute.manager [None req-40b011d4-39ab-4684-a124-b3ba87d489df tempest-ServerPasswordTestJSON-595643156 tempest-ServerPasswordTestJSON-595643156-project-member] [instance: 7c0341f9-7401-42ee-9512-afd832c76940] Terminating instance [ 726.442222] env[61923]: DEBUG oslo_concurrency.lockutils [None req-40b011d4-39ab-4684-a124-b3ba87d489df tempest-ServerPasswordTestJSON-595643156 tempest-ServerPasswordTestJSON-595643156-project-member] Acquiring lock "refresh_cache-7c0341f9-7401-42ee-9512-afd832c76940" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 726.499244] env[61923]: DEBUG nova.network.neutron [None req-24801cc4-3974-42da-978f-a22cb8f70af0 tempest-SecurityGroupsTestJSON-509490211 tempest-SecurityGroupsTestJSON-509490211-project-member] [instance: 2939d1e5-07fe-48b3-bc87-5f67fa1a7ae4] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 726.738881] env[61923]: DEBUG nova.network.neutron [req-38284c6a-a394-489f-90cf-5522e145f16b req-3e65c3e6-f7b4-4217-a2bc-5aaddadfb933 service nova] [instance: 7c0341f9-7401-42ee-9512-afd832c76940] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 726.873312] env[61923]: DEBUG nova.network.neutron [req-38284c6a-a394-489f-90cf-5522e145f16b req-3e65c3e6-f7b4-4217-a2bc-5aaddadfb933 service nova] [instance: 7c0341f9-7401-42ee-9512-afd832c76940] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 727.003744] env[61923]: DEBUG oslo_concurrency.lockutils [None req-24801cc4-3974-42da-978f-a22cb8f70af0 tempest-SecurityGroupsTestJSON-509490211 tempest-SecurityGroupsTestJSON-509490211-project-member] Releasing lock "refresh_cache-2939d1e5-07fe-48b3-bc87-5f67fa1a7ae4" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 727.003744] env[61923]: DEBUG nova.compute.manager [None req-24801cc4-3974-42da-978f-a22cb8f70af0 tempest-SecurityGroupsTestJSON-509490211 tempest-SecurityGroupsTestJSON-509490211-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61923) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 727.003744] env[61923]: DEBUG nova.compute.manager [None req-24801cc4-3974-42da-978f-a22cb8f70af0 tempest-SecurityGroupsTestJSON-509490211 tempest-SecurityGroupsTestJSON-509490211-project-member] [instance: 2939d1e5-07fe-48b3-bc87-5f67fa1a7ae4] Deallocating network for instance {{(pid=61923) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 727.003744] env[61923]: DEBUG nova.network.neutron [None req-24801cc4-3974-42da-978f-a22cb8f70af0 tempest-SecurityGroupsTestJSON-509490211 tempest-SecurityGroupsTestJSON-509490211-project-member] [instance: 2939d1e5-07fe-48b3-bc87-5f67fa1a7ae4] deallocate_for_instance() {{(pid=61923) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 727.029389] env[61923]: DEBUG nova.network.neutron [None req-24801cc4-3974-42da-978f-a22cb8f70af0 tempest-SecurityGroupsTestJSON-509490211 tempest-SecurityGroupsTestJSON-509490211-project-member] [instance: 2939d1e5-07fe-48b3-bc87-5f67fa1a7ae4] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 727.245855] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f26184c-3f3a-4462-946b-05c54b053f38 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.254021] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85e5b2bf-a091-4c96-819b-b86a14169d69 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.284657] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b192c60-4877-4b3d-ba75-b8bbf35944cf {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.292373] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa7f597f-5528-4434-8525-62a1d95daebb {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.306223] env[61923]: DEBUG nova.compute.provider_tree [None req-13202777-73e8-40d3-a6ad-7bb49e61267a tempest-ImagesNegativeTestJSON-1202128281 tempest-ImagesNegativeTestJSON-1202128281-project-member] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 727.376702] env[61923]: DEBUG oslo_concurrency.lockutils [req-38284c6a-a394-489f-90cf-5522e145f16b req-3e65c3e6-f7b4-4217-a2bc-5aaddadfb933 service nova] Releasing lock "refresh_cache-7c0341f9-7401-42ee-9512-afd832c76940" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 727.377255] env[61923]: DEBUG oslo_concurrency.lockutils [None req-40b011d4-39ab-4684-a124-b3ba87d489df tempest-ServerPasswordTestJSON-595643156 tempest-ServerPasswordTestJSON-595643156-project-member] Acquired lock "refresh_cache-7c0341f9-7401-42ee-9512-afd832c76940" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 727.377478] env[61923]: DEBUG nova.network.neutron [None req-40b011d4-39ab-4684-a124-b3ba87d489df tempest-ServerPasswordTestJSON-595643156 tempest-ServerPasswordTestJSON-595643156-project-member] [instance: 7c0341f9-7401-42ee-9512-afd832c76940] Building network info cache for instance {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 727.532015] env[61923]: DEBUG nova.network.neutron [None req-24801cc4-3974-42da-978f-a22cb8f70af0 tempest-SecurityGroupsTestJSON-509490211 tempest-SecurityGroupsTestJSON-509490211-project-member] [instance: 2939d1e5-07fe-48b3-bc87-5f67fa1a7ae4] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 727.809626] env[61923]: DEBUG nova.scheduler.client.report [None req-13202777-73e8-40d3-a6ad-7bb49e61267a tempest-ImagesNegativeTestJSON-1202128281 tempest-ImagesNegativeTestJSON-1202128281-project-member] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 727.896359] env[61923]: DEBUG nova.network.neutron [None req-40b011d4-39ab-4684-a124-b3ba87d489df tempest-ServerPasswordTestJSON-595643156 tempest-ServerPasswordTestJSON-595643156-project-member] [instance: 7c0341f9-7401-42ee-9512-afd832c76940] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 727.995911] env[61923]: DEBUG nova.network.neutron [None req-40b011d4-39ab-4684-a124-b3ba87d489df tempest-ServerPasswordTestJSON-595643156 tempest-ServerPasswordTestJSON-595643156-project-member] [instance: 7c0341f9-7401-42ee-9512-afd832c76940] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 728.035047] env[61923]: INFO nova.compute.manager [None req-24801cc4-3974-42da-978f-a22cb8f70af0 tempest-SecurityGroupsTestJSON-509490211 tempest-SecurityGroupsTestJSON-509490211-project-member] [instance: 2939d1e5-07fe-48b3-bc87-5f67fa1a7ae4] Took 1.03 seconds to deallocate network for instance. [ 728.271296] env[61923]: DEBUG nova.compute.manager [req-f898c097-3a34-4290-9746-c88a6f0173ba req-95b2c2ce-a590-4e82-aa9e-0ad6c48bc1da service nova] [instance: 7c0341f9-7401-42ee-9512-afd832c76940] Received event network-vif-deleted-e84bf904-db99-4abc-90aa-4be30a8f32e8 {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 728.315124] env[61923]: DEBUG oslo_concurrency.lockutils [None req-13202777-73e8-40d3-a6ad-7bb49e61267a tempest-ImagesNegativeTestJSON-1202128281 tempest-ImagesNegativeTestJSON-1202128281-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.470s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 728.315702] env[61923]: DEBUG nova.compute.manager [None req-13202777-73e8-40d3-a6ad-7bb49e61267a tempest-ImagesNegativeTestJSON-1202128281 tempest-ImagesNegativeTestJSON-1202128281-project-member] [instance: f7e1d517-3e39-4381-83be-0152368b8120] Start building networks asynchronously for instance. {{(pid=61923) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 728.319081] env[61923]: DEBUG oslo_concurrency.lockutils [None req-f7b52f4c-b2e3-49f8-b056-b7254f72bb20 tempest-AttachInterfacesUnderV243Test-498638506 tempest-AttachInterfacesUnderV243Test-498638506-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.431s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 728.320438] env[61923]: INFO nova.compute.claims [None req-f7b52f4c-b2e3-49f8-b056-b7254f72bb20 tempest-AttachInterfacesUnderV243Test-498638506 tempest-AttachInterfacesUnderV243Test-498638506-project-member] [instance: 0eab2200-2528-4a89-821a-cdd6a3c99297] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 728.499177] env[61923]: DEBUG oslo_concurrency.lockutils [None req-40b011d4-39ab-4684-a124-b3ba87d489df tempest-ServerPasswordTestJSON-595643156 tempest-ServerPasswordTestJSON-595643156-project-member] Releasing lock "refresh_cache-7c0341f9-7401-42ee-9512-afd832c76940" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 728.500049] env[61923]: DEBUG nova.compute.manager [None req-40b011d4-39ab-4684-a124-b3ba87d489df tempest-ServerPasswordTestJSON-595643156 tempest-ServerPasswordTestJSON-595643156-project-member] [instance: 7c0341f9-7401-42ee-9512-afd832c76940] Start destroying the instance on the hypervisor. {{(pid=61923) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 728.500049] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-40b011d4-39ab-4684-a124-b3ba87d489df tempest-ServerPasswordTestJSON-595643156 tempest-ServerPasswordTestJSON-595643156-project-member] [instance: 7c0341f9-7401-42ee-9512-afd832c76940] Destroying instance {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 728.500489] env[61923]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-cd417d1d-b756-4bd9-b804-c263555dafd1 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.509873] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb20746f-b058-4ef3-8387-9fd26be0afc0 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.532361] env[61923]: WARNING nova.virt.vmwareapi.vmops [None req-40b011d4-39ab-4684-a124-b3ba87d489df tempest-ServerPasswordTestJSON-595643156 tempest-ServerPasswordTestJSON-595643156-project-member] [instance: 7c0341f9-7401-42ee-9512-afd832c76940] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 7c0341f9-7401-42ee-9512-afd832c76940 could not be found. [ 728.532361] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-40b011d4-39ab-4684-a124-b3ba87d489df tempest-ServerPasswordTestJSON-595643156 tempest-ServerPasswordTestJSON-595643156-project-member] [instance: 7c0341f9-7401-42ee-9512-afd832c76940] Instance destroyed {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 728.532560] env[61923]: INFO nova.compute.manager [None req-40b011d4-39ab-4684-a124-b3ba87d489df tempest-ServerPasswordTestJSON-595643156 tempest-ServerPasswordTestJSON-595643156-project-member] [instance: 7c0341f9-7401-42ee-9512-afd832c76940] Took 0.03 seconds to destroy the instance on the hypervisor. [ 728.532700] env[61923]: DEBUG oslo.service.loopingcall [None req-40b011d4-39ab-4684-a124-b3ba87d489df tempest-ServerPasswordTestJSON-595643156 tempest-ServerPasswordTestJSON-595643156-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61923) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 728.533273] env[61923]: DEBUG nova.compute.manager [-] [instance: 7c0341f9-7401-42ee-9512-afd832c76940] Deallocating network for instance {{(pid=61923) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 728.533373] env[61923]: DEBUG nova.network.neutron [-] [instance: 7c0341f9-7401-42ee-9512-afd832c76940] deallocate_for_instance() {{(pid=61923) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 728.551819] env[61923]: DEBUG nova.network.neutron [-] [instance: 7c0341f9-7401-42ee-9512-afd832c76940] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 728.820546] env[61923]: DEBUG nova.compute.utils [None req-13202777-73e8-40d3-a6ad-7bb49e61267a tempest-ImagesNegativeTestJSON-1202128281 tempest-ImagesNegativeTestJSON-1202128281-project-member] Using /dev/sd instead of None {{(pid=61923) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 728.821971] env[61923]: DEBUG nova.compute.manager [None req-13202777-73e8-40d3-a6ad-7bb49e61267a tempest-ImagesNegativeTestJSON-1202128281 tempest-ImagesNegativeTestJSON-1202128281-project-member] [instance: f7e1d517-3e39-4381-83be-0152368b8120] Allocating IP information in the background. {{(pid=61923) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 728.822157] env[61923]: DEBUG nova.network.neutron [None req-13202777-73e8-40d3-a6ad-7bb49e61267a tempest-ImagesNegativeTestJSON-1202128281 tempest-ImagesNegativeTestJSON-1202128281-project-member] [instance: f7e1d517-3e39-4381-83be-0152368b8120] allocate_for_instance() {{(pid=61923) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 728.865791] env[61923]: DEBUG nova.policy [None req-13202777-73e8-40d3-a6ad-7bb49e61267a tempest-ImagesNegativeTestJSON-1202128281 tempest-ImagesNegativeTestJSON-1202128281-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '388917d2958b4545a0e72d9b3e8d4e79', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '590f198ea30e4ce1acbf1911cd251e8b', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61923) authorize /opt/stack/nova/nova/policy.py:201}} [ 729.053766] env[61923]: DEBUG nova.network.neutron [-] [instance: 7c0341f9-7401-42ee-9512-afd832c76940] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 729.076060] env[61923]: INFO nova.scheduler.client.report [None req-24801cc4-3974-42da-978f-a22cb8f70af0 tempest-SecurityGroupsTestJSON-509490211 tempest-SecurityGroupsTestJSON-509490211-project-member] Deleted allocations for instance 2939d1e5-07fe-48b3-bc87-5f67fa1a7ae4 [ 729.263922] env[61923]: DEBUG nova.network.neutron [None req-13202777-73e8-40d3-a6ad-7bb49e61267a tempest-ImagesNegativeTestJSON-1202128281 tempest-ImagesNegativeTestJSON-1202128281-project-member] [instance: f7e1d517-3e39-4381-83be-0152368b8120] Successfully created port: 66c485f3-bd28-41e5-a69c-fc82fa725d19 {{(pid=61923) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 729.330023] env[61923]: DEBUG nova.compute.manager [None req-13202777-73e8-40d3-a6ad-7bb49e61267a tempest-ImagesNegativeTestJSON-1202128281 tempest-ImagesNegativeTestJSON-1202128281-project-member] [instance: f7e1d517-3e39-4381-83be-0152368b8120] Start building block device mappings for instance. {{(pid=61923) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 729.556168] env[61923]: INFO nova.compute.manager [-] [instance: 7c0341f9-7401-42ee-9512-afd832c76940] Took 1.02 seconds to deallocate network for instance. [ 729.560306] env[61923]: DEBUG nova.compute.claims [None req-40b011d4-39ab-4684-a124-b3ba87d489df tempest-ServerPasswordTestJSON-595643156 tempest-ServerPasswordTestJSON-595643156-project-member] [instance: 7c0341f9-7401-42ee-9512-afd832c76940] Aborting claim: {{(pid=61923) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 729.560511] env[61923]: DEBUG oslo_concurrency.lockutils [None req-40b011d4-39ab-4684-a124-b3ba87d489df tempest-ServerPasswordTestJSON-595643156 tempest-ServerPasswordTestJSON-595643156-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 729.584449] env[61923]: DEBUG oslo_concurrency.lockutils [None req-24801cc4-3974-42da-978f-a22cb8f70af0 tempest-SecurityGroupsTestJSON-509490211 tempest-SecurityGroupsTestJSON-509490211-project-member] Lock "2939d1e5-07fe-48b3-bc87-5f67fa1a7ae4" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 133.496s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 729.629525] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b521d99-827e-4833-a591-de5d9984cf52 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.638496] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b41e1c8-1c15-4380-bc94-07cfa4d0c500 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.676947] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8c95f59-54ce-48ac-805e-a7b744041fb9 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.686021] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ec82ae7-912d-422e-a34c-d6ffedeedd82 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.704260] env[61923]: DEBUG nova.compute.provider_tree [None req-f7b52f4c-b2e3-49f8-b056-b7254f72bb20 tempest-AttachInterfacesUnderV243Test-498638506 tempest-AttachInterfacesUnderV243Test-498638506-project-member] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 730.087224] env[61923]: DEBUG nova.compute.manager [None req-f1ee1de3-0f72-403f-9698-da4536b6f37f tempest-ServerDiagnosticsV248Test-1822646244 tempest-ServerDiagnosticsV248Test-1822646244-project-member] [instance: 655b01ae-76a5-4c09-9245-e629fec8fd4b] Starting instance... {{(pid=61923) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 730.210027] env[61923]: DEBUG nova.scheduler.client.report [None req-f7b52f4c-b2e3-49f8-b056-b7254f72bb20 tempest-AttachInterfacesUnderV243Test-498638506 tempest-AttachInterfacesUnderV243Test-498638506-project-member] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 730.344222] env[61923]: DEBUG nova.compute.manager [None req-13202777-73e8-40d3-a6ad-7bb49e61267a tempest-ImagesNegativeTestJSON-1202128281 tempest-ImagesNegativeTestJSON-1202128281-project-member] [instance: f7e1d517-3e39-4381-83be-0152368b8120] Start spawning the instance on the hypervisor. {{(pid=61923) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 730.380037] env[61923]: DEBUG nova.virt.hardware [None req-13202777-73e8-40d3-a6ad-7bb49e61267a tempest-ImagesNegativeTestJSON-1202128281 tempest-ImagesNegativeTestJSON-1202128281-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-29T20:07:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-29T20:07:35Z,direct_url=,disk_format='vmdk',id=0f153f63-ae0a-45b1-b7f5-7f9b673c947f,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='4e7b5e3b3c3443c8bd5c70f8a15739f5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-29T20:07:36Z,virtual_size=,visibility=), allow threads: False {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 730.380037] env[61923]: DEBUG nova.virt.hardware [None req-13202777-73e8-40d3-a6ad-7bb49e61267a tempest-ImagesNegativeTestJSON-1202128281 tempest-ImagesNegativeTestJSON-1202128281-project-member] Flavor limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 730.380037] env[61923]: DEBUG nova.virt.hardware [None req-13202777-73e8-40d3-a6ad-7bb49e61267a tempest-ImagesNegativeTestJSON-1202128281 tempest-ImagesNegativeTestJSON-1202128281-project-member] Image limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 730.380453] env[61923]: DEBUG nova.virt.hardware [None req-13202777-73e8-40d3-a6ad-7bb49e61267a tempest-ImagesNegativeTestJSON-1202128281 tempest-ImagesNegativeTestJSON-1202128281-project-member] Flavor pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 730.380453] env[61923]: DEBUG nova.virt.hardware [None req-13202777-73e8-40d3-a6ad-7bb49e61267a tempest-ImagesNegativeTestJSON-1202128281 tempest-ImagesNegativeTestJSON-1202128281-project-member] Image pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 730.380453] env[61923]: DEBUG nova.virt.hardware [None req-13202777-73e8-40d3-a6ad-7bb49e61267a tempest-ImagesNegativeTestJSON-1202128281 tempest-ImagesNegativeTestJSON-1202128281-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 730.380453] env[61923]: DEBUG nova.virt.hardware [None req-13202777-73e8-40d3-a6ad-7bb49e61267a tempest-ImagesNegativeTestJSON-1202128281 tempest-ImagesNegativeTestJSON-1202128281-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 730.380453] env[61923]: DEBUG nova.virt.hardware [None req-13202777-73e8-40d3-a6ad-7bb49e61267a tempest-ImagesNegativeTestJSON-1202128281 tempest-ImagesNegativeTestJSON-1202128281-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 730.380745] env[61923]: DEBUG nova.virt.hardware [None req-13202777-73e8-40d3-a6ad-7bb49e61267a tempest-ImagesNegativeTestJSON-1202128281 tempest-ImagesNegativeTestJSON-1202128281-project-member] Got 1 possible topologies {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 730.381104] env[61923]: DEBUG nova.virt.hardware [None req-13202777-73e8-40d3-a6ad-7bb49e61267a tempest-ImagesNegativeTestJSON-1202128281 tempest-ImagesNegativeTestJSON-1202128281-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 730.381490] env[61923]: DEBUG nova.virt.hardware [None req-13202777-73e8-40d3-a6ad-7bb49e61267a tempest-ImagesNegativeTestJSON-1202128281 tempest-ImagesNegativeTestJSON-1202128281-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 730.382586] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26dc284b-9473-44a1-b179-0f794e6e32c1 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.392297] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d1c100d-b9e0-41d7-a658-c1375d66783d {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 730.623217] env[61923]: DEBUG oslo_concurrency.lockutils [None req-f1ee1de3-0f72-403f-9698-da4536b6f37f tempest-ServerDiagnosticsV248Test-1822646244 tempest-ServerDiagnosticsV248Test-1822646244-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 730.714970] env[61923]: DEBUG oslo_concurrency.lockutils [None req-f7b52f4c-b2e3-49f8-b056-b7254f72bb20 tempest-AttachInterfacesUnderV243Test-498638506 tempest-AttachInterfacesUnderV243Test-498638506-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.395s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 730.715337] env[61923]: DEBUG nova.compute.manager [None req-f7b52f4c-b2e3-49f8-b056-b7254f72bb20 tempest-AttachInterfacesUnderV243Test-498638506 tempest-AttachInterfacesUnderV243Test-498638506-project-member] [instance: 0eab2200-2528-4a89-821a-cdd6a3c99297] Start building networks asynchronously for instance. {{(pid=61923) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 730.719223] env[61923]: DEBUG oslo_concurrency.lockutils [None req-4a9f7dc2-7221-4911-9bc8-b39b7a22ddda tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.574s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 730.721278] env[61923]: INFO nova.compute.claims [None req-4a9f7dc2-7221-4911-9bc8-b39b7a22ddda tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] [instance: 22d1e8ed-75c5-42e9-ae55-12d59e565914] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 731.221760] env[61923]: DEBUG nova.compute.utils [None req-f7b52f4c-b2e3-49f8-b056-b7254f72bb20 tempest-AttachInterfacesUnderV243Test-498638506 tempest-AttachInterfacesUnderV243Test-498638506-project-member] Using /dev/sd instead of None {{(pid=61923) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 731.226020] env[61923]: DEBUG nova.compute.manager [None req-f7b52f4c-b2e3-49f8-b056-b7254f72bb20 tempest-AttachInterfacesUnderV243Test-498638506 tempest-AttachInterfacesUnderV243Test-498638506-project-member] [instance: 0eab2200-2528-4a89-821a-cdd6a3c99297] Allocating IP information in the background. {{(pid=61923) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 731.226020] env[61923]: DEBUG nova.network.neutron [None req-f7b52f4c-b2e3-49f8-b056-b7254f72bb20 tempest-AttachInterfacesUnderV243Test-498638506 tempest-AttachInterfacesUnderV243Test-498638506-project-member] [instance: 0eab2200-2528-4a89-821a-cdd6a3c99297] allocate_for_instance() {{(pid=61923) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 731.345358] env[61923]: DEBUG nova.policy [None req-f7b52f4c-b2e3-49f8-b056-b7254f72bb20 tempest-AttachInterfacesUnderV243Test-498638506 tempest-AttachInterfacesUnderV243Test-498638506-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5707e34fe99f4c449b55c738a046f303', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd9a6fd5f7d7940f2a5778cd9b8e9f42e', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61923) authorize /opt/stack/nova/nova/policy.py:201}} [ 731.437695] env[61923]: DEBUG nova.compute.manager [req-a53ecab8-5876-4ee2-aa54-c57a4f57194f req-03b7ff25-eefa-4fae-bfaa-821134c8fa8a service nova] [instance: f7e1d517-3e39-4381-83be-0152368b8120] Received event network-changed-66c485f3-bd28-41e5-a69c-fc82fa725d19 {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 731.437879] env[61923]: DEBUG nova.compute.manager [req-a53ecab8-5876-4ee2-aa54-c57a4f57194f req-03b7ff25-eefa-4fae-bfaa-821134c8fa8a service nova] [instance: f7e1d517-3e39-4381-83be-0152368b8120] Refreshing instance network info cache due to event network-changed-66c485f3-bd28-41e5-a69c-fc82fa725d19. {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 731.438187] env[61923]: DEBUG oslo_concurrency.lockutils [req-a53ecab8-5876-4ee2-aa54-c57a4f57194f req-03b7ff25-eefa-4fae-bfaa-821134c8fa8a service nova] Acquiring lock "refresh_cache-f7e1d517-3e39-4381-83be-0152368b8120" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 731.438329] env[61923]: DEBUG oslo_concurrency.lockutils [req-a53ecab8-5876-4ee2-aa54-c57a4f57194f req-03b7ff25-eefa-4fae-bfaa-821134c8fa8a service nova] Acquired lock "refresh_cache-f7e1d517-3e39-4381-83be-0152368b8120" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 731.438493] env[61923]: DEBUG nova.network.neutron [req-a53ecab8-5876-4ee2-aa54-c57a4f57194f req-03b7ff25-eefa-4fae-bfaa-821134c8fa8a service nova] [instance: f7e1d517-3e39-4381-83be-0152368b8120] Refreshing network info cache for port 66c485f3-bd28-41e5-a69c-fc82fa725d19 {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 731.551545] env[61923]: DEBUG oslo_service.periodic_task [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61923) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 731.551545] env[61923]: DEBUG oslo_service.periodic_task [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=61923) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 731.729536] env[61923]: DEBUG nova.compute.manager [None req-f7b52f4c-b2e3-49f8-b056-b7254f72bb20 tempest-AttachInterfacesUnderV243Test-498638506 tempest-AttachInterfacesUnderV243Test-498638506-project-member] [instance: 0eab2200-2528-4a89-821a-cdd6a3c99297] Start building block device mappings for instance. {{(pid=61923) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 731.736734] env[61923]: ERROR nova.compute.manager [None req-13202777-73e8-40d3-a6ad-7bb49e61267a tempest-ImagesNegativeTestJSON-1202128281 tempest-ImagesNegativeTestJSON-1202128281-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 66c485f3-bd28-41e5-a69c-fc82fa725d19, please check neutron logs for more information. [ 731.736734] env[61923]: ERROR nova.compute.manager Traceback (most recent call last): [ 731.736734] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 731.736734] env[61923]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 731.736734] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 731.736734] env[61923]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 731.736734] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 731.736734] env[61923]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 731.736734] env[61923]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 731.736734] env[61923]: ERROR nova.compute.manager self.force_reraise() [ 731.736734] env[61923]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 731.736734] env[61923]: ERROR nova.compute.manager raise self.value [ 731.736734] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 731.736734] env[61923]: ERROR nova.compute.manager updated_port = self._update_port( [ 731.736734] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 731.736734] env[61923]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 731.737290] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 731.737290] env[61923]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 731.737290] env[61923]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 66c485f3-bd28-41e5-a69c-fc82fa725d19, please check neutron logs for more information. [ 731.737290] env[61923]: ERROR nova.compute.manager [ 731.737290] env[61923]: Traceback (most recent call last): [ 731.737290] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 731.737290] env[61923]: listener.cb(fileno) [ 731.737290] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 731.737290] env[61923]: result = function(*args, **kwargs) [ 731.737290] env[61923]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 731.737290] env[61923]: return func(*args, **kwargs) [ 731.737290] env[61923]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 731.737290] env[61923]: raise e [ 731.737290] env[61923]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 731.737290] env[61923]: nwinfo = self.network_api.allocate_for_instance( [ 731.737290] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 731.737290] env[61923]: created_port_ids = self._update_ports_for_instance( [ 731.737290] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 731.737290] env[61923]: with excutils.save_and_reraise_exception(): [ 731.737290] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 731.737290] env[61923]: self.force_reraise() [ 731.737290] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 731.737290] env[61923]: raise self.value [ 731.737290] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 731.737290] env[61923]: updated_port = self._update_port( [ 731.737290] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 731.737290] env[61923]: _ensure_no_port_binding_failure(port) [ 731.737290] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 731.737290] env[61923]: raise exception.PortBindingFailed(port_id=port['id']) [ 731.738182] env[61923]: nova.exception.PortBindingFailed: Binding failed for port 66c485f3-bd28-41e5-a69c-fc82fa725d19, please check neutron logs for more information. [ 731.738182] env[61923]: Removing descriptor: 14 [ 731.743502] env[61923]: ERROR nova.compute.manager [None req-13202777-73e8-40d3-a6ad-7bb49e61267a tempest-ImagesNegativeTestJSON-1202128281 tempest-ImagesNegativeTestJSON-1202128281-project-member] [instance: f7e1d517-3e39-4381-83be-0152368b8120] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 66c485f3-bd28-41e5-a69c-fc82fa725d19, please check neutron logs for more information. [ 731.743502] env[61923]: ERROR nova.compute.manager [instance: f7e1d517-3e39-4381-83be-0152368b8120] Traceback (most recent call last): [ 731.743502] env[61923]: ERROR nova.compute.manager [instance: f7e1d517-3e39-4381-83be-0152368b8120] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 731.743502] env[61923]: ERROR nova.compute.manager [instance: f7e1d517-3e39-4381-83be-0152368b8120] yield resources [ 731.743502] env[61923]: ERROR nova.compute.manager [instance: f7e1d517-3e39-4381-83be-0152368b8120] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 731.743502] env[61923]: ERROR nova.compute.manager [instance: f7e1d517-3e39-4381-83be-0152368b8120] self.driver.spawn(context, instance, image_meta, [ 731.743502] env[61923]: ERROR nova.compute.manager [instance: f7e1d517-3e39-4381-83be-0152368b8120] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 731.743502] env[61923]: ERROR nova.compute.manager [instance: f7e1d517-3e39-4381-83be-0152368b8120] self._vmops.spawn(context, instance, image_meta, injected_files, [ 731.743502] env[61923]: ERROR nova.compute.manager [instance: f7e1d517-3e39-4381-83be-0152368b8120] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 731.743502] env[61923]: ERROR nova.compute.manager [instance: f7e1d517-3e39-4381-83be-0152368b8120] vm_ref = self.build_virtual_machine(instance, [ 731.743502] env[61923]: ERROR nova.compute.manager [instance: f7e1d517-3e39-4381-83be-0152368b8120] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 731.743941] env[61923]: ERROR nova.compute.manager [instance: f7e1d517-3e39-4381-83be-0152368b8120] vif_infos = vmwarevif.get_vif_info(self._session, [ 731.743941] env[61923]: ERROR nova.compute.manager [instance: f7e1d517-3e39-4381-83be-0152368b8120] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 731.743941] env[61923]: ERROR nova.compute.manager [instance: f7e1d517-3e39-4381-83be-0152368b8120] for vif in network_info: [ 731.743941] env[61923]: ERROR nova.compute.manager [instance: f7e1d517-3e39-4381-83be-0152368b8120] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 731.743941] env[61923]: ERROR nova.compute.manager [instance: f7e1d517-3e39-4381-83be-0152368b8120] return self._sync_wrapper(fn, *args, **kwargs) [ 731.743941] env[61923]: ERROR nova.compute.manager [instance: f7e1d517-3e39-4381-83be-0152368b8120] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 731.743941] env[61923]: ERROR nova.compute.manager [instance: f7e1d517-3e39-4381-83be-0152368b8120] self.wait() [ 731.743941] env[61923]: ERROR nova.compute.manager [instance: f7e1d517-3e39-4381-83be-0152368b8120] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 731.743941] env[61923]: ERROR nova.compute.manager [instance: f7e1d517-3e39-4381-83be-0152368b8120] self[:] = self._gt.wait() [ 731.743941] env[61923]: ERROR nova.compute.manager [instance: f7e1d517-3e39-4381-83be-0152368b8120] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 731.743941] env[61923]: ERROR nova.compute.manager [instance: f7e1d517-3e39-4381-83be-0152368b8120] return self._exit_event.wait() [ 731.743941] env[61923]: ERROR nova.compute.manager [instance: f7e1d517-3e39-4381-83be-0152368b8120] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 731.743941] env[61923]: ERROR nova.compute.manager [instance: f7e1d517-3e39-4381-83be-0152368b8120] result = hub.switch() [ 731.744395] env[61923]: ERROR nova.compute.manager [instance: f7e1d517-3e39-4381-83be-0152368b8120] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 731.744395] env[61923]: ERROR nova.compute.manager [instance: f7e1d517-3e39-4381-83be-0152368b8120] return self.greenlet.switch() [ 731.744395] env[61923]: ERROR nova.compute.manager [instance: f7e1d517-3e39-4381-83be-0152368b8120] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 731.744395] env[61923]: ERROR nova.compute.manager [instance: f7e1d517-3e39-4381-83be-0152368b8120] result = function(*args, **kwargs) [ 731.744395] env[61923]: ERROR nova.compute.manager [instance: f7e1d517-3e39-4381-83be-0152368b8120] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 731.744395] env[61923]: ERROR nova.compute.manager [instance: f7e1d517-3e39-4381-83be-0152368b8120] return func(*args, **kwargs) [ 731.744395] env[61923]: ERROR nova.compute.manager [instance: f7e1d517-3e39-4381-83be-0152368b8120] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 731.744395] env[61923]: ERROR nova.compute.manager [instance: f7e1d517-3e39-4381-83be-0152368b8120] raise e [ 731.744395] env[61923]: ERROR nova.compute.manager [instance: f7e1d517-3e39-4381-83be-0152368b8120] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 731.744395] env[61923]: ERROR nova.compute.manager [instance: f7e1d517-3e39-4381-83be-0152368b8120] nwinfo = self.network_api.allocate_for_instance( [ 731.744395] env[61923]: ERROR nova.compute.manager [instance: f7e1d517-3e39-4381-83be-0152368b8120] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 731.744395] env[61923]: ERROR nova.compute.manager [instance: f7e1d517-3e39-4381-83be-0152368b8120] created_port_ids = self._update_ports_for_instance( [ 731.744395] env[61923]: ERROR nova.compute.manager [instance: f7e1d517-3e39-4381-83be-0152368b8120] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 731.744819] env[61923]: ERROR nova.compute.manager [instance: f7e1d517-3e39-4381-83be-0152368b8120] with excutils.save_and_reraise_exception(): [ 731.744819] env[61923]: ERROR nova.compute.manager [instance: f7e1d517-3e39-4381-83be-0152368b8120] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 731.744819] env[61923]: ERROR nova.compute.manager [instance: f7e1d517-3e39-4381-83be-0152368b8120] self.force_reraise() [ 731.744819] env[61923]: ERROR nova.compute.manager [instance: f7e1d517-3e39-4381-83be-0152368b8120] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 731.744819] env[61923]: ERROR nova.compute.manager [instance: f7e1d517-3e39-4381-83be-0152368b8120] raise self.value [ 731.744819] env[61923]: ERROR nova.compute.manager [instance: f7e1d517-3e39-4381-83be-0152368b8120] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 731.744819] env[61923]: ERROR nova.compute.manager [instance: f7e1d517-3e39-4381-83be-0152368b8120] updated_port = self._update_port( [ 731.744819] env[61923]: ERROR nova.compute.manager [instance: f7e1d517-3e39-4381-83be-0152368b8120] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 731.744819] env[61923]: ERROR nova.compute.manager [instance: f7e1d517-3e39-4381-83be-0152368b8120] _ensure_no_port_binding_failure(port) [ 731.744819] env[61923]: ERROR nova.compute.manager [instance: f7e1d517-3e39-4381-83be-0152368b8120] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 731.744819] env[61923]: ERROR nova.compute.manager [instance: f7e1d517-3e39-4381-83be-0152368b8120] raise exception.PortBindingFailed(port_id=port['id']) [ 731.744819] env[61923]: ERROR nova.compute.manager [instance: f7e1d517-3e39-4381-83be-0152368b8120] nova.exception.PortBindingFailed: Binding failed for port 66c485f3-bd28-41e5-a69c-fc82fa725d19, please check neutron logs for more information. [ 731.744819] env[61923]: ERROR nova.compute.manager [instance: f7e1d517-3e39-4381-83be-0152368b8120] [ 731.745242] env[61923]: INFO nova.compute.manager [None req-13202777-73e8-40d3-a6ad-7bb49e61267a tempest-ImagesNegativeTestJSON-1202128281 tempest-ImagesNegativeTestJSON-1202128281-project-member] [instance: f7e1d517-3e39-4381-83be-0152368b8120] Terminating instance [ 731.749560] env[61923]: DEBUG oslo_concurrency.lockutils [None req-13202777-73e8-40d3-a6ad-7bb49e61267a tempest-ImagesNegativeTestJSON-1202128281 tempest-ImagesNegativeTestJSON-1202128281-project-member] Acquiring lock "refresh_cache-f7e1d517-3e39-4381-83be-0152368b8120" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 731.989232] env[61923]: DEBUG nova.network.neutron [req-a53ecab8-5876-4ee2-aa54-c57a4f57194f req-03b7ff25-eefa-4fae-bfaa-821134c8fa8a service nova] [instance: f7e1d517-3e39-4381-83be-0152368b8120] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 732.052740] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dbd2ff34-52d6-4e77-822e-6006ce44aadb {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.060283] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0eb152d5-d864-4bc8-a134-81841f7ae8a1 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.063852] env[61923]: DEBUG oslo_service.periodic_task [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61923) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 732.064601] env[61923]: DEBUG nova.compute.manager [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Starting heal instance info cache {{(pid=61923) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 732.064601] env[61923]: DEBUG nova.compute.manager [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Rebuilding the list of instances to heal {{(pid=61923) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 732.101668] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8b79dc9-7ca3-4a24-9f10-bcdf6f0378a3 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.109986] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da828917-4595-4efa-aa7f-bf3761548621 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.126966] env[61923]: DEBUG nova.compute.provider_tree [None req-4a9f7dc2-7221-4911-9bc8-b39b7a22ddda tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 732.156983] env[61923]: DEBUG nova.network.neutron [None req-f7b52f4c-b2e3-49f8-b056-b7254f72bb20 tempest-AttachInterfacesUnderV243Test-498638506 tempest-AttachInterfacesUnderV243Test-498638506-project-member] [instance: 0eab2200-2528-4a89-821a-cdd6a3c99297] Successfully created port: 4c6d06bd-7574-4546-acd9-9d362bc379b1 {{(pid=61923) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 732.240558] env[61923]: DEBUG nova.network.neutron [req-a53ecab8-5876-4ee2-aa54-c57a4f57194f req-03b7ff25-eefa-4fae-bfaa-821134c8fa8a service nova] [instance: f7e1d517-3e39-4381-83be-0152368b8120] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 732.568198] env[61923]: DEBUG nova.compute.manager [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] [instance: 74b32299-983f-41f7-b3f4-3a12815a83e6] Skipping network cache update for instance because it is Building. {{(pid=61923) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 732.568488] env[61923]: DEBUG nova.compute.manager [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] [instance: 7c0341f9-7401-42ee-9512-afd832c76940] Skipping network cache update for instance because it is Building. {{(pid=61923) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 732.568658] env[61923]: DEBUG nova.compute.manager [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] [instance: f7e1d517-3e39-4381-83be-0152368b8120] Skipping network cache update for instance because it is Building. {{(pid=61923) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 732.568791] env[61923]: DEBUG nova.compute.manager [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] [instance: 0eab2200-2528-4a89-821a-cdd6a3c99297] Skipping network cache update for instance because it is Building. {{(pid=61923) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 732.568944] env[61923]: DEBUG nova.compute.manager [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] [instance: 22d1e8ed-75c5-42e9-ae55-12d59e565914] Skipping network cache update for instance because it is Building. {{(pid=61923) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 732.569125] env[61923]: DEBUG nova.compute.manager [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Didn't find any instances for network info cache update. {{(pid=61923) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10014}} [ 732.569329] env[61923]: DEBUG oslo_service.periodic_task [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61923) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 732.569480] env[61923]: DEBUG oslo_service.periodic_task [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61923) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 732.569621] env[61923]: DEBUG oslo_service.periodic_task [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61923) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 732.569761] env[61923]: DEBUG oslo_service.periodic_task [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61923) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 732.569928] env[61923]: DEBUG oslo_service.periodic_task [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61923) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 732.570049] env[61923]: DEBUG oslo_service.periodic_task [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61923) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 732.570179] env[61923]: DEBUG nova.compute.manager [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61923) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 732.570340] env[61923]: DEBUG oslo_service.periodic_task [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Running periodic task ComputeManager.update_available_resource {{(pid=61923) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 732.630341] env[61923]: DEBUG nova.scheduler.client.report [None req-4a9f7dc2-7221-4911-9bc8-b39b7a22ddda tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 732.747241] env[61923]: DEBUG oslo_concurrency.lockutils [req-a53ecab8-5876-4ee2-aa54-c57a4f57194f req-03b7ff25-eefa-4fae-bfaa-821134c8fa8a service nova] Releasing lock "refresh_cache-f7e1d517-3e39-4381-83be-0152368b8120" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 732.747716] env[61923]: DEBUG oslo_concurrency.lockutils [None req-13202777-73e8-40d3-a6ad-7bb49e61267a tempest-ImagesNegativeTestJSON-1202128281 tempest-ImagesNegativeTestJSON-1202128281-project-member] Acquired lock "refresh_cache-f7e1d517-3e39-4381-83be-0152368b8120" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 732.747965] env[61923]: DEBUG nova.network.neutron [None req-13202777-73e8-40d3-a6ad-7bb49e61267a tempest-ImagesNegativeTestJSON-1202128281 tempest-ImagesNegativeTestJSON-1202128281-project-member] [instance: f7e1d517-3e39-4381-83be-0152368b8120] Building network info cache for instance {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 732.757602] env[61923]: DEBUG nova.compute.manager [None req-f7b52f4c-b2e3-49f8-b056-b7254f72bb20 tempest-AttachInterfacesUnderV243Test-498638506 tempest-AttachInterfacesUnderV243Test-498638506-project-member] [instance: 0eab2200-2528-4a89-821a-cdd6a3c99297] Start spawning the instance on the hypervisor. {{(pid=61923) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 732.784072] env[61923]: DEBUG nova.virt.hardware [None req-f7b52f4c-b2e3-49f8-b056-b7254f72bb20 tempest-AttachInterfacesUnderV243Test-498638506 tempest-AttachInterfacesUnderV243Test-498638506-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-29T20:07:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-29T20:07:35Z,direct_url=,disk_format='vmdk',id=0f153f63-ae0a-45b1-b7f5-7f9b673c947f,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='4e7b5e3b3c3443c8bd5c70f8a15739f5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-29T20:07:36Z,virtual_size=,visibility=), allow threads: False {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 732.784323] env[61923]: DEBUG nova.virt.hardware [None req-f7b52f4c-b2e3-49f8-b056-b7254f72bb20 tempest-AttachInterfacesUnderV243Test-498638506 tempest-AttachInterfacesUnderV243Test-498638506-project-member] Flavor limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 732.784470] env[61923]: DEBUG nova.virt.hardware [None req-f7b52f4c-b2e3-49f8-b056-b7254f72bb20 tempest-AttachInterfacesUnderV243Test-498638506 tempest-AttachInterfacesUnderV243Test-498638506-project-member] Image limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 732.784640] env[61923]: DEBUG nova.virt.hardware [None req-f7b52f4c-b2e3-49f8-b056-b7254f72bb20 tempest-AttachInterfacesUnderV243Test-498638506 tempest-AttachInterfacesUnderV243Test-498638506-project-member] Flavor pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 732.784775] env[61923]: DEBUG nova.virt.hardware [None req-f7b52f4c-b2e3-49f8-b056-b7254f72bb20 tempest-AttachInterfacesUnderV243Test-498638506 tempest-AttachInterfacesUnderV243Test-498638506-project-member] Image pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 732.784921] env[61923]: DEBUG nova.virt.hardware [None req-f7b52f4c-b2e3-49f8-b056-b7254f72bb20 tempest-AttachInterfacesUnderV243Test-498638506 tempest-AttachInterfacesUnderV243Test-498638506-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 732.785224] env[61923]: DEBUG nova.virt.hardware [None req-f7b52f4c-b2e3-49f8-b056-b7254f72bb20 tempest-AttachInterfacesUnderV243Test-498638506 tempest-AttachInterfacesUnderV243Test-498638506-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 732.785389] env[61923]: DEBUG nova.virt.hardware [None req-f7b52f4c-b2e3-49f8-b056-b7254f72bb20 tempest-AttachInterfacesUnderV243Test-498638506 tempest-AttachInterfacesUnderV243Test-498638506-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 732.785549] env[61923]: DEBUG nova.virt.hardware [None req-f7b52f4c-b2e3-49f8-b056-b7254f72bb20 tempest-AttachInterfacesUnderV243Test-498638506 tempest-AttachInterfacesUnderV243Test-498638506-project-member] Got 1 possible topologies {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 732.785704] env[61923]: DEBUG nova.virt.hardware [None req-f7b52f4c-b2e3-49f8-b056-b7254f72bb20 tempest-AttachInterfacesUnderV243Test-498638506 tempest-AttachInterfacesUnderV243Test-498638506-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 732.785868] env[61923]: DEBUG nova.virt.hardware [None req-f7b52f4c-b2e3-49f8-b056-b7254f72bb20 tempest-AttachInterfacesUnderV243Test-498638506 tempest-AttachInterfacesUnderV243Test-498638506-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 732.786770] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc6f884c-a8fd-48c1-a0ab-592d179dbb24 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 732.794772] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1f61606-ebc5-4228-9739-dcfa27c0703c {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.074297] env[61923]: DEBUG oslo_concurrency.lockutils [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 733.135451] env[61923]: DEBUG oslo_concurrency.lockutils [None req-4a9f7dc2-7221-4911-9bc8-b39b7a22ddda tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.416s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 733.136579] env[61923]: DEBUG nova.compute.manager [None req-4a9f7dc2-7221-4911-9bc8-b39b7a22ddda tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] [instance: 22d1e8ed-75c5-42e9-ae55-12d59e565914] Start building networks asynchronously for instance. {{(pid=61923) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 733.139514] env[61923]: DEBUG oslo_concurrency.lockutils [None req-5bb1ca17-5442-4564-86ab-53f866a97e1d tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.236s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 733.140986] env[61923]: INFO nova.compute.claims [None req-5bb1ca17-5442-4564-86ab-53f866a97e1d tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: 15eaaa7c-77be-4b8b-89d0-0b2b11a0c49e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 733.286910] env[61923]: DEBUG nova.network.neutron [None req-13202777-73e8-40d3-a6ad-7bb49e61267a tempest-ImagesNegativeTestJSON-1202128281 tempest-ImagesNegativeTestJSON-1202128281-project-member] [instance: f7e1d517-3e39-4381-83be-0152368b8120] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 733.533117] env[61923]: DEBUG nova.network.neutron [None req-13202777-73e8-40d3-a6ad-7bb49e61267a tempest-ImagesNegativeTestJSON-1202128281 tempest-ImagesNegativeTestJSON-1202128281-project-member] [instance: f7e1d517-3e39-4381-83be-0152368b8120] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 733.655279] env[61923]: DEBUG nova.compute.utils [None req-4a9f7dc2-7221-4911-9bc8-b39b7a22ddda tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Using /dev/sd instead of None {{(pid=61923) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 733.655279] env[61923]: DEBUG nova.compute.manager [None req-4a9f7dc2-7221-4911-9bc8-b39b7a22ddda tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] [instance: 22d1e8ed-75c5-42e9-ae55-12d59e565914] Allocating IP information in the background. {{(pid=61923) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 733.655279] env[61923]: DEBUG nova.network.neutron [None req-4a9f7dc2-7221-4911-9bc8-b39b7a22ddda tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] [instance: 22d1e8ed-75c5-42e9-ae55-12d59e565914] allocate_for_instance() {{(pid=61923) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 733.674748] env[61923]: DEBUG nova.compute.manager [req-672956e5-a564-4f8d-8e11-624f8314695c req-c32dddfc-2fdb-4e93-bc23-2c9b6ed252be service nova] [instance: f7e1d517-3e39-4381-83be-0152368b8120] Received event network-vif-deleted-66c485f3-bd28-41e5-a69c-fc82fa725d19 {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 733.809925] env[61923]: DEBUG nova.policy [None req-4a9f7dc2-7221-4911-9bc8-b39b7a22ddda tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'cd5039bf8f4344e490df828e643d2779', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '912300a09b80452e85e3cd13c4a644cc', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61923) authorize /opt/stack/nova/nova/policy.py:201}} [ 734.037235] env[61923]: DEBUG oslo_concurrency.lockutils [None req-13202777-73e8-40d3-a6ad-7bb49e61267a tempest-ImagesNegativeTestJSON-1202128281 tempest-ImagesNegativeTestJSON-1202128281-project-member] Releasing lock "refresh_cache-f7e1d517-3e39-4381-83be-0152368b8120" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 734.037235] env[61923]: DEBUG nova.compute.manager [None req-13202777-73e8-40d3-a6ad-7bb49e61267a tempest-ImagesNegativeTestJSON-1202128281 tempest-ImagesNegativeTestJSON-1202128281-project-member] [instance: f7e1d517-3e39-4381-83be-0152368b8120] Start destroying the instance on the hypervisor. {{(pid=61923) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 734.037235] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-13202777-73e8-40d3-a6ad-7bb49e61267a tempest-ImagesNegativeTestJSON-1202128281 tempest-ImagesNegativeTestJSON-1202128281-project-member] [instance: f7e1d517-3e39-4381-83be-0152368b8120] Destroying instance {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 734.037235] env[61923]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ed3ed09c-3bcb-4c07-9c43-948227e562b5 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.050489] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2bdd71dc-50ae-4fc9-92cc-b8afb2df7e01 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.076211] env[61923]: WARNING nova.virt.vmwareapi.vmops [None req-13202777-73e8-40d3-a6ad-7bb49e61267a tempest-ImagesNegativeTestJSON-1202128281 tempest-ImagesNegativeTestJSON-1202128281-project-member] [instance: f7e1d517-3e39-4381-83be-0152368b8120] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance f7e1d517-3e39-4381-83be-0152368b8120 could not be found. [ 734.076707] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-13202777-73e8-40d3-a6ad-7bb49e61267a tempest-ImagesNegativeTestJSON-1202128281 tempest-ImagesNegativeTestJSON-1202128281-project-member] [instance: f7e1d517-3e39-4381-83be-0152368b8120] Instance destroyed {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 734.077062] env[61923]: INFO nova.compute.manager [None req-13202777-73e8-40d3-a6ad-7bb49e61267a tempest-ImagesNegativeTestJSON-1202128281 tempest-ImagesNegativeTestJSON-1202128281-project-member] [instance: f7e1d517-3e39-4381-83be-0152368b8120] Took 0.04 seconds to destroy the instance on the hypervisor. [ 734.077462] env[61923]: DEBUG oslo.service.loopingcall [None req-13202777-73e8-40d3-a6ad-7bb49e61267a tempest-ImagesNegativeTestJSON-1202128281 tempest-ImagesNegativeTestJSON-1202128281-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61923) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 734.077829] env[61923]: DEBUG nova.compute.manager [-] [instance: f7e1d517-3e39-4381-83be-0152368b8120] Deallocating network for instance {{(pid=61923) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 734.078180] env[61923]: DEBUG nova.network.neutron [-] [instance: f7e1d517-3e39-4381-83be-0152368b8120] deallocate_for_instance() {{(pid=61923) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 734.101054] env[61923]: DEBUG nova.network.neutron [-] [instance: f7e1d517-3e39-4381-83be-0152368b8120] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 734.162624] env[61923]: DEBUG nova.compute.manager [None req-4a9f7dc2-7221-4911-9bc8-b39b7a22ddda tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] [instance: 22d1e8ed-75c5-42e9-ae55-12d59e565914] Start building block device mappings for instance. {{(pid=61923) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 734.530582] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e74ce89e-28f8-49c9-bbb0-80e717cb389a {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.538316] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-691ce71c-2a07-47a7-bb16-ce705c77ec4b {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.568217] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5fef131-a02f-4f9e-a161-ecc318d4308f {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.577019] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61132a34-aab6-4896-81eb-b50909c4b942 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 734.589638] env[61923]: DEBUG nova.compute.provider_tree [None req-5bb1ca17-5442-4564-86ab-53f866a97e1d tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 734.605861] env[61923]: DEBUG nova.network.neutron [-] [instance: f7e1d517-3e39-4381-83be-0152368b8120] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 734.856310] env[61923]: DEBUG nova.network.neutron [None req-4a9f7dc2-7221-4911-9bc8-b39b7a22ddda tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] [instance: 22d1e8ed-75c5-42e9-ae55-12d59e565914] Successfully created port: 32585aee-c6bc-4a95-9516-646bb26bf0a0 {{(pid=61923) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 735.096570] env[61923]: DEBUG nova.scheduler.client.report [None req-5bb1ca17-5442-4564-86ab-53f866a97e1d tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 735.109863] env[61923]: INFO nova.compute.manager [-] [instance: f7e1d517-3e39-4381-83be-0152368b8120] Took 1.03 seconds to deallocate network for instance. [ 735.117843] env[61923]: DEBUG nova.compute.claims [None req-13202777-73e8-40d3-a6ad-7bb49e61267a tempest-ImagesNegativeTestJSON-1202128281 tempest-ImagesNegativeTestJSON-1202128281-project-member] [instance: f7e1d517-3e39-4381-83be-0152368b8120] Aborting claim: {{(pid=61923) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 735.117843] env[61923]: DEBUG oslo_concurrency.lockutils [None req-13202777-73e8-40d3-a6ad-7bb49e61267a tempest-ImagesNegativeTestJSON-1202128281 tempest-ImagesNegativeTestJSON-1202128281-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 735.176842] env[61923]: DEBUG nova.compute.manager [None req-4a9f7dc2-7221-4911-9bc8-b39b7a22ddda tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] [instance: 22d1e8ed-75c5-42e9-ae55-12d59e565914] Start spawning the instance on the hypervisor. {{(pid=61923) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 735.219013] env[61923]: DEBUG nova.virt.hardware [None req-4a9f7dc2-7221-4911-9bc8-b39b7a22ddda tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-29T20:07:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-29T20:07:35Z,direct_url=,disk_format='vmdk',id=0f153f63-ae0a-45b1-b7f5-7f9b673c947f,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='4e7b5e3b3c3443c8bd5c70f8a15739f5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-29T20:07:36Z,virtual_size=,visibility=), allow threads: False {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 735.219013] env[61923]: DEBUG nova.virt.hardware [None req-4a9f7dc2-7221-4911-9bc8-b39b7a22ddda tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Flavor limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 735.219013] env[61923]: DEBUG nova.virt.hardware [None req-4a9f7dc2-7221-4911-9bc8-b39b7a22ddda tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Image limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 735.219251] env[61923]: DEBUG nova.virt.hardware [None req-4a9f7dc2-7221-4911-9bc8-b39b7a22ddda tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Flavor pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 735.219251] env[61923]: DEBUG nova.virt.hardware [None req-4a9f7dc2-7221-4911-9bc8-b39b7a22ddda tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Image pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 735.219251] env[61923]: DEBUG nova.virt.hardware [None req-4a9f7dc2-7221-4911-9bc8-b39b7a22ddda tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 735.219251] env[61923]: DEBUG nova.virt.hardware [None req-4a9f7dc2-7221-4911-9bc8-b39b7a22ddda tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 735.219251] env[61923]: DEBUG nova.virt.hardware [None req-4a9f7dc2-7221-4911-9bc8-b39b7a22ddda tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 735.219413] env[61923]: DEBUG nova.virt.hardware [None req-4a9f7dc2-7221-4911-9bc8-b39b7a22ddda tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Got 1 possible topologies {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 735.219413] env[61923]: DEBUG nova.virt.hardware [None req-4a9f7dc2-7221-4911-9bc8-b39b7a22ddda tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 735.219478] env[61923]: DEBUG nova.virt.hardware [None req-4a9f7dc2-7221-4911-9bc8-b39b7a22ddda tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 735.220373] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5ab9fbd-1073-4859-bca3-831fc95e8f44 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.235617] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83bd8fa5-a1cb-4515-865c-e8f43d949a41 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.606101] env[61923]: DEBUG oslo_concurrency.lockutils [None req-5bb1ca17-5442-4564-86ab-53f866a97e1d tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.466s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 735.606658] env[61923]: DEBUG nova.compute.manager [None req-5bb1ca17-5442-4564-86ab-53f866a97e1d tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: 15eaaa7c-77be-4b8b-89d0-0b2b11a0c49e] Start building networks asynchronously for instance. {{(pid=61923) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 735.609596] env[61923]: DEBUG oslo_concurrency.lockutils [None req-425a3606-915c-4f9b-8507-50f1c8d0d8d9 tempest-ServerRescueTestJSON-521767381 tempest-ServerRescueTestJSON-521767381-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.292s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 735.620317] env[61923]: ERROR nova.compute.manager [None req-f7b52f4c-b2e3-49f8-b056-b7254f72bb20 tempest-AttachInterfacesUnderV243Test-498638506 tempest-AttachInterfacesUnderV243Test-498638506-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 4c6d06bd-7574-4546-acd9-9d362bc379b1, please check neutron logs for more information. [ 735.620317] env[61923]: ERROR nova.compute.manager Traceback (most recent call last): [ 735.620317] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 735.620317] env[61923]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 735.620317] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 735.620317] env[61923]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 735.620317] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 735.620317] env[61923]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 735.620317] env[61923]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 735.620317] env[61923]: ERROR nova.compute.manager self.force_reraise() [ 735.620317] env[61923]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 735.620317] env[61923]: ERROR nova.compute.manager raise self.value [ 735.620317] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 735.620317] env[61923]: ERROR nova.compute.manager updated_port = self._update_port( [ 735.620317] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 735.620317] env[61923]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 735.620766] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 735.620766] env[61923]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 735.620766] env[61923]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 4c6d06bd-7574-4546-acd9-9d362bc379b1, please check neutron logs for more information. [ 735.620766] env[61923]: ERROR nova.compute.manager [ 735.620766] env[61923]: Traceback (most recent call last): [ 735.620766] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 735.620766] env[61923]: listener.cb(fileno) [ 735.620766] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 735.620766] env[61923]: result = function(*args, **kwargs) [ 735.620766] env[61923]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 735.620766] env[61923]: return func(*args, **kwargs) [ 735.620766] env[61923]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 735.620766] env[61923]: raise e [ 735.620766] env[61923]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 735.620766] env[61923]: nwinfo = self.network_api.allocate_for_instance( [ 735.620766] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 735.620766] env[61923]: created_port_ids = self._update_ports_for_instance( [ 735.620766] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 735.620766] env[61923]: with excutils.save_and_reraise_exception(): [ 735.620766] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 735.620766] env[61923]: self.force_reraise() [ 735.620766] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 735.620766] env[61923]: raise self.value [ 735.620766] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 735.620766] env[61923]: updated_port = self._update_port( [ 735.620766] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 735.620766] env[61923]: _ensure_no_port_binding_failure(port) [ 735.620766] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 735.620766] env[61923]: raise exception.PortBindingFailed(port_id=port['id']) [ 735.621580] env[61923]: nova.exception.PortBindingFailed: Binding failed for port 4c6d06bd-7574-4546-acd9-9d362bc379b1, please check neutron logs for more information. [ 735.621580] env[61923]: Removing descriptor: 17 [ 735.621580] env[61923]: ERROR nova.compute.manager [None req-f7b52f4c-b2e3-49f8-b056-b7254f72bb20 tempest-AttachInterfacesUnderV243Test-498638506 tempest-AttachInterfacesUnderV243Test-498638506-project-member] [instance: 0eab2200-2528-4a89-821a-cdd6a3c99297] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 4c6d06bd-7574-4546-acd9-9d362bc379b1, please check neutron logs for more information. [ 735.621580] env[61923]: ERROR nova.compute.manager [instance: 0eab2200-2528-4a89-821a-cdd6a3c99297] Traceback (most recent call last): [ 735.621580] env[61923]: ERROR nova.compute.manager [instance: 0eab2200-2528-4a89-821a-cdd6a3c99297] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 735.621580] env[61923]: ERROR nova.compute.manager [instance: 0eab2200-2528-4a89-821a-cdd6a3c99297] yield resources [ 735.621580] env[61923]: ERROR nova.compute.manager [instance: 0eab2200-2528-4a89-821a-cdd6a3c99297] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 735.621580] env[61923]: ERROR nova.compute.manager [instance: 0eab2200-2528-4a89-821a-cdd6a3c99297] self.driver.spawn(context, instance, image_meta, [ 735.621580] env[61923]: ERROR nova.compute.manager [instance: 0eab2200-2528-4a89-821a-cdd6a3c99297] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 735.621580] env[61923]: ERROR nova.compute.manager [instance: 0eab2200-2528-4a89-821a-cdd6a3c99297] self._vmops.spawn(context, instance, image_meta, injected_files, [ 735.621580] env[61923]: ERROR nova.compute.manager [instance: 0eab2200-2528-4a89-821a-cdd6a3c99297] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 735.621580] env[61923]: ERROR nova.compute.manager [instance: 0eab2200-2528-4a89-821a-cdd6a3c99297] vm_ref = self.build_virtual_machine(instance, [ 735.622063] env[61923]: ERROR nova.compute.manager [instance: 0eab2200-2528-4a89-821a-cdd6a3c99297] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 735.622063] env[61923]: ERROR nova.compute.manager [instance: 0eab2200-2528-4a89-821a-cdd6a3c99297] vif_infos = vmwarevif.get_vif_info(self._session, [ 735.622063] env[61923]: ERROR nova.compute.manager [instance: 0eab2200-2528-4a89-821a-cdd6a3c99297] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 735.622063] env[61923]: ERROR nova.compute.manager [instance: 0eab2200-2528-4a89-821a-cdd6a3c99297] for vif in network_info: [ 735.622063] env[61923]: ERROR nova.compute.manager [instance: 0eab2200-2528-4a89-821a-cdd6a3c99297] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 735.622063] env[61923]: ERROR nova.compute.manager [instance: 0eab2200-2528-4a89-821a-cdd6a3c99297] return self._sync_wrapper(fn, *args, **kwargs) [ 735.622063] env[61923]: ERROR nova.compute.manager [instance: 0eab2200-2528-4a89-821a-cdd6a3c99297] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 735.622063] env[61923]: ERROR nova.compute.manager [instance: 0eab2200-2528-4a89-821a-cdd6a3c99297] self.wait() [ 735.622063] env[61923]: ERROR nova.compute.manager [instance: 0eab2200-2528-4a89-821a-cdd6a3c99297] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 735.622063] env[61923]: ERROR nova.compute.manager [instance: 0eab2200-2528-4a89-821a-cdd6a3c99297] self[:] = self._gt.wait() [ 735.622063] env[61923]: ERROR nova.compute.manager [instance: 0eab2200-2528-4a89-821a-cdd6a3c99297] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 735.622063] env[61923]: ERROR nova.compute.manager [instance: 0eab2200-2528-4a89-821a-cdd6a3c99297] return self._exit_event.wait() [ 735.622063] env[61923]: ERROR nova.compute.manager [instance: 0eab2200-2528-4a89-821a-cdd6a3c99297] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 735.622419] env[61923]: ERROR nova.compute.manager [instance: 0eab2200-2528-4a89-821a-cdd6a3c99297] result = hub.switch() [ 735.622419] env[61923]: ERROR nova.compute.manager [instance: 0eab2200-2528-4a89-821a-cdd6a3c99297] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 735.622419] env[61923]: ERROR nova.compute.manager [instance: 0eab2200-2528-4a89-821a-cdd6a3c99297] return self.greenlet.switch() [ 735.622419] env[61923]: ERROR nova.compute.manager [instance: 0eab2200-2528-4a89-821a-cdd6a3c99297] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 735.622419] env[61923]: ERROR nova.compute.manager [instance: 0eab2200-2528-4a89-821a-cdd6a3c99297] result = function(*args, **kwargs) [ 735.622419] env[61923]: ERROR nova.compute.manager [instance: 0eab2200-2528-4a89-821a-cdd6a3c99297] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 735.622419] env[61923]: ERROR nova.compute.manager [instance: 0eab2200-2528-4a89-821a-cdd6a3c99297] return func(*args, **kwargs) [ 735.622419] env[61923]: ERROR nova.compute.manager [instance: 0eab2200-2528-4a89-821a-cdd6a3c99297] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 735.622419] env[61923]: ERROR nova.compute.manager [instance: 0eab2200-2528-4a89-821a-cdd6a3c99297] raise e [ 735.622419] env[61923]: ERROR nova.compute.manager [instance: 0eab2200-2528-4a89-821a-cdd6a3c99297] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 735.622419] env[61923]: ERROR nova.compute.manager [instance: 0eab2200-2528-4a89-821a-cdd6a3c99297] nwinfo = self.network_api.allocate_for_instance( [ 735.622419] env[61923]: ERROR nova.compute.manager [instance: 0eab2200-2528-4a89-821a-cdd6a3c99297] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 735.622419] env[61923]: ERROR nova.compute.manager [instance: 0eab2200-2528-4a89-821a-cdd6a3c99297] created_port_ids = self._update_ports_for_instance( [ 735.625529] env[61923]: ERROR nova.compute.manager [instance: 0eab2200-2528-4a89-821a-cdd6a3c99297] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 735.625529] env[61923]: ERROR nova.compute.manager [instance: 0eab2200-2528-4a89-821a-cdd6a3c99297] with excutils.save_and_reraise_exception(): [ 735.625529] env[61923]: ERROR nova.compute.manager [instance: 0eab2200-2528-4a89-821a-cdd6a3c99297] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 735.625529] env[61923]: ERROR nova.compute.manager [instance: 0eab2200-2528-4a89-821a-cdd6a3c99297] self.force_reraise() [ 735.625529] env[61923]: ERROR nova.compute.manager [instance: 0eab2200-2528-4a89-821a-cdd6a3c99297] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 735.625529] env[61923]: ERROR nova.compute.manager [instance: 0eab2200-2528-4a89-821a-cdd6a3c99297] raise self.value [ 735.625529] env[61923]: ERROR nova.compute.manager [instance: 0eab2200-2528-4a89-821a-cdd6a3c99297] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 735.625529] env[61923]: ERROR nova.compute.manager [instance: 0eab2200-2528-4a89-821a-cdd6a3c99297] updated_port = self._update_port( [ 735.625529] env[61923]: ERROR nova.compute.manager [instance: 0eab2200-2528-4a89-821a-cdd6a3c99297] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 735.625529] env[61923]: ERROR nova.compute.manager [instance: 0eab2200-2528-4a89-821a-cdd6a3c99297] _ensure_no_port_binding_failure(port) [ 735.625529] env[61923]: ERROR nova.compute.manager [instance: 0eab2200-2528-4a89-821a-cdd6a3c99297] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 735.625529] env[61923]: ERROR nova.compute.manager [instance: 0eab2200-2528-4a89-821a-cdd6a3c99297] raise exception.PortBindingFailed(port_id=port['id']) [ 735.625872] env[61923]: ERROR nova.compute.manager [instance: 0eab2200-2528-4a89-821a-cdd6a3c99297] nova.exception.PortBindingFailed: Binding failed for port 4c6d06bd-7574-4546-acd9-9d362bc379b1, please check neutron logs for more information. [ 735.625872] env[61923]: ERROR nova.compute.manager [instance: 0eab2200-2528-4a89-821a-cdd6a3c99297] [ 735.625872] env[61923]: INFO nova.compute.manager [None req-f7b52f4c-b2e3-49f8-b056-b7254f72bb20 tempest-AttachInterfacesUnderV243Test-498638506 tempest-AttachInterfacesUnderV243Test-498638506-project-member] [instance: 0eab2200-2528-4a89-821a-cdd6a3c99297] Terminating instance [ 735.633444] env[61923]: DEBUG oslo_concurrency.lockutils [None req-f7b52f4c-b2e3-49f8-b056-b7254f72bb20 tempest-AttachInterfacesUnderV243Test-498638506 tempest-AttachInterfacesUnderV243Test-498638506-project-member] Acquiring lock "refresh_cache-0eab2200-2528-4a89-821a-cdd6a3c99297" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 735.633444] env[61923]: DEBUG oslo_concurrency.lockutils [None req-f7b52f4c-b2e3-49f8-b056-b7254f72bb20 tempest-AttachInterfacesUnderV243Test-498638506 tempest-AttachInterfacesUnderV243Test-498638506-project-member] Acquired lock "refresh_cache-0eab2200-2528-4a89-821a-cdd6a3c99297" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 735.633444] env[61923]: DEBUG nova.network.neutron [None req-f7b52f4c-b2e3-49f8-b056-b7254f72bb20 tempest-AttachInterfacesUnderV243Test-498638506 tempest-AttachInterfacesUnderV243Test-498638506-project-member] [instance: 0eab2200-2528-4a89-821a-cdd6a3c99297] Building network info cache for instance {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 735.798314] env[61923]: DEBUG nova.compute.manager [req-d23832b7-642f-48ce-916c-004edf8d3f1b req-f01fa3fc-e518-4ba0-9439-c28932347e79 service nova] [instance: 0eab2200-2528-4a89-821a-cdd6a3c99297] Received event network-changed-4c6d06bd-7574-4546-acd9-9d362bc379b1 {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 735.798508] env[61923]: DEBUG nova.compute.manager [req-d23832b7-642f-48ce-916c-004edf8d3f1b req-f01fa3fc-e518-4ba0-9439-c28932347e79 service nova] [instance: 0eab2200-2528-4a89-821a-cdd6a3c99297] Refreshing instance network info cache due to event network-changed-4c6d06bd-7574-4546-acd9-9d362bc379b1. {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 735.798690] env[61923]: DEBUG oslo_concurrency.lockutils [req-d23832b7-642f-48ce-916c-004edf8d3f1b req-f01fa3fc-e518-4ba0-9439-c28932347e79 service nova] Acquiring lock "refresh_cache-0eab2200-2528-4a89-821a-cdd6a3c99297" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 736.115855] env[61923]: DEBUG nova.compute.utils [None req-5bb1ca17-5442-4564-86ab-53f866a97e1d tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Using /dev/sd instead of None {{(pid=61923) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 736.117595] env[61923]: DEBUG nova.compute.manager [None req-5bb1ca17-5442-4564-86ab-53f866a97e1d tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: 15eaaa7c-77be-4b8b-89d0-0b2b11a0c49e] Allocating IP information in the background. {{(pid=61923) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 736.117763] env[61923]: DEBUG nova.network.neutron [None req-5bb1ca17-5442-4564-86ab-53f866a97e1d tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: 15eaaa7c-77be-4b8b-89d0-0b2b11a0c49e] allocate_for_instance() {{(pid=61923) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 736.159062] env[61923]: DEBUG nova.network.neutron [None req-f7b52f4c-b2e3-49f8-b056-b7254f72bb20 tempest-AttachInterfacesUnderV243Test-498638506 tempest-AttachInterfacesUnderV243Test-498638506-project-member] [instance: 0eab2200-2528-4a89-821a-cdd6a3c99297] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 736.237544] env[61923]: DEBUG nova.policy [None req-5bb1ca17-5442-4564-86ab-53f866a97e1d tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5b8a639e99914e75857b4571f0d58a3c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '69cc941eb6dd4780ac12aa29656c37f7', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61923) authorize /opt/stack/nova/nova/policy.py:201}} [ 736.280979] env[61923]: DEBUG nova.network.neutron [None req-f7b52f4c-b2e3-49f8-b056-b7254f72bb20 tempest-AttachInterfacesUnderV243Test-498638506 tempest-AttachInterfacesUnderV243Test-498638506-project-member] [instance: 0eab2200-2528-4a89-821a-cdd6a3c99297] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 736.467244] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24e65639-9374-4ecc-bb3b-fde73f3f146c {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.475788] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-002a9ae0-7f66-471b-8cf2-640e8ebee48b {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.516045] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a60516ed-7cb3-4a5e-889d-e13e230b0b05 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.524198] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f1ab4c9-cf7c-4ab5-83c1-1516407f6a40 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.538115] env[61923]: DEBUG nova.compute.provider_tree [None req-425a3606-915c-4f9b-8507-50f1c8d0d8d9 tempest-ServerRescueTestJSON-521767381 tempest-ServerRescueTestJSON-521767381-project-member] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 736.628679] env[61923]: DEBUG nova.compute.manager [None req-5bb1ca17-5442-4564-86ab-53f866a97e1d tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: 15eaaa7c-77be-4b8b-89d0-0b2b11a0c49e] Start building block device mappings for instance. {{(pid=61923) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 736.784508] env[61923]: DEBUG oslo_concurrency.lockutils [None req-f7b52f4c-b2e3-49f8-b056-b7254f72bb20 tempest-AttachInterfacesUnderV243Test-498638506 tempest-AttachInterfacesUnderV243Test-498638506-project-member] Releasing lock "refresh_cache-0eab2200-2528-4a89-821a-cdd6a3c99297" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 736.784979] env[61923]: DEBUG nova.compute.manager [None req-f7b52f4c-b2e3-49f8-b056-b7254f72bb20 tempest-AttachInterfacesUnderV243Test-498638506 tempest-AttachInterfacesUnderV243Test-498638506-project-member] [instance: 0eab2200-2528-4a89-821a-cdd6a3c99297] Start destroying the instance on the hypervisor. {{(pid=61923) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 736.785207] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-f7b52f4c-b2e3-49f8-b056-b7254f72bb20 tempest-AttachInterfacesUnderV243Test-498638506 tempest-AttachInterfacesUnderV243Test-498638506-project-member] [instance: 0eab2200-2528-4a89-821a-cdd6a3c99297] Destroying instance {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 736.785528] env[61923]: DEBUG oslo_concurrency.lockutils [req-d23832b7-642f-48ce-916c-004edf8d3f1b req-f01fa3fc-e518-4ba0-9439-c28932347e79 service nova] Acquired lock "refresh_cache-0eab2200-2528-4a89-821a-cdd6a3c99297" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 736.785697] env[61923]: DEBUG nova.network.neutron [req-d23832b7-642f-48ce-916c-004edf8d3f1b req-f01fa3fc-e518-4ba0-9439-c28932347e79 service nova] [instance: 0eab2200-2528-4a89-821a-cdd6a3c99297] Refreshing network info cache for port 4c6d06bd-7574-4546-acd9-9d362bc379b1 {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 736.786762] env[61923]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-9df0a55a-d6b1-40fc-8d28-348b94b85084 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.796418] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93c39ebb-9d28-441e-9dff-cf3e7b4945a3 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.822595] env[61923]: WARNING nova.virt.vmwareapi.vmops [None req-f7b52f4c-b2e3-49f8-b056-b7254f72bb20 tempest-AttachInterfacesUnderV243Test-498638506 tempest-AttachInterfacesUnderV243Test-498638506-project-member] [instance: 0eab2200-2528-4a89-821a-cdd6a3c99297] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 0eab2200-2528-4a89-821a-cdd6a3c99297 could not be found. [ 736.823049] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-f7b52f4c-b2e3-49f8-b056-b7254f72bb20 tempest-AttachInterfacesUnderV243Test-498638506 tempest-AttachInterfacesUnderV243Test-498638506-project-member] [instance: 0eab2200-2528-4a89-821a-cdd6a3c99297] Instance destroyed {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 736.823280] env[61923]: INFO nova.compute.manager [None req-f7b52f4c-b2e3-49f8-b056-b7254f72bb20 tempest-AttachInterfacesUnderV243Test-498638506 tempest-AttachInterfacesUnderV243Test-498638506-project-member] [instance: 0eab2200-2528-4a89-821a-cdd6a3c99297] Took 0.04 seconds to destroy the instance on the hypervisor. [ 736.823556] env[61923]: DEBUG oslo.service.loopingcall [None req-f7b52f4c-b2e3-49f8-b056-b7254f72bb20 tempest-AttachInterfacesUnderV243Test-498638506 tempest-AttachInterfacesUnderV243Test-498638506-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61923) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 736.823813] env[61923]: DEBUG nova.compute.manager [-] [instance: 0eab2200-2528-4a89-821a-cdd6a3c99297] Deallocating network for instance {{(pid=61923) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 736.823912] env[61923]: DEBUG nova.network.neutron [-] [instance: 0eab2200-2528-4a89-821a-cdd6a3c99297] deallocate_for_instance() {{(pid=61923) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 736.848356] env[61923]: DEBUG nova.network.neutron [None req-5bb1ca17-5442-4564-86ab-53f866a97e1d tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: 15eaaa7c-77be-4b8b-89d0-0b2b11a0c49e] Successfully created port: 439b85a4-1206-4dc7-9225-473a7f7704d9 {{(pid=61923) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 736.908149] env[61923]: DEBUG nova.network.neutron [-] [instance: 0eab2200-2528-4a89-821a-cdd6a3c99297] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 737.047099] env[61923]: DEBUG nova.scheduler.client.report [None req-425a3606-915c-4f9b-8507-50f1c8d0d8d9 tempest-ServerRescueTestJSON-521767381 tempest-ServerRescueTestJSON-521767381-project-member] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 737.312624] env[61923]: DEBUG nova.network.neutron [req-d23832b7-642f-48ce-916c-004edf8d3f1b req-f01fa3fc-e518-4ba0-9439-c28932347e79 service nova] [instance: 0eab2200-2528-4a89-821a-cdd6a3c99297] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 737.410468] env[61923]: DEBUG nova.network.neutron [-] [instance: 0eab2200-2528-4a89-821a-cdd6a3c99297] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 737.446712] env[61923]: DEBUG nova.network.neutron [req-d23832b7-642f-48ce-916c-004edf8d3f1b req-f01fa3fc-e518-4ba0-9439-c28932347e79 service nova] [instance: 0eab2200-2528-4a89-821a-cdd6a3c99297] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 737.552831] env[61923]: DEBUG oslo_concurrency.lockutils [None req-425a3606-915c-4f9b-8507-50f1c8d0d8d9 tempest-ServerRescueTestJSON-521767381 tempest-ServerRescueTestJSON-521767381-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.943s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 737.553577] env[61923]: ERROR nova.compute.manager [None req-425a3606-915c-4f9b-8507-50f1c8d0d8d9 tempest-ServerRescueTestJSON-521767381 tempest-ServerRescueTestJSON-521767381-project-member] [instance: 74b32299-983f-41f7-b3f4-3a12815a83e6] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 8065e5e4-32c9-4969-8aff-a6bc9953f4bc, please check neutron logs for more information. [ 737.553577] env[61923]: ERROR nova.compute.manager [instance: 74b32299-983f-41f7-b3f4-3a12815a83e6] Traceback (most recent call last): [ 737.553577] env[61923]: ERROR nova.compute.manager [instance: 74b32299-983f-41f7-b3f4-3a12815a83e6] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 737.553577] env[61923]: ERROR nova.compute.manager [instance: 74b32299-983f-41f7-b3f4-3a12815a83e6] self.driver.spawn(context, instance, image_meta, [ 737.553577] env[61923]: ERROR nova.compute.manager [instance: 74b32299-983f-41f7-b3f4-3a12815a83e6] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 737.553577] env[61923]: ERROR nova.compute.manager [instance: 74b32299-983f-41f7-b3f4-3a12815a83e6] self._vmops.spawn(context, instance, image_meta, injected_files, [ 737.553577] env[61923]: ERROR nova.compute.manager [instance: 74b32299-983f-41f7-b3f4-3a12815a83e6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 737.553577] env[61923]: ERROR nova.compute.manager [instance: 74b32299-983f-41f7-b3f4-3a12815a83e6] vm_ref = self.build_virtual_machine(instance, [ 737.553577] env[61923]: ERROR nova.compute.manager [instance: 74b32299-983f-41f7-b3f4-3a12815a83e6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 737.553577] env[61923]: ERROR nova.compute.manager [instance: 74b32299-983f-41f7-b3f4-3a12815a83e6] vif_infos = vmwarevif.get_vif_info(self._session, [ 737.553577] env[61923]: ERROR nova.compute.manager [instance: 74b32299-983f-41f7-b3f4-3a12815a83e6] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 737.553967] env[61923]: ERROR nova.compute.manager [instance: 74b32299-983f-41f7-b3f4-3a12815a83e6] for vif in network_info: [ 737.553967] env[61923]: ERROR nova.compute.manager [instance: 74b32299-983f-41f7-b3f4-3a12815a83e6] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 737.553967] env[61923]: ERROR nova.compute.manager [instance: 74b32299-983f-41f7-b3f4-3a12815a83e6] return self._sync_wrapper(fn, *args, **kwargs) [ 737.553967] env[61923]: ERROR nova.compute.manager [instance: 74b32299-983f-41f7-b3f4-3a12815a83e6] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 737.553967] env[61923]: ERROR nova.compute.manager [instance: 74b32299-983f-41f7-b3f4-3a12815a83e6] self.wait() [ 737.553967] env[61923]: ERROR nova.compute.manager [instance: 74b32299-983f-41f7-b3f4-3a12815a83e6] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 737.553967] env[61923]: ERROR nova.compute.manager [instance: 74b32299-983f-41f7-b3f4-3a12815a83e6] self[:] = self._gt.wait() [ 737.553967] env[61923]: ERROR nova.compute.manager [instance: 74b32299-983f-41f7-b3f4-3a12815a83e6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 737.553967] env[61923]: ERROR nova.compute.manager [instance: 74b32299-983f-41f7-b3f4-3a12815a83e6] return self._exit_event.wait() [ 737.553967] env[61923]: ERROR nova.compute.manager [instance: 74b32299-983f-41f7-b3f4-3a12815a83e6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 737.553967] env[61923]: ERROR nova.compute.manager [instance: 74b32299-983f-41f7-b3f4-3a12815a83e6] result = hub.switch() [ 737.553967] env[61923]: ERROR nova.compute.manager [instance: 74b32299-983f-41f7-b3f4-3a12815a83e6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 737.553967] env[61923]: ERROR nova.compute.manager [instance: 74b32299-983f-41f7-b3f4-3a12815a83e6] return self.greenlet.switch() [ 737.554380] env[61923]: ERROR nova.compute.manager [instance: 74b32299-983f-41f7-b3f4-3a12815a83e6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 737.554380] env[61923]: ERROR nova.compute.manager [instance: 74b32299-983f-41f7-b3f4-3a12815a83e6] result = function(*args, **kwargs) [ 737.554380] env[61923]: ERROR nova.compute.manager [instance: 74b32299-983f-41f7-b3f4-3a12815a83e6] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 737.554380] env[61923]: ERROR nova.compute.manager [instance: 74b32299-983f-41f7-b3f4-3a12815a83e6] return func(*args, **kwargs) [ 737.554380] env[61923]: ERROR nova.compute.manager [instance: 74b32299-983f-41f7-b3f4-3a12815a83e6] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 737.554380] env[61923]: ERROR nova.compute.manager [instance: 74b32299-983f-41f7-b3f4-3a12815a83e6] raise e [ 737.554380] env[61923]: ERROR nova.compute.manager [instance: 74b32299-983f-41f7-b3f4-3a12815a83e6] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 737.554380] env[61923]: ERROR nova.compute.manager [instance: 74b32299-983f-41f7-b3f4-3a12815a83e6] nwinfo = self.network_api.allocate_for_instance( [ 737.554380] env[61923]: ERROR nova.compute.manager [instance: 74b32299-983f-41f7-b3f4-3a12815a83e6] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 737.554380] env[61923]: ERROR nova.compute.manager [instance: 74b32299-983f-41f7-b3f4-3a12815a83e6] created_port_ids = self._update_ports_for_instance( [ 737.554380] env[61923]: ERROR nova.compute.manager [instance: 74b32299-983f-41f7-b3f4-3a12815a83e6] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 737.554380] env[61923]: ERROR nova.compute.manager [instance: 74b32299-983f-41f7-b3f4-3a12815a83e6] with excutils.save_and_reraise_exception(): [ 737.554380] env[61923]: ERROR nova.compute.manager [instance: 74b32299-983f-41f7-b3f4-3a12815a83e6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 737.554797] env[61923]: ERROR nova.compute.manager [instance: 74b32299-983f-41f7-b3f4-3a12815a83e6] self.force_reraise() [ 737.554797] env[61923]: ERROR nova.compute.manager [instance: 74b32299-983f-41f7-b3f4-3a12815a83e6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 737.554797] env[61923]: ERROR nova.compute.manager [instance: 74b32299-983f-41f7-b3f4-3a12815a83e6] raise self.value [ 737.554797] env[61923]: ERROR nova.compute.manager [instance: 74b32299-983f-41f7-b3f4-3a12815a83e6] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 737.554797] env[61923]: ERROR nova.compute.manager [instance: 74b32299-983f-41f7-b3f4-3a12815a83e6] updated_port = self._update_port( [ 737.554797] env[61923]: ERROR nova.compute.manager [instance: 74b32299-983f-41f7-b3f4-3a12815a83e6] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 737.554797] env[61923]: ERROR nova.compute.manager [instance: 74b32299-983f-41f7-b3f4-3a12815a83e6] _ensure_no_port_binding_failure(port) [ 737.554797] env[61923]: ERROR nova.compute.manager [instance: 74b32299-983f-41f7-b3f4-3a12815a83e6] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 737.554797] env[61923]: ERROR nova.compute.manager [instance: 74b32299-983f-41f7-b3f4-3a12815a83e6] raise exception.PortBindingFailed(port_id=port['id']) [ 737.554797] env[61923]: ERROR nova.compute.manager [instance: 74b32299-983f-41f7-b3f4-3a12815a83e6] nova.exception.PortBindingFailed: Binding failed for port 8065e5e4-32c9-4969-8aff-a6bc9953f4bc, please check neutron logs for more information. [ 737.554797] env[61923]: ERROR nova.compute.manager [instance: 74b32299-983f-41f7-b3f4-3a12815a83e6] [ 737.555176] env[61923]: DEBUG nova.compute.utils [None req-425a3606-915c-4f9b-8507-50f1c8d0d8d9 tempest-ServerRescueTestJSON-521767381 tempest-ServerRescueTestJSON-521767381-project-member] [instance: 74b32299-983f-41f7-b3f4-3a12815a83e6] Binding failed for port 8065e5e4-32c9-4969-8aff-a6bc9953f4bc, please check neutron logs for more information. {{(pid=61923) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 737.556513] env[61923]: DEBUG oslo_concurrency.lockutils [None req-0708225f-80bb-41ec-b3ed-7dd45d589b95 tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.257s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 737.560726] env[61923]: INFO nova.compute.claims [None req-0708225f-80bb-41ec-b3ed-7dd45d589b95 tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] [instance: 62d8a170-e7ad-4d14-90d8-6f6ce32b7cf8] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 737.568161] env[61923]: DEBUG nova.compute.manager [None req-425a3606-915c-4f9b-8507-50f1c8d0d8d9 tempest-ServerRescueTestJSON-521767381 tempest-ServerRescueTestJSON-521767381-project-member] [instance: 74b32299-983f-41f7-b3f4-3a12815a83e6] Build of instance 74b32299-983f-41f7-b3f4-3a12815a83e6 was re-scheduled: Binding failed for port 8065e5e4-32c9-4969-8aff-a6bc9953f4bc, please check neutron logs for more information. {{(pid=61923) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 737.568629] env[61923]: DEBUG nova.compute.manager [None req-425a3606-915c-4f9b-8507-50f1c8d0d8d9 tempest-ServerRescueTestJSON-521767381 tempest-ServerRescueTestJSON-521767381-project-member] [instance: 74b32299-983f-41f7-b3f4-3a12815a83e6] Unplugging VIFs for instance {{(pid=61923) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 737.568852] env[61923]: DEBUG oslo_concurrency.lockutils [None req-425a3606-915c-4f9b-8507-50f1c8d0d8d9 tempest-ServerRescueTestJSON-521767381 tempest-ServerRescueTestJSON-521767381-project-member] Acquiring lock "refresh_cache-74b32299-983f-41f7-b3f4-3a12815a83e6" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 737.569070] env[61923]: DEBUG oslo_concurrency.lockutils [None req-425a3606-915c-4f9b-8507-50f1c8d0d8d9 tempest-ServerRescueTestJSON-521767381 tempest-ServerRescueTestJSON-521767381-project-member] Acquired lock "refresh_cache-74b32299-983f-41f7-b3f4-3a12815a83e6" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 737.569208] env[61923]: DEBUG nova.network.neutron [None req-425a3606-915c-4f9b-8507-50f1c8d0d8d9 tempest-ServerRescueTestJSON-521767381 tempest-ServerRescueTestJSON-521767381-project-member] [instance: 74b32299-983f-41f7-b3f4-3a12815a83e6] Building network info cache for instance {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 737.590014] env[61923]: ERROR nova.compute.manager [None req-4a9f7dc2-7221-4911-9bc8-b39b7a22ddda tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 32585aee-c6bc-4a95-9516-646bb26bf0a0, please check neutron logs for more information. [ 737.590014] env[61923]: ERROR nova.compute.manager Traceback (most recent call last): [ 737.590014] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 737.590014] env[61923]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 737.590014] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 737.590014] env[61923]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 737.590014] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 737.590014] env[61923]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 737.590014] env[61923]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 737.590014] env[61923]: ERROR nova.compute.manager self.force_reraise() [ 737.590014] env[61923]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 737.590014] env[61923]: ERROR nova.compute.manager raise self.value [ 737.590014] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 737.590014] env[61923]: ERROR nova.compute.manager updated_port = self._update_port( [ 737.590014] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 737.590014] env[61923]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 737.590615] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 737.590615] env[61923]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 737.590615] env[61923]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 32585aee-c6bc-4a95-9516-646bb26bf0a0, please check neutron logs for more information. [ 737.590615] env[61923]: ERROR nova.compute.manager [ 737.590615] env[61923]: Traceback (most recent call last): [ 737.590615] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 737.590615] env[61923]: listener.cb(fileno) [ 737.590615] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 737.590615] env[61923]: result = function(*args, **kwargs) [ 737.590615] env[61923]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 737.590615] env[61923]: return func(*args, **kwargs) [ 737.590615] env[61923]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 737.590615] env[61923]: raise e [ 737.590615] env[61923]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 737.590615] env[61923]: nwinfo = self.network_api.allocate_for_instance( [ 737.590615] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 737.590615] env[61923]: created_port_ids = self._update_ports_for_instance( [ 737.590615] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 737.590615] env[61923]: with excutils.save_and_reraise_exception(): [ 737.590615] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 737.590615] env[61923]: self.force_reraise() [ 737.590615] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 737.590615] env[61923]: raise self.value [ 737.590615] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 737.590615] env[61923]: updated_port = self._update_port( [ 737.590615] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 737.590615] env[61923]: _ensure_no_port_binding_failure(port) [ 737.590615] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 737.590615] env[61923]: raise exception.PortBindingFailed(port_id=port['id']) [ 737.591571] env[61923]: nova.exception.PortBindingFailed: Binding failed for port 32585aee-c6bc-4a95-9516-646bb26bf0a0, please check neutron logs for more information. [ 737.591571] env[61923]: Removing descriptor: 14 [ 737.591571] env[61923]: ERROR nova.compute.manager [None req-4a9f7dc2-7221-4911-9bc8-b39b7a22ddda tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] [instance: 22d1e8ed-75c5-42e9-ae55-12d59e565914] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 32585aee-c6bc-4a95-9516-646bb26bf0a0, please check neutron logs for more information. [ 737.591571] env[61923]: ERROR nova.compute.manager [instance: 22d1e8ed-75c5-42e9-ae55-12d59e565914] Traceback (most recent call last): [ 737.591571] env[61923]: ERROR nova.compute.manager [instance: 22d1e8ed-75c5-42e9-ae55-12d59e565914] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 737.591571] env[61923]: ERROR nova.compute.manager [instance: 22d1e8ed-75c5-42e9-ae55-12d59e565914] yield resources [ 737.591571] env[61923]: ERROR nova.compute.manager [instance: 22d1e8ed-75c5-42e9-ae55-12d59e565914] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 737.591571] env[61923]: ERROR nova.compute.manager [instance: 22d1e8ed-75c5-42e9-ae55-12d59e565914] self.driver.spawn(context, instance, image_meta, [ 737.591571] env[61923]: ERROR nova.compute.manager [instance: 22d1e8ed-75c5-42e9-ae55-12d59e565914] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 737.591571] env[61923]: ERROR nova.compute.manager [instance: 22d1e8ed-75c5-42e9-ae55-12d59e565914] self._vmops.spawn(context, instance, image_meta, injected_files, [ 737.591571] env[61923]: ERROR nova.compute.manager [instance: 22d1e8ed-75c5-42e9-ae55-12d59e565914] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 737.591571] env[61923]: ERROR nova.compute.manager [instance: 22d1e8ed-75c5-42e9-ae55-12d59e565914] vm_ref = self.build_virtual_machine(instance, [ 737.592027] env[61923]: ERROR nova.compute.manager [instance: 22d1e8ed-75c5-42e9-ae55-12d59e565914] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 737.592027] env[61923]: ERROR nova.compute.manager [instance: 22d1e8ed-75c5-42e9-ae55-12d59e565914] vif_infos = vmwarevif.get_vif_info(self._session, [ 737.592027] env[61923]: ERROR nova.compute.manager [instance: 22d1e8ed-75c5-42e9-ae55-12d59e565914] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 737.592027] env[61923]: ERROR nova.compute.manager [instance: 22d1e8ed-75c5-42e9-ae55-12d59e565914] for vif in network_info: [ 737.592027] env[61923]: ERROR nova.compute.manager [instance: 22d1e8ed-75c5-42e9-ae55-12d59e565914] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 737.592027] env[61923]: ERROR nova.compute.manager [instance: 22d1e8ed-75c5-42e9-ae55-12d59e565914] return self._sync_wrapper(fn, *args, **kwargs) [ 737.592027] env[61923]: ERROR nova.compute.manager [instance: 22d1e8ed-75c5-42e9-ae55-12d59e565914] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 737.592027] env[61923]: ERROR nova.compute.manager [instance: 22d1e8ed-75c5-42e9-ae55-12d59e565914] self.wait() [ 737.592027] env[61923]: ERROR nova.compute.manager [instance: 22d1e8ed-75c5-42e9-ae55-12d59e565914] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 737.592027] env[61923]: ERROR nova.compute.manager [instance: 22d1e8ed-75c5-42e9-ae55-12d59e565914] self[:] = self._gt.wait() [ 737.592027] env[61923]: ERROR nova.compute.manager [instance: 22d1e8ed-75c5-42e9-ae55-12d59e565914] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 737.592027] env[61923]: ERROR nova.compute.manager [instance: 22d1e8ed-75c5-42e9-ae55-12d59e565914] return self._exit_event.wait() [ 737.592027] env[61923]: ERROR nova.compute.manager [instance: 22d1e8ed-75c5-42e9-ae55-12d59e565914] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 737.592416] env[61923]: ERROR nova.compute.manager [instance: 22d1e8ed-75c5-42e9-ae55-12d59e565914] result = hub.switch() [ 737.592416] env[61923]: ERROR nova.compute.manager [instance: 22d1e8ed-75c5-42e9-ae55-12d59e565914] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 737.592416] env[61923]: ERROR nova.compute.manager [instance: 22d1e8ed-75c5-42e9-ae55-12d59e565914] return self.greenlet.switch() [ 737.592416] env[61923]: ERROR nova.compute.manager [instance: 22d1e8ed-75c5-42e9-ae55-12d59e565914] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 737.592416] env[61923]: ERROR nova.compute.manager [instance: 22d1e8ed-75c5-42e9-ae55-12d59e565914] result = function(*args, **kwargs) [ 737.592416] env[61923]: ERROR nova.compute.manager [instance: 22d1e8ed-75c5-42e9-ae55-12d59e565914] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 737.592416] env[61923]: ERROR nova.compute.manager [instance: 22d1e8ed-75c5-42e9-ae55-12d59e565914] return func(*args, **kwargs) [ 737.592416] env[61923]: ERROR nova.compute.manager [instance: 22d1e8ed-75c5-42e9-ae55-12d59e565914] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 737.592416] env[61923]: ERROR nova.compute.manager [instance: 22d1e8ed-75c5-42e9-ae55-12d59e565914] raise e [ 737.592416] env[61923]: ERROR nova.compute.manager [instance: 22d1e8ed-75c5-42e9-ae55-12d59e565914] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 737.592416] env[61923]: ERROR nova.compute.manager [instance: 22d1e8ed-75c5-42e9-ae55-12d59e565914] nwinfo = self.network_api.allocate_for_instance( [ 737.592416] env[61923]: ERROR nova.compute.manager [instance: 22d1e8ed-75c5-42e9-ae55-12d59e565914] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 737.592416] env[61923]: ERROR nova.compute.manager [instance: 22d1e8ed-75c5-42e9-ae55-12d59e565914] created_port_ids = self._update_ports_for_instance( [ 737.592784] env[61923]: ERROR nova.compute.manager [instance: 22d1e8ed-75c5-42e9-ae55-12d59e565914] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 737.592784] env[61923]: ERROR nova.compute.manager [instance: 22d1e8ed-75c5-42e9-ae55-12d59e565914] with excutils.save_and_reraise_exception(): [ 737.592784] env[61923]: ERROR nova.compute.manager [instance: 22d1e8ed-75c5-42e9-ae55-12d59e565914] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 737.592784] env[61923]: ERROR nova.compute.manager [instance: 22d1e8ed-75c5-42e9-ae55-12d59e565914] self.force_reraise() [ 737.592784] env[61923]: ERROR nova.compute.manager [instance: 22d1e8ed-75c5-42e9-ae55-12d59e565914] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 737.592784] env[61923]: ERROR nova.compute.manager [instance: 22d1e8ed-75c5-42e9-ae55-12d59e565914] raise self.value [ 737.592784] env[61923]: ERROR nova.compute.manager [instance: 22d1e8ed-75c5-42e9-ae55-12d59e565914] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 737.592784] env[61923]: ERROR nova.compute.manager [instance: 22d1e8ed-75c5-42e9-ae55-12d59e565914] updated_port = self._update_port( [ 737.592784] env[61923]: ERROR nova.compute.manager [instance: 22d1e8ed-75c5-42e9-ae55-12d59e565914] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 737.592784] env[61923]: ERROR nova.compute.manager [instance: 22d1e8ed-75c5-42e9-ae55-12d59e565914] _ensure_no_port_binding_failure(port) [ 737.592784] env[61923]: ERROR nova.compute.manager [instance: 22d1e8ed-75c5-42e9-ae55-12d59e565914] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 737.592784] env[61923]: ERROR nova.compute.manager [instance: 22d1e8ed-75c5-42e9-ae55-12d59e565914] raise exception.PortBindingFailed(port_id=port['id']) [ 737.593126] env[61923]: ERROR nova.compute.manager [instance: 22d1e8ed-75c5-42e9-ae55-12d59e565914] nova.exception.PortBindingFailed: Binding failed for port 32585aee-c6bc-4a95-9516-646bb26bf0a0, please check neutron logs for more information. [ 737.593126] env[61923]: ERROR nova.compute.manager [instance: 22d1e8ed-75c5-42e9-ae55-12d59e565914] [ 737.593126] env[61923]: INFO nova.compute.manager [None req-4a9f7dc2-7221-4911-9bc8-b39b7a22ddda tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] [instance: 22d1e8ed-75c5-42e9-ae55-12d59e565914] Terminating instance [ 737.593620] env[61923]: DEBUG oslo_concurrency.lockutils [None req-4a9f7dc2-7221-4911-9bc8-b39b7a22ddda tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Acquiring lock "refresh_cache-22d1e8ed-75c5-42e9-ae55-12d59e565914" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 737.593776] env[61923]: DEBUG oslo_concurrency.lockutils [None req-4a9f7dc2-7221-4911-9bc8-b39b7a22ddda tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Acquired lock "refresh_cache-22d1e8ed-75c5-42e9-ae55-12d59e565914" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 737.593941] env[61923]: DEBUG nova.network.neutron [None req-4a9f7dc2-7221-4911-9bc8-b39b7a22ddda tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] [instance: 22d1e8ed-75c5-42e9-ae55-12d59e565914] Building network info cache for instance {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 737.639088] env[61923]: DEBUG nova.compute.manager [None req-5bb1ca17-5442-4564-86ab-53f866a97e1d tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: 15eaaa7c-77be-4b8b-89d0-0b2b11a0c49e] Start spawning the instance on the hypervisor. {{(pid=61923) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 737.682043] env[61923]: DEBUG nova.virt.hardware [None req-5bb1ca17-5442-4564-86ab-53f866a97e1d tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-29T20:07:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-29T20:07:35Z,direct_url=,disk_format='vmdk',id=0f153f63-ae0a-45b1-b7f5-7f9b673c947f,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='4e7b5e3b3c3443c8bd5c70f8a15739f5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-29T20:07:36Z,virtual_size=,visibility=), allow threads: False {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 737.682043] env[61923]: DEBUG nova.virt.hardware [None req-5bb1ca17-5442-4564-86ab-53f866a97e1d tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Flavor limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 737.682043] env[61923]: DEBUG nova.virt.hardware [None req-5bb1ca17-5442-4564-86ab-53f866a97e1d tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Image limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 737.682043] env[61923]: DEBUG nova.virt.hardware [None req-5bb1ca17-5442-4564-86ab-53f866a97e1d tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Flavor pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 737.682377] env[61923]: DEBUG nova.virt.hardware [None req-5bb1ca17-5442-4564-86ab-53f866a97e1d tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Image pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 737.682377] env[61923]: DEBUG nova.virt.hardware [None req-5bb1ca17-5442-4564-86ab-53f866a97e1d tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 737.682377] env[61923]: DEBUG nova.virt.hardware [None req-5bb1ca17-5442-4564-86ab-53f866a97e1d tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 737.682377] env[61923]: DEBUG nova.virt.hardware [None req-5bb1ca17-5442-4564-86ab-53f866a97e1d tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 737.682377] env[61923]: DEBUG nova.virt.hardware [None req-5bb1ca17-5442-4564-86ab-53f866a97e1d tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Got 1 possible topologies {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 737.682554] env[61923]: DEBUG nova.virt.hardware [None req-5bb1ca17-5442-4564-86ab-53f866a97e1d tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 737.682554] env[61923]: DEBUG nova.virt.hardware [None req-5bb1ca17-5442-4564-86ab-53f866a97e1d tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 737.682912] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52ca589a-4297-4691-b945-c01c4bc747a4 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.693206] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-350f6f05-f6e6-47a4-9527-656e2a275b83 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.870743] env[61923]: DEBUG oslo_concurrency.lockutils [None req-8e8efa20-a2ed-4da9-89f1-5ff799c0382c tempest-ServersNegativeTestMultiTenantJSON-1497120864 tempest-ServersNegativeTestMultiTenantJSON-1497120864-project-member] Acquiring lock "1d1b1dc4-c701-42b2-beba-344d622aef41" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 737.870743] env[61923]: DEBUG oslo_concurrency.lockutils [None req-8e8efa20-a2ed-4da9-89f1-5ff799c0382c tempest-ServersNegativeTestMultiTenantJSON-1497120864 tempest-ServersNegativeTestMultiTenantJSON-1497120864-project-member] Lock "1d1b1dc4-c701-42b2-beba-344d622aef41" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 737.917040] env[61923]: INFO nova.compute.manager [-] [instance: 0eab2200-2528-4a89-821a-cdd6a3c99297] Took 1.09 seconds to deallocate network for instance. [ 737.918475] env[61923]: DEBUG nova.compute.claims [None req-f7b52f4c-b2e3-49f8-b056-b7254f72bb20 tempest-AttachInterfacesUnderV243Test-498638506 tempest-AttachInterfacesUnderV243Test-498638506-project-member] [instance: 0eab2200-2528-4a89-821a-cdd6a3c99297] Aborting claim: {{(pid=61923) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 737.918940] env[61923]: DEBUG oslo_concurrency.lockutils [None req-f7b52f4c-b2e3-49f8-b056-b7254f72bb20 tempest-AttachInterfacesUnderV243Test-498638506 tempest-AttachInterfacesUnderV243Test-498638506-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 737.951731] env[61923]: DEBUG oslo_concurrency.lockutils [req-d23832b7-642f-48ce-916c-004edf8d3f1b req-f01fa3fc-e518-4ba0-9439-c28932347e79 service nova] Releasing lock "refresh_cache-0eab2200-2528-4a89-821a-cdd6a3c99297" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 737.952414] env[61923]: DEBUG nova.compute.manager [req-d23832b7-642f-48ce-916c-004edf8d3f1b req-f01fa3fc-e518-4ba0-9439-c28932347e79 service nova] [instance: 0eab2200-2528-4a89-821a-cdd6a3c99297] Received event network-vif-deleted-4c6d06bd-7574-4546-acd9-9d362bc379b1 {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 737.962747] env[61923]: DEBUG nova.compute.manager [req-0eb03853-755e-455e-86cf-38fd4212c8ba req-51260118-2de3-4167-a05f-af294ccdec9f service nova] [instance: 22d1e8ed-75c5-42e9-ae55-12d59e565914] Received event network-changed-32585aee-c6bc-4a95-9516-646bb26bf0a0 {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 737.963273] env[61923]: DEBUG nova.compute.manager [req-0eb03853-755e-455e-86cf-38fd4212c8ba req-51260118-2de3-4167-a05f-af294ccdec9f service nova] [instance: 22d1e8ed-75c5-42e9-ae55-12d59e565914] Refreshing instance network info cache due to event network-changed-32585aee-c6bc-4a95-9516-646bb26bf0a0. {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 737.963499] env[61923]: DEBUG oslo_concurrency.lockutils [req-0eb03853-755e-455e-86cf-38fd4212c8ba req-51260118-2de3-4167-a05f-af294ccdec9f service nova] Acquiring lock "refresh_cache-22d1e8ed-75c5-42e9-ae55-12d59e565914" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 738.100045] env[61923]: DEBUG nova.network.neutron [None req-425a3606-915c-4f9b-8507-50f1c8d0d8d9 tempest-ServerRescueTestJSON-521767381 tempest-ServerRescueTestJSON-521767381-project-member] [instance: 74b32299-983f-41f7-b3f4-3a12815a83e6] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 738.115694] env[61923]: DEBUG nova.network.neutron [None req-4a9f7dc2-7221-4911-9bc8-b39b7a22ddda tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] [instance: 22d1e8ed-75c5-42e9-ae55-12d59e565914] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 738.189177] env[61923]: DEBUG nova.network.neutron [None req-4a9f7dc2-7221-4911-9bc8-b39b7a22ddda tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] [instance: 22d1e8ed-75c5-42e9-ae55-12d59e565914] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 738.263355] env[61923]: DEBUG nova.network.neutron [None req-425a3606-915c-4f9b-8507-50f1c8d0d8d9 tempest-ServerRescueTestJSON-521767381 tempest-ServerRescueTestJSON-521767381-project-member] [instance: 74b32299-983f-41f7-b3f4-3a12815a83e6] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 738.605426] env[61923]: ERROR nova.compute.manager [None req-5bb1ca17-5442-4564-86ab-53f866a97e1d tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 439b85a4-1206-4dc7-9225-473a7f7704d9, please check neutron logs for more information. [ 738.605426] env[61923]: ERROR nova.compute.manager Traceback (most recent call last): [ 738.605426] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 738.605426] env[61923]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 738.605426] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 738.605426] env[61923]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 738.605426] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 738.605426] env[61923]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 738.605426] env[61923]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 738.605426] env[61923]: ERROR nova.compute.manager self.force_reraise() [ 738.605426] env[61923]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 738.605426] env[61923]: ERROR nova.compute.manager raise self.value [ 738.605426] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 738.605426] env[61923]: ERROR nova.compute.manager updated_port = self._update_port( [ 738.605426] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 738.605426] env[61923]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 738.605951] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 738.605951] env[61923]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 738.605951] env[61923]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 439b85a4-1206-4dc7-9225-473a7f7704d9, please check neutron logs for more information. [ 738.605951] env[61923]: ERROR nova.compute.manager [ 738.605951] env[61923]: Traceback (most recent call last): [ 738.605951] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 738.605951] env[61923]: listener.cb(fileno) [ 738.605951] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 738.605951] env[61923]: result = function(*args, **kwargs) [ 738.605951] env[61923]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 738.605951] env[61923]: return func(*args, **kwargs) [ 738.605951] env[61923]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 738.605951] env[61923]: raise e [ 738.605951] env[61923]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 738.605951] env[61923]: nwinfo = self.network_api.allocate_for_instance( [ 738.605951] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 738.605951] env[61923]: created_port_ids = self._update_ports_for_instance( [ 738.605951] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 738.605951] env[61923]: with excutils.save_and_reraise_exception(): [ 738.605951] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 738.605951] env[61923]: self.force_reraise() [ 738.605951] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 738.605951] env[61923]: raise self.value [ 738.605951] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 738.605951] env[61923]: updated_port = self._update_port( [ 738.605951] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 738.605951] env[61923]: _ensure_no_port_binding_failure(port) [ 738.605951] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 738.605951] env[61923]: raise exception.PortBindingFailed(port_id=port['id']) [ 738.606762] env[61923]: nova.exception.PortBindingFailed: Binding failed for port 439b85a4-1206-4dc7-9225-473a7f7704d9, please check neutron logs for more information. [ 738.606762] env[61923]: Removing descriptor: 17 [ 738.606762] env[61923]: ERROR nova.compute.manager [None req-5bb1ca17-5442-4564-86ab-53f866a97e1d tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: 15eaaa7c-77be-4b8b-89d0-0b2b11a0c49e] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 439b85a4-1206-4dc7-9225-473a7f7704d9, please check neutron logs for more information. [ 738.606762] env[61923]: ERROR nova.compute.manager [instance: 15eaaa7c-77be-4b8b-89d0-0b2b11a0c49e] Traceback (most recent call last): [ 738.606762] env[61923]: ERROR nova.compute.manager [instance: 15eaaa7c-77be-4b8b-89d0-0b2b11a0c49e] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 738.606762] env[61923]: ERROR nova.compute.manager [instance: 15eaaa7c-77be-4b8b-89d0-0b2b11a0c49e] yield resources [ 738.606762] env[61923]: ERROR nova.compute.manager [instance: 15eaaa7c-77be-4b8b-89d0-0b2b11a0c49e] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 738.606762] env[61923]: ERROR nova.compute.manager [instance: 15eaaa7c-77be-4b8b-89d0-0b2b11a0c49e] self.driver.spawn(context, instance, image_meta, [ 738.606762] env[61923]: ERROR nova.compute.manager [instance: 15eaaa7c-77be-4b8b-89d0-0b2b11a0c49e] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 738.606762] env[61923]: ERROR nova.compute.manager [instance: 15eaaa7c-77be-4b8b-89d0-0b2b11a0c49e] self._vmops.spawn(context, instance, image_meta, injected_files, [ 738.606762] env[61923]: ERROR nova.compute.manager [instance: 15eaaa7c-77be-4b8b-89d0-0b2b11a0c49e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 738.606762] env[61923]: ERROR nova.compute.manager [instance: 15eaaa7c-77be-4b8b-89d0-0b2b11a0c49e] vm_ref = self.build_virtual_machine(instance, [ 738.607343] env[61923]: ERROR nova.compute.manager [instance: 15eaaa7c-77be-4b8b-89d0-0b2b11a0c49e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 738.607343] env[61923]: ERROR nova.compute.manager [instance: 15eaaa7c-77be-4b8b-89d0-0b2b11a0c49e] vif_infos = vmwarevif.get_vif_info(self._session, [ 738.607343] env[61923]: ERROR nova.compute.manager [instance: 15eaaa7c-77be-4b8b-89d0-0b2b11a0c49e] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 738.607343] env[61923]: ERROR nova.compute.manager [instance: 15eaaa7c-77be-4b8b-89d0-0b2b11a0c49e] for vif in network_info: [ 738.607343] env[61923]: ERROR nova.compute.manager [instance: 15eaaa7c-77be-4b8b-89d0-0b2b11a0c49e] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 738.607343] env[61923]: ERROR nova.compute.manager [instance: 15eaaa7c-77be-4b8b-89d0-0b2b11a0c49e] return self._sync_wrapper(fn, *args, **kwargs) [ 738.607343] env[61923]: ERROR nova.compute.manager [instance: 15eaaa7c-77be-4b8b-89d0-0b2b11a0c49e] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 738.607343] env[61923]: ERROR nova.compute.manager [instance: 15eaaa7c-77be-4b8b-89d0-0b2b11a0c49e] self.wait() [ 738.607343] env[61923]: ERROR nova.compute.manager [instance: 15eaaa7c-77be-4b8b-89d0-0b2b11a0c49e] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 738.607343] env[61923]: ERROR nova.compute.manager [instance: 15eaaa7c-77be-4b8b-89d0-0b2b11a0c49e] self[:] = self._gt.wait() [ 738.607343] env[61923]: ERROR nova.compute.manager [instance: 15eaaa7c-77be-4b8b-89d0-0b2b11a0c49e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 738.607343] env[61923]: ERROR nova.compute.manager [instance: 15eaaa7c-77be-4b8b-89d0-0b2b11a0c49e] return self._exit_event.wait() [ 738.607343] env[61923]: ERROR nova.compute.manager [instance: 15eaaa7c-77be-4b8b-89d0-0b2b11a0c49e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 738.607701] env[61923]: ERROR nova.compute.manager [instance: 15eaaa7c-77be-4b8b-89d0-0b2b11a0c49e] result = hub.switch() [ 738.607701] env[61923]: ERROR nova.compute.manager [instance: 15eaaa7c-77be-4b8b-89d0-0b2b11a0c49e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 738.607701] env[61923]: ERROR nova.compute.manager [instance: 15eaaa7c-77be-4b8b-89d0-0b2b11a0c49e] return self.greenlet.switch() [ 738.607701] env[61923]: ERROR nova.compute.manager [instance: 15eaaa7c-77be-4b8b-89d0-0b2b11a0c49e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 738.607701] env[61923]: ERROR nova.compute.manager [instance: 15eaaa7c-77be-4b8b-89d0-0b2b11a0c49e] result = function(*args, **kwargs) [ 738.607701] env[61923]: ERROR nova.compute.manager [instance: 15eaaa7c-77be-4b8b-89d0-0b2b11a0c49e] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 738.607701] env[61923]: ERROR nova.compute.manager [instance: 15eaaa7c-77be-4b8b-89d0-0b2b11a0c49e] return func(*args, **kwargs) [ 738.607701] env[61923]: ERROR nova.compute.manager [instance: 15eaaa7c-77be-4b8b-89d0-0b2b11a0c49e] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 738.607701] env[61923]: ERROR nova.compute.manager [instance: 15eaaa7c-77be-4b8b-89d0-0b2b11a0c49e] raise e [ 738.607701] env[61923]: ERROR nova.compute.manager [instance: 15eaaa7c-77be-4b8b-89d0-0b2b11a0c49e] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 738.607701] env[61923]: ERROR nova.compute.manager [instance: 15eaaa7c-77be-4b8b-89d0-0b2b11a0c49e] nwinfo = self.network_api.allocate_for_instance( [ 738.607701] env[61923]: ERROR nova.compute.manager [instance: 15eaaa7c-77be-4b8b-89d0-0b2b11a0c49e] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 738.607701] env[61923]: ERROR nova.compute.manager [instance: 15eaaa7c-77be-4b8b-89d0-0b2b11a0c49e] created_port_ids = self._update_ports_for_instance( [ 738.608214] env[61923]: ERROR nova.compute.manager [instance: 15eaaa7c-77be-4b8b-89d0-0b2b11a0c49e] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 738.608214] env[61923]: ERROR nova.compute.manager [instance: 15eaaa7c-77be-4b8b-89d0-0b2b11a0c49e] with excutils.save_and_reraise_exception(): [ 738.608214] env[61923]: ERROR nova.compute.manager [instance: 15eaaa7c-77be-4b8b-89d0-0b2b11a0c49e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 738.608214] env[61923]: ERROR nova.compute.manager [instance: 15eaaa7c-77be-4b8b-89d0-0b2b11a0c49e] self.force_reraise() [ 738.608214] env[61923]: ERROR nova.compute.manager [instance: 15eaaa7c-77be-4b8b-89d0-0b2b11a0c49e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 738.608214] env[61923]: ERROR nova.compute.manager [instance: 15eaaa7c-77be-4b8b-89d0-0b2b11a0c49e] raise self.value [ 738.608214] env[61923]: ERROR nova.compute.manager [instance: 15eaaa7c-77be-4b8b-89d0-0b2b11a0c49e] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 738.608214] env[61923]: ERROR nova.compute.manager [instance: 15eaaa7c-77be-4b8b-89d0-0b2b11a0c49e] updated_port = self._update_port( [ 738.608214] env[61923]: ERROR nova.compute.manager [instance: 15eaaa7c-77be-4b8b-89d0-0b2b11a0c49e] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 738.608214] env[61923]: ERROR nova.compute.manager [instance: 15eaaa7c-77be-4b8b-89d0-0b2b11a0c49e] _ensure_no_port_binding_failure(port) [ 738.608214] env[61923]: ERROR nova.compute.manager [instance: 15eaaa7c-77be-4b8b-89d0-0b2b11a0c49e] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 738.608214] env[61923]: ERROR nova.compute.manager [instance: 15eaaa7c-77be-4b8b-89d0-0b2b11a0c49e] raise exception.PortBindingFailed(port_id=port['id']) [ 738.608702] env[61923]: ERROR nova.compute.manager [instance: 15eaaa7c-77be-4b8b-89d0-0b2b11a0c49e] nova.exception.PortBindingFailed: Binding failed for port 439b85a4-1206-4dc7-9225-473a7f7704d9, please check neutron logs for more information. [ 738.608702] env[61923]: ERROR nova.compute.manager [instance: 15eaaa7c-77be-4b8b-89d0-0b2b11a0c49e] [ 738.608702] env[61923]: INFO nova.compute.manager [None req-5bb1ca17-5442-4564-86ab-53f866a97e1d tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: 15eaaa7c-77be-4b8b-89d0-0b2b11a0c49e] Terminating instance [ 738.611820] env[61923]: DEBUG oslo_concurrency.lockutils [None req-5bb1ca17-5442-4564-86ab-53f866a97e1d tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Acquiring lock "refresh_cache-15eaaa7c-77be-4b8b-89d0-0b2b11a0c49e" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 738.611820] env[61923]: DEBUG oslo_concurrency.lockutils [None req-5bb1ca17-5442-4564-86ab-53f866a97e1d tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Acquired lock "refresh_cache-15eaaa7c-77be-4b8b-89d0-0b2b11a0c49e" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 738.611820] env[61923]: DEBUG nova.network.neutron [None req-5bb1ca17-5442-4564-86ab-53f866a97e1d tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: 15eaaa7c-77be-4b8b-89d0-0b2b11a0c49e] Building network info cache for instance {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 738.693606] env[61923]: DEBUG oslo_concurrency.lockutils [None req-4a9f7dc2-7221-4911-9bc8-b39b7a22ddda tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Releasing lock "refresh_cache-22d1e8ed-75c5-42e9-ae55-12d59e565914" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 738.694618] env[61923]: DEBUG nova.compute.manager [None req-4a9f7dc2-7221-4911-9bc8-b39b7a22ddda tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] [instance: 22d1e8ed-75c5-42e9-ae55-12d59e565914] Start destroying the instance on the hypervisor. {{(pid=61923) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 738.694618] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-4a9f7dc2-7221-4911-9bc8-b39b7a22ddda tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] [instance: 22d1e8ed-75c5-42e9-ae55-12d59e565914] Destroying instance {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 738.694895] env[61923]: DEBUG oslo_concurrency.lockutils [req-0eb03853-755e-455e-86cf-38fd4212c8ba req-51260118-2de3-4167-a05f-af294ccdec9f service nova] Acquired lock "refresh_cache-22d1e8ed-75c5-42e9-ae55-12d59e565914" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 738.695102] env[61923]: DEBUG nova.network.neutron [req-0eb03853-755e-455e-86cf-38fd4212c8ba req-51260118-2de3-4167-a05f-af294ccdec9f service nova] [instance: 22d1e8ed-75c5-42e9-ae55-12d59e565914] Refreshing network info cache for port 32585aee-c6bc-4a95-9516-646bb26bf0a0 {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 738.699441] env[61923]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ebc59a3b-d408-472f-bac4-1aa842ff7dc4 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.710830] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9c1d4ce-3d1a-40b5-9e7f-e0b50c195eab {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.735904] env[61923]: WARNING nova.virt.vmwareapi.vmops [None req-4a9f7dc2-7221-4911-9bc8-b39b7a22ddda tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] [instance: 22d1e8ed-75c5-42e9-ae55-12d59e565914] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 22d1e8ed-75c5-42e9-ae55-12d59e565914 could not be found. [ 738.736203] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-4a9f7dc2-7221-4911-9bc8-b39b7a22ddda tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] [instance: 22d1e8ed-75c5-42e9-ae55-12d59e565914] Instance destroyed {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 738.736422] env[61923]: INFO nova.compute.manager [None req-4a9f7dc2-7221-4911-9bc8-b39b7a22ddda tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] [instance: 22d1e8ed-75c5-42e9-ae55-12d59e565914] Took 0.04 seconds to destroy the instance on the hypervisor. [ 738.737233] env[61923]: DEBUG oslo.service.loopingcall [None req-4a9f7dc2-7221-4911-9bc8-b39b7a22ddda tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61923) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 738.737233] env[61923]: DEBUG nova.compute.manager [-] [instance: 22d1e8ed-75c5-42e9-ae55-12d59e565914] Deallocating network for instance {{(pid=61923) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 738.737233] env[61923]: DEBUG nova.network.neutron [-] [instance: 22d1e8ed-75c5-42e9-ae55-12d59e565914] deallocate_for_instance() {{(pid=61923) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 738.767598] env[61923]: DEBUG oslo_concurrency.lockutils [None req-425a3606-915c-4f9b-8507-50f1c8d0d8d9 tempest-ServerRescueTestJSON-521767381 tempest-ServerRescueTestJSON-521767381-project-member] Releasing lock "refresh_cache-74b32299-983f-41f7-b3f4-3a12815a83e6" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 738.767824] env[61923]: DEBUG nova.compute.manager [None req-425a3606-915c-4f9b-8507-50f1c8d0d8d9 tempest-ServerRescueTestJSON-521767381 tempest-ServerRescueTestJSON-521767381-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61923) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 738.768015] env[61923]: DEBUG nova.compute.manager [None req-425a3606-915c-4f9b-8507-50f1c8d0d8d9 tempest-ServerRescueTestJSON-521767381 tempest-ServerRescueTestJSON-521767381-project-member] [instance: 74b32299-983f-41f7-b3f4-3a12815a83e6] Deallocating network for instance {{(pid=61923) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 738.769078] env[61923]: DEBUG nova.network.neutron [None req-425a3606-915c-4f9b-8507-50f1c8d0d8d9 tempest-ServerRescueTestJSON-521767381 tempest-ServerRescueTestJSON-521767381-project-member] [instance: 74b32299-983f-41f7-b3f4-3a12815a83e6] deallocate_for_instance() {{(pid=61923) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 738.770902] env[61923]: DEBUG nova.network.neutron [-] [instance: 22d1e8ed-75c5-42e9-ae55-12d59e565914] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 738.791365] env[61923]: DEBUG nova.network.neutron [None req-425a3606-915c-4f9b-8507-50f1c8d0d8d9 tempest-ServerRescueTestJSON-521767381 tempest-ServerRescueTestJSON-521767381-project-member] [instance: 74b32299-983f-41f7-b3f4-3a12815a83e6] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 739.014324] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d310fdf-6a3d-434d-827d-3cbdcf4414dd {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.021876] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12109f3e-ff91-40b2-a064-8e402506ffeb {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.056981] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff065c38-4355-4ae7-9ce9-762863c0967d {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.068145] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-259d2042-b523-4144-83de-42ac8d1770b7 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.082566] env[61923]: DEBUG nova.compute.provider_tree [None req-0708225f-80bb-41ec-b3ed-7dd45d589b95 tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 739.137264] env[61923]: DEBUG nova.network.neutron [None req-5bb1ca17-5442-4564-86ab-53f866a97e1d tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: 15eaaa7c-77be-4b8b-89d0-0b2b11a0c49e] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 739.226289] env[61923]: DEBUG nova.network.neutron [req-0eb03853-755e-455e-86cf-38fd4212c8ba req-51260118-2de3-4167-a05f-af294ccdec9f service nova] [instance: 22d1e8ed-75c5-42e9-ae55-12d59e565914] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 739.274936] env[61923]: DEBUG nova.network.neutron [-] [instance: 22d1e8ed-75c5-42e9-ae55-12d59e565914] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 739.296207] env[61923]: DEBUG nova.network.neutron [None req-425a3606-915c-4f9b-8507-50f1c8d0d8d9 tempest-ServerRescueTestJSON-521767381 tempest-ServerRescueTestJSON-521767381-project-member] [instance: 74b32299-983f-41f7-b3f4-3a12815a83e6] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 739.333925] env[61923]: DEBUG nova.network.neutron [None req-5bb1ca17-5442-4564-86ab-53f866a97e1d tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: 15eaaa7c-77be-4b8b-89d0-0b2b11a0c49e] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 739.423864] env[61923]: DEBUG nova.network.neutron [req-0eb03853-755e-455e-86cf-38fd4212c8ba req-51260118-2de3-4167-a05f-af294ccdec9f service nova] [instance: 22d1e8ed-75c5-42e9-ae55-12d59e565914] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 739.586484] env[61923]: DEBUG nova.scheduler.client.report [None req-0708225f-80bb-41ec-b3ed-7dd45d589b95 tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 739.780193] env[61923]: INFO nova.compute.manager [-] [instance: 22d1e8ed-75c5-42e9-ae55-12d59e565914] Took 1.04 seconds to deallocate network for instance. [ 739.782500] env[61923]: DEBUG nova.compute.claims [None req-4a9f7dc2-7221-4911-9bc8-b39b7a22ddda tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] [instance: 22d1e8ed-75c5-42e9-ae55-12d59e565914] Aborting claim: {{(pid=61923) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 739.782500] env[61923]: DEBUG oslo_concurrency.lockutils [None req-4a9f7dc2-7221-4911-9bc8-b39b7a22ddda tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 739.798462] env[61923]: INFO nova.compute.manager [None req-425a3606-915c-4f9b-8507-50f1c8d0d8d9 tempest-ServerRescueTestJSON-521767381 tempest-ServerRescueTestJSON-521767381-project-member] [instance: 74b32299-983f-41f7-b3f4-3a12815a83e6] Took 1.03 seconds to deallocate network for instance. [ 739.835513] env[61923]: DEBUG oslo_concurrency.lockutils [None req-5bb1ca17-5442-4564-86ab-53f866a97e1d tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Releasing lock "refresh_cache-15eaaa7c-77be-4b8b-89d0-0b2b11a0c49e" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 739.835957] env[61923]: DEBUG nova.compute.manager [None req-5bb1ca17-5442-4564-86ab-53f866a97e1d tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: 15eaaa7c-77be-4b8b-89d0-0b2b11a0c49e] Start destroying the instance on the hypervisor. {{(pid=61923) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 739.836218] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-5bb1ca17-5442-4564-86ab-53f866a97e1d tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: 15eaaa7c-77be-4b8b-89d0-0b2b11a0c49e] Destroying instance {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 739.836534] env[61923]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-7f3e5f19-e252-42c6-a7b8-9431d682d31c {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.846187] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd26319b-60e6-437c-947b-f007476ea417 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.867818] env[61923]: WARNING nova.virt.vmwareapi.vmops [None req-5bb1ca17-5442-4564-86ab-53f866a97e1d tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: 15eaaa7c-77be-4b8b-89d0-0b2b11a0c49e] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 15eaaa7c-77be-4b8b-89d0-0b2b11a0c49e could not be found. [ 739.868064] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-5bb1ca17-5442-4564-86ab-53f866a97e1d tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: 15eaaa7c-77be-4b8b-89d0-0b2b11a0c49e] Instance destroyed {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 739.868254] env[61923]: INFO nova.compute.manager [None req-5bb1ca17-5442-4564-86ab-53f866a97e1d tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: 15eaaa7c-77be-4b8b-89d0-0b2b11a0c49e] Took 0.03 seconds to destroy the instance on the hypervisor. [ 739.868497] env[61923]: DEBUG oslo.service.loopingcall [None req-5bb1ca17-5442-4564-86ab-53f866a97e1d tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61923) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 739.869337] env[61923]: DEBUG nova.compute.manager [-] [instance: 15eaaa7c-77be-4b8b-89d0-0b2b11a0c49e] Deallocating network for instance {{(pid=61923) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 739.869434] env[61923]: DEBUG nova.network.neutron [-] [instance: 15eaaa7c-77be-4b8b-89d0-0b2b11a0c49e] deallocate_for_instance() {{(pid=61923) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 739.882744] env[61923]: DEBUG nova.network.neutron [-] [instance: 15eaaa7c-77be-4b8b-89d0-0b2b11a0c49e] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 739.926854] env[61923]: DEBUG oslo_concurrency.lockutils [req-0eb03853-755e-455e-86cf-38fd4212c8ba req-51260118-2de3-4167-a05f-af294ccdec9f service nova] Releasing lock "refresh_cache-22d1e8ed-75c5-42e9-ae55-12d59e565914" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 739.927130] env[61923]: DEBUG nova.compute.manager [req-0eb03853-755e-455e-86cf-38fd4212c8ba req-51260118-2de3-4167-a05f-af294ccdec9f service nova] [instance: 22d1e8ed-75c5-42e9-ae55-12d59e565914] Received event network-vif-deleted-32585aee-c6bc-4a95-9516-646bb26bf0a0 {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 739.990331] env[61923]: DEBUG nova.compute.manager [req-4262cb7f-380a-4ab4-a3ed-f9457f0142af req-97a10385-449b-47ad-a99c-b696b0f037eb service nova] [instance: 15eaaa7c-77be-4b8b-89d0-0b2b11a0c49e] Received event network-changed-439b85a4-1206-4dc7-9225-473a7f7704d9 {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 739.990567] env[61923]: DEBUG nova.compute.manager [req-4262cb7f-380a-4ab4-a3ed-f9457f0142af req-97a10385-449b-47ad-a99c-b696b0f037eb service nova] [instance: 15eaaa7c-77be-4b8b-89d0-0b2b11a0c49e] Refreshing instance network info cache due to event network-changed-439b85a4-1206-4dc7-9225-473a7f7704d9. {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 739.990726] env[61923]: DEBUG oslo_concurrency.lockutils [req-4262cb7f-380a-4ab4-a3ed-f9457f0142af req-97a10385-449b-47ad-a99c-b696b0f037eb service nova] Acquiring lock "refresh_cache-15eaaa7c-77be-4b8b-89d0-0b2b11a0c49e" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 739.990864] env[61923]: DEBUG oslo_concurrency.lockutils [req-4262cb7f-380a-4ab4-a3ed-f9457f0142af req-97a10385-449b-47ad-a99c-b696b0f037eb service nova] Acquired lock "refresh_cache-15eaaa7c-77be-4b8b-89d0-0b2b11a0c49e" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 739.991068] env[61923]: DEBUG nova.network.neutron [req-4262cb7f-380a-4ab4-a3ed-f9457f0142af req-97a10385-449b-47ad-a99c-b696b0f037eb service nova] [instance: 15eaaa7c-77be-4b8b-89d0-0b2b11a0c49e] Refreshing network info cache for port 439b85a4-1206-4dc7-9225-473a7f7704d9 {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 740.091769] env[61923]: DEBUG oslo_concurrency.lockutils [None req-0708225f-80bb-41ec-b3ed-7dd45d589b95 tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.535s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 740.092306] env[61923]: DEBUG nova.compute.manager [None req-0708225f-80bb-41ec-b3ed-7dd45d589b95 tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] [instance: 62d8a170-e7ad-4d14-90d8-6f6ce32b7cf8] Start building networks asynchronously for instance. {{(pid=61923) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 740.097472] env[61923]: DEBUG oslo_concurrency.lockutils [None req-ab7b3da0-3a01-44d4-9dfd-3e5d2ad27432 tempest-TenantUsagesTestJSON-871253541 tempest-TenantUsagesTestJSON-871253541-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.890s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 740.097472] env[61923]: INFO nova.compute.claims [None req-ab7b3da0-3a01-44d4-9dfd-3e5d2ad27432 tempest-TenantUsagesTestJSON-871253541 tempest-TenantUsagesTestJSON-871253541-project-member] [instance: b6712878-3270-43f4-9518-d41eb77124b7] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 740.185008] env[61923]: DEBUG oslo_concurrency.lockutils [None req-458eed44-c3ca-4049-bcef-fbcaf60c50ed tempest-ServersTestManualDisk-2012652654 tempest-ServersTestManualDisk-2012652654-project-member] Acquiring lock "f1839f61-4314-48fe-8ab6-14b5e49d569d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 740.185258] env[61923]: DEBUG oslo_concurrency.lockutils [None req-458eed44-c3ca-4049-bcef-fbcaf60c50ed tempest-ServersTestManualDisk-2012652654 tempest-ServersTestManualDisk-2012652654-project-member] Lock "f1839f61-4314-48fe-8ab6-14b5e49d569d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 740.384624] env[61923]: DEBUG nova.network.neutron [-] [instance: 15eaaa7c-77be-4b8b-89d0-0b2b11a0c49e] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 740.519776] env[61923]: DEBUG nova.network.neutron [req-4262cb7f-380a-4ab4-a3ed-f9457f0142af req-97a10385-449b-47ad-a99c-b696b0f037eb service nova] [instance: 15eaaa7c-77be-4b8b-89d0-0b2b11a0c49e] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 740.601261] env[61923]: DEBUG nova.compute.utils [None req-0708225f-80bb-41ec-b3ed-7dd45d589b95 tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Using /dev/sd instead of None {{(pid=61923) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 740.605065] env[61923]: DEBUG nova.compute.manager [None req-0708225f-80bb-41ec-b3ed-7dd45d589b95 tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] [instance: 62d8a170-e7ad-4d14-90d8-6f6ce32b7cf8] Allocating IP information in the background. {{(pid=61923) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 740.605065] env[61923]: DEBUG nova.network.neutron [None req-0708225f-80bb-41ec-b3ed-7dd45d589b95 tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] [instance: 62d8a170-e7ad-4d14-90d8-6f6ce32b7cf8] allocate_for_instance() {{(pid=61923) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 740.611773] env[61923]: DEBUG nova.network.neutron [req-4262cb7f-380a-4ab4-a3ed-f9457f0142af req-97a10385-449b-47ad-a99c-b696b0f037eb service nova] [instance: 15eaaa7c-77be-4b8b-89d0-0b2b11a0c49e] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 740.660270] env[61923]: DEBUG nova.policy [None req-0708225f-80bb-41ec-b3ed-7dd45d589b95 tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'eb00f8aff92b44e9a0e739366a1d2662', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7abfe95660904cd4b8824257795d8e8a', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61923) authorize /opt/stack/nova/nova/policy.py:201}} [ 740.827849] env[61923]: INFO nova.scheduler.client.report [None req-425a3606-915c-4f9b-8507-50f1c8d0d8d9 tempest-ServerRescueTestJSON-521767381 tempest-ServerRescueTestJSON-521767381-project-member] Deleted allocations for instance 74b32299-983f-41f7-b3f4-3a12815a83e6 [ 740.888182] env[61923]: INFO nova.compute.manager [-] [instance: 15eaaa7c-77be-4b8b-89d0-0b2b11a0c49e] Took 1.02 seconds to deallocate network for instance. [ 740.892871] env[61923]: DEBUG nova.compute.claims [None req-5bb1ca17-5442-4564-86ab-53f866a97e1d tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: 15eaaa7c-77be-4b8b-89d0-0b2b11a0c49e] Aborting claim: {{(pid=61923) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 740.893057] env[61923]: DEBUG oslo_concurrency.lockutils [None req-5bb1ca17-5442-4564-86ab-53f866a97e1d tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 741.036246] env[61923]: DEBUG nova.network.neutron [None req-0708225f-80bb-41ec-b3ed-7dd45d589b95 tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] [instance: 62d8a170-e7ad-4d14-90d8-6f6ce32b7cf8] Successfully created port: fbdf8b6e-0d86-4121-8aeb-b977a59f1186 {{(pid=61923) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 741.105220] env[61923]: DEBUG nova.compute.manager [None req-0708225f-80bb-41ec-b3ed-7dd45d589b95 tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] [instance: 62d8a170-e7ad-4d14-90d8-6f6ce32b7cf8] Start building block device mappings for instance. {{(pid=61923) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 741.119432] env[61923]: DEBUG oslo_concurrency.lockutils [req-4262cb7f-380a-4ab4-a3ed-f9457f0142af req-97a10385-449b-47ad-a99c-b696b0f037eb service nova] Releasing lock "refresh_cache-15eaaa7c-77be-4b8b-89d0-0b2b11a0c49e" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 741.119432] env[61923]: DEBUG nova.compute.manager [req-4262cb7f-380a-4ab4-a3ed-f9457f0142af req-97a10385-449b-47ad-a99c-b696b0f037eb service nova] [instance: 15eaaa7c-77be-4b8b-89d0-0b2b11a0c49e] Received event network-vif-deleted-439b85a4-1206-4dc7-9225-473a7f7704d9 {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 741.339665] env[61923]: DEBUG oslo_concurrency.lockutils [None req-425a3606-915c-4f9b-8507-50f1c8d0d8d9 tempest-ServerRescueTestJSON-521767381 tempest-ServerRescueTestJSON-521767381-project-member] Lock "74b32299-983f-41f7-b3f4-3a12815a83e6" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 140.551s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 741.423429] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15245326-c8aa-4d4b-9fd8-223979b5607d {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.431637] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63785200-5581-42a4-8b6f-9f0dc6828c31 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.465518] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6d44068-be45-4b1d-a225-1c9a509cd376 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.473676] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0fa0792-c68c-4f15-a8f5-8545d150daef {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.489528] env[61923]: DEBUG nova.compute.provider_tree [None req-ab7b3da0-3a01-44d4-9dfd-3e5d2ad27432 tempest-TenantUsagesTestJSON-871253541 tempest-TenantUsagesTestJSON-871253541-project-member] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 741.845388] env[61923]: DEBUG nova.compute.manager [None req-347b83bf-b4ed-4904-9bd3-d6021183ae1c tempest-ServerShowV254Test-921228709 tempest-ServerShowV254Test-921228709-project-member] [instance: ecedb77f-4a2a-4047-964f-80796954d847] Starting instance... {{(pid=61923) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 741.994644] env[61923]: DEBUG nova.scheduler.client.report [None req-ab7b3da0-3a01-44d4-9dfd-3e5d2ad27432 tempest-TenantUsagesTestJSON-871253541 tempest-TenantUsagesTestJSON-871253541-project-member] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 742.025685] env[61923]: DEBUG nova.compute.manager [req-c054fa0f-e84e-49a3-8e3e-ff33152cec08 req-20c76a27-cda2-4c0f-8c83-cb9027e6acda service nova] [instance: 62d8a170-e7ad-4d14-90d8-6f6ce32b7cf8] Received event network-changed-fbdf8b6e-0d86-4121-8aeb-b977a59f1186 {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 742.026079] env[61923]: DEBUG nova.compute.manager [req-c054fa0f-e84e-49a3-8e3e-ff33152cec08 req-20c76a27-cda2-4c0f-8c83-cb9027e6acda service nova] [instance: 62d8a170-e7ad-4d14-90d8-6f6ce32b7cf8] Refreshing instance network info cache due to event network-changed-fbdf8b6e-0d86-4121-8aeb-b977a59f1186. {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 742.026546] env[61923]: DEBUG oslo_concurrency.lockutils [req-c054fa0f-e84e-49a3-8e3e-ff33152cec08 req-20c76a27-cda2-4c0f-8c83-cb9027e6acda service nova] Acquiring lock "refresh_cache-62d8a170-e7ad-4d14-90d8-6f6ce32b7cf8" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 742.027017] env[61923]: DEBUG oslo_concurrency.lockutils [req-c054fa0f-e84e-49a3-8e3e-ff33152cec08 req-20c76a27-cda2-4c0f-8c83-cb9027e6acda service nova] Acquired lock "refresh_cache-62d8a170-e7ad-4d14-90d8-6f6ce32b7cf8" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 742.027119] env[61923]: DEBUG nova.network.neutron [req-c054fa0f-e84e-49a3-8e3e-ff33152cec08 req-20c76a27-cda2-4c0f-8c83-cb9027e6acda service nova] [instance: 62d8a170-e7ad-4d14-90d8-6f6ce32b7cf8] Refreshing network info cache for port fbdf8b6e-0d86-4121-8aeb-b977a59f1186 {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 742.129804] env[61923]: DEBUG nova.compute.manager [None req-0708225f-80bb-41ec-b3ed-7dd45d589b95 tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] [instance: 62d8a170-e7ad-4d14-90d8-6f6ce32b7cf8] Start spawning the instance on the hypervisor. {{(pid=61923) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 742.154245] env[61923]: DEBUG nova.virt.hardware [None req-0708225f-80bb-41ec-b3ed-7dd45d589b95 tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-29T20:07:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-29T20:07:35Z,direct_url=,disk_format='vmdk',id=0f153f63-ae0a-45b1-b7f5-7f9b673c947f,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='4e7b5e3b3c3443c8bd5c70f8a15739f5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-29T20:07:36Z,virtual_size=,visibility=), allow threads: False {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 742.154699] env[61923]: DEBUG nova.virt.hardware [None req-0708225f-80bb-41ec-b3ed-7dd45d589b95 tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Flavor limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 742.154978] env[61923]: DEBUG nova.virt.hardware [None req-0708225f-80bb-41ec-b3ed-7dd45d589b95 tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Image limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 742.159018] env[61923]: DEBUG nova.virt.hardware [None req-0708225f-80bb-41ec-b3ed-7dd45d589b95 tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Flavor pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 742.159018] env[61923]: DEBUG nova.virt.hardware [None req-0708225f-80bb-41ec-b3ed-7dd45d589b95 tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Image pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 742.159018] env[61923]: DEBUG nova.virt.hardware [None req-0708225f-80bb-41ec-b3ed-7dd45d589b95 tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 742.159018] env[61923]: DEBUG nova.virt.hardware [None req-0708225f-80bb-41ec-b3ed-7dd45d589b95 tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 742.159018] env[61923]: DEBUG nova.virt.hardware [None req-0708225f-80bb-41ec-b3ed-7dd45d589b95 tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 742.159018] env[61923]: DEBUG nova.virt.hardware [None req-0708225f-80bb-41ec-b3ed-7dd45d589b95 tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Got 1 possible topologies {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 742.159410] env[61923]: DEBUG nova.virt.hardware [None req-0708225f-80bb-41ec-b3ed-7dd45d589b95 tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 742.159410] env[61923]: DEBUG nova.virt.hardware [None req-0708225f-80bb-41ec-b3ed-7dd45d589b95 tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 742.159410] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13d18924-535f-400f-a98b-e5df928cd85e {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.166605] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e702aa8-a4cd-43f1-bd8e-e8a416a1e5bd {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.189788] env[61923]: ERROR nova.compute.manager [None req-0708225f-80bb-41ec-b3ed-7dd45d589b95 tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port fbdf8b6e-0d86-4121-8aeb-b977a59f1186, please check neutron logs for more information. [ 742.189788] env[61923]: ERROR nova.compute.manager Traceback (most recent call last): [ 742.189788] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 742.189788] env[61923]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 742.189788] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 742.189788] env[61923]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 742.189788] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 742.189788] env[61923]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 742.189788] env[61923]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 742.189788] env[61923]: ERROR nova.compute.manager self.force_reraise() [ 742.189788] env[61923]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 742.189788] env[61923]: ERROR nova.compute.manager raise self.value [ 742.189788] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 742.189788] env[61923]: ERROR nova.compute.manager updated_port = self._update_port( [ 742.189788] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 742.189788] env[61923]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 742.190340] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 742.190340] env[61923]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 742.190340] env[61923]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port fbdf8b6e-0d86-4121-8aeb-b977a59f1186, please check neutron logs for more information. [ 742.190340] env[61923]: ERROR nova.compute.manager [ 742.190340] env[61923]: Traceback (most recent call last): [ 742.190340] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 742.190340] env[61923]: listener.cb(fileno) [ 742.190340] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 742.190340] env[61923]: result = function(*args, **kwargs) [ 742.190340] env[61923]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 742.190340] env[61923]: return func(*args, **kwargs) [ 742.190340] env[61923]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 742.190340] env[61923]: raise e [ 742.190340] env[61923]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 742.190340] env[61923]: nwinfo = self.network_api.allocate_for_instance( [ 742.190340] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 742.190340] env[61923]: created_port_ids = self._update_ports_for_instance( [ 742.190340] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 742.190340] env[61923]: with excutils.save_and_reraise_exception(): [ 742.190340] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 742.190340] env[61923]: self.force_reraise() [ 742.190340] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 742.190340] env[61923]: raise self.value [ 742.190340] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 742.190340] env[61923]: updated_port = self._update_port( [ 742.190340] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 742.190340] env[61923]: _ensure_no_port_binding_failure(port) [ 742.190340] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 742.190340] env[61923]: raise exception.PortBindingFailed(port_id=port['id']) [ 742.191279] env[61923]: nova.exception.PortBindingFailed: Binding failed for port fbdf8b6e-0d86-4121-8aeb-b977a59f1186, please check neutron logs for more information. [ 742.191279] env[61923]: Removing descriptor: 14 [ 742.191279] env[61923]: ERROR nova.compute.manager [None req-0708225f-80bb-41ec-b3ed-7dd45d589b95 tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] [instance: 62d8a170-e7ad-4d14-90d8-6f6ce32b7cf8] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port fbdf8b6e-0d86-4121-8aeb-b977a59f1186, please check neutron logs for more information. [ 742.191279] env[61923]: ERROR nova.compute.manager [instance: 62d8a170-e7ad-4d14-90d8-6f6ce32b7cf8] Traceback (most recent call last): [ 742.191279] env[61923]: ERROR nova.compute.manager [instance: 62d8a170-e7ad-4d14-90d8-6f6ce32b7cf8] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 742.191279] env[61923]: ERROR nova.compute.manager [instance: 62d8a170-e7ad-4d14-90d8-6f6ce32b7cf8] yield resources [ 742.191279] env[61923]: ERROR nova.compute.manager [instance: 62d8a170-e7ad-4d14-90d8-6f6ce32b7cf8] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 742.191279] env[61923]: ERROR nova.compute.manager [instance: 62d8a170-e7ad-4d14-90d8-6f6ce32b7cf8] self.driver.spawn(context, instance, image_meta, [ 742.191279] env[61923]: ERROR nova.compute.manager [instance: 62d8a170-e7ad-4d14-90d8-6f6ce32b7cf8] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 742.191279] env[61923]: ERROR nova.compute.manager [instance: 62d8a170-e7ad-4d14-90d8-6f6ce32b7cf8] self._vmops.spawn(context, instance, image_meta, injected_files, [ 742.191279] env[61923]: ERROR nova.compute.manager [instance: 62d8a170-e7ad-4d14-90d8-6f6ce32b7cf8] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 742.191279] env[61923]: ERROR nova.compute.manager [instance: 62d8a170-e7ad-4d14-90d8-6f6ce32b7cf8] vm_ref = self.build_virtual_machine(instance, [ 742.191664] env[61923]: ERROR nova.compute.manager [instance: 62d8a170-e7ad-4d14-90d8-6f6ce32b7cf8] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 742.191664] env[61923]: ERROR nova.compute.manager [instance: 62d8a170-e7ad-4d14-90d8-6f6ce32b7cf8] vif_infos = vmwarevif.get_vif_info(self._session, [ 742.191664] env[61923]: ERROR nova.compute.manager [instance: 62d8a170-e7ad-4d14-90d8-6f6ce32b7cf8] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 742.191664] env[61923]: ERROR nova.compute.manager [instance: 62d8a170-e7ad-4d14-90d8-6f6ce32b7cf8] for vif in network_info: [ 742.191664] env[61923]: ERROR nova.compute.manager [instance: 62d8a170-e7ad-4d14-90d8-6f6ce32b7cf8] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 742.191664] env[61923]: ERROR nova.compute.manager [instance: 62d8a170-e7ad-4d14-90d8-6f6ce32b7cf8] return self._sync_wrapper(fn, *args, **kwargs) [ 742.191664] env[61923]: ERROR nova.compute.manager [instance: 62d8a170-e7ad-4d14-90d8-6f6ce32b7cf8] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 742.191664] env[61923]: ERROR nova.compute.manager [instance: 62d8a170-e7ad-4d14-90d8-6f6ce32b7cf8] self.wait() [ 742.191664] env[61923]: ERROR nova.compute.manager [instance: 62d8a170-e7ad-4d14-90d8-6f6ce32b7cf8] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 742.191664] env[61923]: ERROR nova.compute.manager [instance: 62d8a170-e7ad-4d14-90d8-6f6ce32b7cf8] self[:] = self._gt.wait() [ 742.191664] env[61923]: ERROR nova.compute.manager [instance: 62d8a170-e7ad-4d14-90d8-6f6ce32b7cf8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 742.191664] env[61923]: ERROR nova.compute.manager [instance: 62d8a170-e7ad-4d14-90d8-6f6ce32b7cf8] return self._exit_event.wait() [ 742.191664] env[61923]: ERROR nova.compute.manager [instance: 62d8a170-e7ad-4d14-90d8-6f6ce32b7cf8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 742.192123] env[61923]: ERROR nova.compute.manager [instance: 62d8a170-e7ad-4d14-90d8-6f6ce32b7cf8] result = hub.switch() [ 742.192123] env[61923]: ERROR nova.compute.manager [instance: 62d8a170-e7ad-4d14-90d8-6f6ce32b7cf8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 742.192123] env[61923]: ERROR nova.compute.manager [instance: 62d8a170-e7ad-4d14-90d8-6f6ce32b7cf8] return self.greenlet.switch() [ 742.192123] env[61923]: ERROR nova.compute.manager [instance: 62d8a170-e7ad-4d14-90d8-6f6ce32b7cf8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 742.192123] env[61923]: ERROR nova.compute.manager [instance: 62d8a170-e7ad-4d14-90d8-6f6ce32b7cf8] result = function(*args, **kwargs) [ 742.192123] env[61923]: ERROR nova.compute.manager [instance: 62d8a170-e7ad-4d14-90d8-6f6ce32b7cf8] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 742.192123] env[61923]: ERROR nova.compute.manager [instance: 62d8a170-e7ad-4d14-90d8-6f6ce32b7cf8] return func(*args, **kwargs) [ 742.192123] env[61923]: ERROR nova.compute.manager [instance: 62d8a170-e7ad-4d14-90d8-6f6ce32b7cf8] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 742.192123] env[61923]: ERROR nova.compute.manager [instance: 62d8a170-e7ad-4d14-90d8-6f6ce32b7cf8] raise e [ 742.192123] env[61923]: ERROR nova.compute.manager [instance: 62d8a170-e7ad-4d14-90d8-6f6ce32b7cf8] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 742.192123] env[61923]: ERROR nova.compute.manager [instance: 62d8a170-e7ad-4d14-90d8-6f6ce32b7cf8] nwinfo = self.network_api.allocate_for_instance( [ 742.192123] env[61923]: ERROR nova.compute.manager [instance: 62d8a170-e7ad-4d14-90d8-6f6ce32b7cf8] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 742.192123] env[61923]: ERROR nova.compute.manager [instance: 62d8a170-e7ad-4d14-90d8-6f6ce32b7cf8] created_port_ids = self._update_ports_for_instance( [ 742.192538] env[61923]: ERROR nova.compute.manager [instance: 62d8a170-e7ad-4d14-90d8-6f6ce32b7cf8] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 742.192538] env[61923]: ERROR nova.compute.manager [instance: 62d8a170-e7ad-4d14-90d8-6f6ce32b7cf8] with excutils.save_and_reraise_exception(): [ 742.192538] env[61923]: ERROR nova.compute.manager [instance: 62d8a170-e7ad-4d14-90d8-6f6ce32b7cf8] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 742.192538] env[61923]: ERROR nova.compute.manager [instance: 62d8a170-e7ad-4d14-90d8-6f6ce32b7cf8] self.force_reraise() [ 742.192538] env[61923]: ERROR nova.compute.manager [instance: 62d8a170-e7ad-4d14-90d8-6f6ce32b7cf8] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 742.192538] env[61923]: ERROR nova.compute.manager [instance: 62d8a170-e7ad-4d14-90d8-6f6ce32b7cf8] raise self.value [ 742.192538] env[61923]: ERROR nova.compute.manager [instance: 62d8a170-e7ad-4d14-90d8-6f6ce32b7cf8] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 742.192538] env[61923]: ERROR nova.compute.manager [instance: 62d8a170-e7ad-4d14-90d8-6f6ce32b7cf8] updated_port = self._update_port( [ 742.192538] env[61923]: ERROR nova.compute.manager [instance: 62d8a170-e7ad-4d14-90d8-6f6ce32b7cf8] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 742.192538] env[61923]: ERROR nova.compute.manager [instance: 62d8a170-e7ad-4d14-90d8-6f6ce32b7cf8] _ensure_no_port_binding_failure(port) [ 742.192538] env[61923]: ERROR nova.compute.manager [instance: 62d8a170-e7ad-4d14-90d8-6f6ce32b7cf8] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 742.192538] env[61923]: ERROR nova.compute.manager [instance: 62d8a170-e7ad-4d14-90d8-6f6ce32b7cf8] raise exception.PortBindingFailed(port_id=port['id']) [ 742.192898] env[61923]: ERROR nova.compute.manager [instance: 62d8a170-e7ad-4d14-90d8-6f6ce32b7cf8] nova.exception.PortBindingFailed: Binding failed for port fbdf8b6e-0d86-4121-8aeb-b977a59f1186, please check neutron logs for more information. [ 742.192898] env[61923]: ERROR nova.compute.manager [instance: 62d8a170-e7ad-4d14-90d8-6f6ce32b7cf8] [ 742.192898] env[61923]: INFO nova.compute.manager [None req-0708225f-80bb-41ec-b3ed-7dd45d589b95 tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] [instance: 62d8a170-e7ad-4d14-90d8-6f6ce32b7cf8] Terminating instance [ 742.193855] env[61923]: DEBUG oslo_concurrency.lockutils [None req-0708225f-80bb-41ec-b3ed-7dd45d589b95 tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Acquiring lock "refresh_cache-62d8a170-e7ad-4d14-90d8-6f6ce32b7cf8" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 742.369436] env[61923]: DEBUG oslo_concurrency.lockutils [None req-347b83bf-b4ed-4904-9bd3-d6021183ae1c tempest-ServerShowV254Test-921228709 tempest-ServerShowV254Test-921228709-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 742.499670] env[61923]: DEBUG oslo_concurrency.lockutils [None req-ab7b3da0-3a01-44d4-9dfd-3e5d2ad27432 tempest-TenantUsagesTestJSON-871253541 tempest-TenantUsagesTestJSON-871253541-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.404s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 742.500224] env[61923]: DEBUG nova.compute.manager [None req-ab7b3da0-3a01-44d4-9dfd-3e5d2ad27432 tempest-TenantUsagesTestJSON-871253541 tempest-TenantUsagesTestJSON-871253541-project-member] [instance: b6712878-3270-43f4-9518-d41eb77124b7] Start building networks asynchronously for instance. {{(pid=61923) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 742.502970] env[61923]: DEBUG oslo_concurrency.lockutils [None req-5138192d-6349-432b-b7de-64ec5d391f8d tempest-ServerActionsTestOtherB-987045973 tempest-ServerActionsTestOtherB-987045973-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.368s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 742.504531] env[61923]: INFO nova.compute.claims [None req-5138192d-6349-432b-b7de-64ec5d391f8d tempest-ServerActionsTestOtherB-987045973 tempest-ServerActionsTestOtherB-987045973-project-member] [instance: b0bc08c1-38d5-45ee-b521-1cffda41c77c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 742.546588] env[61923]: DEBUG nova.network.neutron [req-c054fa0f-e84e-49a3-8e3e-ff33152cec08 req-20c76a27-cda2-4c0f-8c83-cb9027e6acda service nova] [instance: 62d8a170-e7ad-4d14-90d8-6f6ce32b7cf8] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 742.648492] env[61923]: DEBUG nova.network.neutron [req-c054fa0f-e84e-49a3-8e3e-ff33152cec08 req-20c76a27-cda2-4c0f-8c83-cb9027e6acda service nova] [instance: 62d8a170-e7ad-4d14-90d8-6f6ce32b7cf8] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 743.012082] env[61923]: DEBUG nova.compute.utils [None req-ab7b3da0-3a01-44d4-9dfd-3e5d2ad27432 tempest-TenantUsagesTestJSON-871253541 tempest-TenantUsagesTestJSON-871253541-project-member] Using /dev/sd instead of None {{(pid=61923) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 743.013443] env[61923]: DEBUG nova.compute.manager [None req-ab7b3da0-3a01-44d4-9dfd-3e5d2ad27432 tempest-TenantUsagesTestJSON-871253541 tempest-TenantUsagesTestJSON-871253541-project-member] [instance: b6712878-3270-43f4-9518-d41eb77124b7] Allocating IP information in the background. {{(pid=61923) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 743.013609] env[61923]: DEBUG nova.network.neutron [None req-ab7b3da0-3a01-44d4-9dfd-3e5d2ad27432 tempest-TenantUsagesTestJSON-871253541 tempest-TenantUsagesTestJSON-871253541-project-member] [instance: b6712878-3270-43f4-9518-d41eb77124b7] allocate_for_instance() {{(pid=61923) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 743.151454] env[61923]: DEBUG oslo_concurrency.lockutils [req-c054fa0f-e84e-49a3-8e3e-ff33152cec08 req-20c76a27-cda2-4c0f-8c83-cb9027e6acda service nova] Releasing lock "refresh_cache-62d8a170-e7ad-4d14-90d8-6f6ce32b7cf8" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 743.151954] env[61923]: DEBUG oslo_concurrency.lockutils [None req-0708225f-80bb-41ec-b3ed-7dd45d589b95 tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Acquired lock "refresh_cache-62d8a170-e7ad-4d14-90d8-6f6ce32b7cf8" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 743.152180] env[61923]: DEBUG nova.network.neutron [None req-0708225f-80bb-41ec-b3ed-7dd45d589b95 tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] [instance: 62d8a170-e7ad-4d14-90d8-6f6ce32b7cf8] Building network info cache for instance {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 743.231128] env[61923]: DEBUG nova.policy [None req-ab7b3da0-3a01-44d4-9dfd-3e5d2ad27432 tempest-TenantUsagesTestJSON-871253541 tempest-TenantUsagesTestJSON-871253541-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ad805225a3c64d9d92492ea1cf19860c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a489eaeb7c2d4c1b8dd0916edd2160d2', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61923) authorize /opt/stack/nova/nova/policy.py:201}} [ 743.519270] env[61923]: DEBUG nova.compute.manager [None req-ab7b3da0-3a01-44d4-9dfd-3e5d2ad27432 tempest-TenantUsagesTestJSON-871253541 tempest-TenantUsagesTestJSON-871253541-project-member] [instance: b6712878-3270-43f4-9518-d41eb77124b7] Start building block device mappings for instance. {{(pid=61923) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 743.678400] env[61923]: DEBUG nova.network.neutron [None req-ab7b3da0-3a01-44d4-9dfd-3e5d2ad27432 tempest-TenantUsagesTestJSON-871253541 tempest-TenantUsagesTestJSON-871253541-project-member] [instance: b6712878-3270-43f4-9518-d41eb77124b7] Successfully created port: a4338c60-f3e2-4844-b170-c852adcfcb25 {{(pid=61923) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 743.681291] env[61923]: DEBUG nova.network.neutron [None req-0708225f-80bb-41ec-b3ed-7dd45d589b95 tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] [instance: 62d8a170-e7ad-4d14-90d8-6f6ce32b7cf8] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 743.865741] env[61923]: DEBUG nova.network.neutron [None req-0708225f-80bb-41ec-b3ed-7dd45d589b95 tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] [instance: 62d8a170-e7ad-4d14-90d8-6f6ce32b7cf8] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 743.875071] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46f6be70-d0bf-4650-ab6a-57f147764eae {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.886546] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d3c40ee-6eb9-4a30-875a-f51b8a30aa6c {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.941421] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb272545-1d02-42a3-b2fb-f3d3c8939939 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.952596] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28422848-acaa-4b90-810b-49efc125e187 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.975348] env[61923]: DEBUG nova.compute.provider_tree [None req-5138192d-6349-432b-b7de-64ec5d391f8d tempest-ServerActionsTestOtherB-987045973 tempest-ServerActionsTestOtherB-987045973-project-member] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 744.074447] env[61923]: DEBUG nova.compute.manager [req-5fa287ba-58be-42a6-a097-287014e489d9 req-e1f05f39-dea9-41f5-9f6d-bba410b2d90c service nova] [instance: 62d8a170-e7ad-4d14-90d8-6f6ce32b7cf8] Received event network-vif-deleted-fbdf8b6e-0d86-4121-8aeb-b977a59f1186 {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 744.369427] env[61923]: DEBUG oslo_concurrency.lockutils [None req-0708225f-80bb-41ec-b3ed-7dd45d589b95 tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Releasing lock "refresh_cache-62d8a170-e7ad-4d14-90d8-6f6ce32b7cf8" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 744.369849] env[61923]: DEBUG nova.compute.manager [None req-0708225f-80bb-41ec-b3ed-7dd45d589b95 tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] [instance: 62d8a170-e7ad-4d14-90d8-6f6ce32b7cf8] Start destroying the instance on the hypervisor. {{(pid=61923) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 744.370093] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-0708225f-80bb-41ec-b3ed-7dd45d589b95 tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] [instance: 62d8a170-e7ad-4d14-90d8-6f6ce32b7cf8] Destroying instance {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 744.370350] env[61923]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-71d649b7-35c0-4e06-9d59-8f28e4adb517 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.380692] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6fbfacf7-e8d4-4fe7-8b33-f458ed050bfd {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.406301] env[61923]: WARNING nova.virt.vmwareapi.vmops [None req-0708225f-80bb-41ec-b3ed-7dd45d589b95 tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] [instance: 62d8a170-e7ad-4d14-90d8-6f6ce32b7cf8] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 62d8a170-e7ad-4d14-90d8-6f6ce32b7cf8 could not be found. [ 744.406560] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-0708225f-80bb-41ec-b3ed-7dd45d589b95 tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] [instance: 62d8a170-e7ad-4d14-90d8-6f6ce32b7cf8] Instance destroyed {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 744.406744] env[61923]: INFO nova.compute.manager [None req-0708225f-80bb-41ec-b3ed-7dd45d589b95 tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] [instance: 62d8a170-e7ad-4d14-90d8-6f6ce32b7cf8] Took 0.04 seconds to destroy the instance on the hypervisor. [ 744.407448] env[61923]: DEBUG oslo.service.loopingcall [None req-0708225f-80bb-41ec-b3ed-7dd45d589b95 tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61923) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 744.407448] env[61923]: DEBUG nova.compute.manager [-] [instance: 62d8a170-e7ad-4d14-90d8-6f6ce32b7cf8] Deallocating network for instance {{(pid=61923) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 744.407448] env[61923]: DEBUG nova.network.neutron [-] [instance: 62d8a170-e7ad-4d14-90d8-6f6ce32b7cf8] deallocate_for_instance() {{(pid=61923) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 744.438064] env[61923]: DEBUG nova.network.neutron [-] [instance: 62d8a170-e7ad-4d14-90d8-6f6ce32b7cf8] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 744.479117] env[61923]: DEBUG nova.scheduler.client.report [None req-5138192d-6349-432b-b7de-64ec5d391f8d tempest-ServerActionsTestOtherB-987045973 tempest-ServerActionsTestOtherB-987045973-project-member] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 744.530766] env[61923]: DEBUG nova.compute.manager [None req-ab7b3da0-3a01-44d4-9dfd-3e5d2ad27432 tempest-TenantUsagesTestJSON-871253541 tempest-TenantUsagesTestJSON-871253541-project-member] [instance: b6712878-3270-43f4-9518-d41eb77124b7] Start spawning the instance on the hypervisor. {{(pid=61923) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 744.558136] env[61923]: DEBUG nova.virt.hardware [None req-ab7b3da0-3a01-44d4-9dfd-3e5d2ad27432 tempest-TenantUsagesTestJSON-871253541 tempest-TenantUsagesTestJSON-871253541-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-29T20:07:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-29T20:07:35Z,direct_url=,disk_format='vmdk',id=0f153f63-ae0a-45b1-b7f5-7f9b673c947f,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='4e7b5e3b3c3443c8bd5c70f8a15739f5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-29T20:07:36Z,virtual_size=,visibility=), allow threads: False {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 744.558136] env[61923]: DEBUG nova.virt.hardware [None req-ab7b3da0-3a01-44d4-9dfd-3e5d2ad27432 tempest-TenantUsagesTestJSON-871253541 tempest-TenantUsagesTestJSON-871253541-project-member] Flavor limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 744.558136] env[61923]: DEBUG nova.virt.hardware [None req-ab7b3da0-3a01-44d4-9dfd-3e5d2ad27432 tempest-TenantUsagesTestJSON-871253541 tempest-TenantUsagesTestJSON-871253541-project-member] Image limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 744.558136] env[61923]: DEBUG nova.virt.hardware [None req-ab7b3da0-3a01-44d4-9dfd-3e5d2ad27432 tempest-TenantUsagesTestJSON-871253541 tempest-TenantUsagesTestJSON-871253541-project-member] Flavor pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 744.558345] env[61923]: DEBUG nova.virt.hardware [None req-ab7b3da0-3a01-44d4-9dfd-3e5d2ad27432 tempest-TenantUsagesTestJSON-871253541 tempest-TenantUsagesTestJSON-871253541-project-member] Image pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 744.558345] env[61923]: DEBUG nova.virt.hardware [None req-ab7b3da0-3a01-44d4-9dfd-3e5d2ad27432 tempest-TenantUsagesTestJSON-871253541 tempest-TenantUsagesTestJSON-871253541-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 744.558696] env[61923]: DEBUG nova.virt.hardware [None req-ab7b3da0-3a01-44d4-9dfd-3e5d2ad27432 tempest-TenantUsagesTestJSON-871253541 tempest-TenantUsagesTestJSON-871253541-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 744.558991] env[61923]: DEBUG nova.virt.hardware [None req-ab7b3da0-3a01-44d4-9dfd-3e5d2ad27432 tempest-TenantUsagesTestJSON-871253541 tempest-TenantUsagesTestJSON-871253541-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 744.559324] env[61923]: DEBUG nova.virt.hardware [None req-ab7b3da0-3a01-44d4-9dfd-3e5d2ad27432 tempest-TenantUsagesTestJSON-871253541 tempest-TenantUsagesTestJSON-871253541-project-member] Got 1 possible topologies {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 744.559625] env[61923]: DEBUG nova.virt.hardware [None req-ab7b3da0-3a01-44d4-9dfd-3e5d2ad27432 tempest-TenantUsagesTestJSON-871253541 tempest-TenantUsagesTestJSON-871253541-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 744.559957] env[61923]: DEBUG nova.virt.hardware [None req-ab7b3da0-3a01-44d4-9dfd-3e5d2ad27432 tempest-TenantUsagesTestJSON-871253541 tempest-TenantUsagesTestJSON-871253541-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 744.560893] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ae8be2f-4a37-4691-afba-a21eb06b7fef {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.570469] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d73ee2ec-ab65-49b4-ac4d-0bf91f350ddd {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.851279] env[61923]: ERROR nova.compute.manager [None req-ab7b3da0-3a01-44d4-9dfd-3e5d2ad27432 tempest-TenantUsagesTestJSON-871253541 tempest-TenantUsagesTestJSON-871253541-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port a4338c60-f3e2-4844-b170-c852adcfcb25, please check neutron logs for more information. [ 744.851279] env[61923]: ERROR nova.compute.manager Traceback (most recent call last): [ 744.851279] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 744.851279] env[61923]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 744.851279] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 744.851279] env[61923]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 744.851279] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 744.851279] env[61923]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 744.851279] env[61923]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 744.851279] env[61923]: ERROR nova.compute.manager self.force_reraise() [ 744.851279] env[61923]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 744.851279] env[61923]: ERROR nova.compute.manager raise self.value [ 744.851279] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 744.851279] env[61923]: ERROR nova.compute.manager updated_port = self._update_port( [ 744.851279] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 744.851279] env[61923]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 744.852161] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 744.852161] env[61923]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 744.852161] env[61923]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port a4338c60-f3e2-4844-b170-c852adcfcb25, please check neutron logs for more information. [ 744.852161] env[61923]: ERROR nova.compute.manager [ 744.852161] env[61923]: Traceback (most recent call last): [ 744.852161] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 744.852161] env[61923]: listener.cb(fileno) [ 744.852161] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 744.852161] env[61923]: result = function(*args, **kwargs) [ 744.852161] env[61923]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 744.852161] env[61923]: return func(*args, **kwargs) [ 744.852161] env[61923]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 744.852161] env[61923]: raise e [ 744.852161] env[61923]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 744.852161] env[61923]: nwinfo = self.network_api.allocate_for_instance( [ 744.852161] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 744.852161] env[61923]: created_port_ids = self._update_ports_for_instance( [ 744.852161] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 744.852161] env[61923]: with excutils.save_and_reraise_exception(): [ 744.852161] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 744.852161] env[61923]: self.force_reraise() [ 744.852161] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 744.852161] env[61923]: raise self.value [ 744.852161] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 744.852161] env[61923]: updated_port = self._update_port( [ 744.852161] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 744.852161] env[61923]: _ensure_no_port_binding_failure(port) [ 744.852161] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 744.852161] env[61923]: raise exception.PortBindingFailed(port_id=port['id']) [ 744.853433] env[61923]: nova.exception.PortBindingFailed: Binding failed for port a4338c60-f3e2-4844-b170-c852adcfcb25, please check neutron logs for more information. [ 744.853433] env[61923]: Removing descriptor: 14 [ 744.853433] env[61923]: ERROR nova.compute.manager [None req-ab7b3da0-3a01-44d4-9dfd-3e5d2ad27432 tempest-TenantUsagesTestJSON-871253541 tempest-TenantUsagesTestJSON-871253541-project-member] [instance: b6712878-3270-43f4-9518-d41eb77124b7] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port a4338c60-f3e2-4844-b170-c852adcfcb25, please check neutron logs for more information. [ 744.853433] env[61923]: ERROR nova.compute.manager [instance: b6712878-3270-43f4-9518-d41eb77124b7] Traceback (most recent call last): [ 744.853433] env[61923]: ERROR nova.compute.manager [instance: b6712878-3270-43f4-9518-d41eb77124b7] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 744.853433] env[61923]: ERROR nova.compute.manager [instance: b6712878-3270-43f4-9518-d41eb77124b7] yield resources [ 744.853433] env[61923]: ERROR nova.compute.manager [instance: b6712878-3270-43f4-9518-d41eb77124b7] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 744.853433] env[61923]: ERROR nova.compute.manager [instance: b6712878-3270-43f4-9518-d41eb77124b7] self.driver.spawn(context, instance, image_meta, [ 744.853433] env[61923]: ERROR nova.compute.manager [instance: b6712878-3270-43f4-9518-d41eb77124b7] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 744.853433] env[61923]: ERROR nova.compute.manager [instance: b6712878-3270-43f4-9518-d41eb77124b7] self._vmops.spawn(context, instance, image_meta, injected_files, [ 744.853433] env[61923]: ERROR nova.compute.manager [instance: b6712878-3270-43f4-9518-d41eb77124b7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 744.853433] env[61923]: ERROR nova.compute.manager [instance: b6712878-3270-43f4-9518-d41eb77124b7] vm_ref = self.build_virtual_machine(instance, [ 744.854283] env[61923]: ERROR nova.compute.manager [instance: b6712878-3270-43f4-9518-d41eb77124b7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 744.854283] env[61923]: ERROR nova.compute.manager [instance: b6712878-3270-43f4-9518-d41eb77124b7] vif_infos = vmwarevif.get_vif_info(self._session, [ 744.854283] env[61923]: ERROR nova.compute.manager [instance: b6712878-3270-43f4-9518-d41eb77124b7] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 744.854283] env[61923]: ERROR nova.compute.manager [instance: b6712878-3270-43f4-9518-d41eb77124b7] for vif in network_info: [ 744.854283] env[61923]: ERROR nova.compute.manager [instance: b6712878-3270-43f4-9518-d41eb77124b7] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 744.854283] env[61923]: ERROR nova.compute.manager [instance: b6712878-3270-43f4-9518-d41eb77124b7] return self._sync_wrapper(fn, *args, **kwargs) [ 744.854283] env[61923]: ERROR nova.compute.manager [instance: b6712878-3270-43f4-9518-d41eb77124b7] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 744.854283] env[61923]: ERROR nova.compute.manager [instance: b6712878-3270-43f4-9518-d41eb77124b7] self.wait() [ 744.854283] env[61923]: ERROR nova.compute.manager [instance: b6712878-3270-43f4-9518-d41eb77124b7] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 744.854283] env[61923]: ERROR nova.compute.manager [instance: b6712878-3270-43f4-9518-d41eb77124b7] self[:] = self._gt.wait() [ 744.854283] env[61923]: ERROR nova.compute.manager [instance: b6712878-3270-43f4-9518-d41eb77124b7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 744.854283] env[61923]: ERROR nova.compute.manager [instance: b6712878-3270-43f4-9518-d41eb77124b7] return self._exit_event.wait() [ 744.854283] env[61923]: ERROR nova.compute.manager [instance: b6712878-3270-43f4-9518-d41eb77124b7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 744.854928] env[61923]: ERROR nova.compute.manager [instance: b6712878-3270-43f4-9518-d41eb77124b7] result = hub.switch() [ 744.854928] env[61923]: ERROR nova.compute.manager [instance: b6712878-3270-43f4-9518-d41eb77124b7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 744.854928] env[61923]: ERROR nova.compute.manager [instance: b6712878-3270-43f4-9518-d41eb77124b7] return self.greenlet.switch() [ 744.854928] env[61923]: ERROR nova.compute.manager [instance: b6712878-3270-43f4-9518-d41eb77124b7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 744.854928] env[61923]: ERROR nova.compute.manager [instance: b6712878-3270-43f4-9518-d41eb77124b7] result = function(*args, **kwargs) [ 744.854928] env[61923]: ERROR nova.compute.manager [instance: b6712878-3270-43f4-9518-d41eb77124b7] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 744.854928] env[61923]: ERROR nova.compute.manager [instance: b6712878-3270-43f4-9518-d41eb77124b7] return func(*args, **kwargs) [ 744.854928] env[61923]: ERROR nova.compute.manager [instance: b6712878-3270-43f4-9518-d41eb77124b7] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 744.854928] env[61923]: ERROR nova.compute.manager [instance: b6712878-3270-43f4-9518-d41eb77124b7] raise e [ 744.854928] env[61923]: ERROR nova.compute.manager [instance: b6712878-3270-43f4-9518-d41eb77124b7] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 744.854928] env[61923]: ERROR nova.compute.manager [instance: b6712878-3270-43f4-9518-d41eb77124b7] nwinfo = self.network_api.allocate_for_instance( [ 744.854928] env[61923]: ERROR nova.compute.manager [instance: b6712878-3270-43f4-9518-d41eb77124b7] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 744.854928] env[61923]: ERROR nova.compute.manager [instance: b6712878-3270-43f4-9518-d41eb77124b7] created_port_ids = self._update_ports_for_instance( [ 744.855559] env[61923]: ERROR nova.compute.manager [instance: b6712878-3270-43f4-9518-d41eb77124b7] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 744.855559] env[61923]: ERROR nova.compute.manager [instance: b6712878-3270-43f4-9518-d41eb77124b7] with excutils.save_and_reraise_exception(): [ 744.855559] env[61923]: ERROR nova.compute.manager [instance: b6712878-3270-43f4-9518-d41eb77124b7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 744.855559] env[61923]: ERROR nova.compute.manager [instance: b6712878-3270-43f4-9518-d41eb77124b7] self.force_reraise() [ 744.855559] env[61923]: ERROR nova.compute.manager [instance: b6712878-3270-43f4-9518-d41eb77124b7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 744.855559] env[61923]: ERROR nova.compute.manager [instance: b6712878-3270-43f4-9518-d41eb77124b7] raise self.value [ 744.855559] env[61923]: ERROR nova.compute.manager [instance: b6712878-3270-43f4-9518-d41eb77124b7] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 744.855559] env[61923]: ERROR nova.compute.manager [instance: b6712878-3270-43f4-9518-d41eb77124b7] updated_port = self._update_port( [ 744.855559] env[61923]: ERROR nova.compute.manager [instance: b6712878-3270-43f4-9518-d41eb77124b7] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 744.855559] env[61923]: ERROR nova.compute.manager [instance: b6712878-3270-43f4-9518-d41eb77124b7] _ensure_no_port_binding_failure(port) [ 744.855559] env[61923]: ERROR nova.compute.manager [instance: b6712878-3270-43f4-9518-d41eb77124b7] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 744.855559] env[61923]: ERROR nova.compute.manager [instance: b6712878-3270-43f4-9518-d41eb77124b7] raise exception.PortBindingFailed(port_id=port['id']) [ 744.856148] env[61923]: ERROR nova.compute.manager [instance: b6712878-3270-43f4-9518-d41eb77124b7] nova.exception.PortBindingFailed: Binding failed for port a4338c60-f3e2-4844-b170-c852adcfcb25, please check neutron logs for more information. [ 744.856148] env[61923]: ERROR nova.compute.manager [instance: b6712878-3270-43f4-9518-d41eb77124b7] [ 744.856148] env[61923]: INFO nova.compute.manager [None req-ab7b3da0-3a01-44d4-9dfd-3e5d2ad27432 tempest-TenantUsagesTestJSON-871253541 tempest-TenantUsagesTestJSON-871253541-project-member] [instance: b6712878-3270-43f4-9518-d41eb77124b7] Terminating instance [ 744.856148] env[61923]: DEBUG oslo_concurrency.lockutils [None req-ab7b3da0-3a01-44d4-9dfd-3e5d2ad27432 tempest-TenantUsagesTestJSON-871253541 tempest-TenantUsagesTestJSON-871253541-project-member] Acquiring lock "refresh_cache-b6712878-3270-43f4-9518-d41eb77124b7" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 744.856148] env[61923]: DEBUG oslo_concurrency.lockutils [None req-ab7b3da0-3a01-44d4-9dfd-3e5d2ad27432 tempest-TenantUsagesTestJSON-871253541 tempest-TenantUsagesTestJSON-871253541-project-member] Acquired lock "refresh_cache-b6712878-3270-43f4-9518-d41eb77124b7" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 744.856148] env[61923]: DEBUG nova.network.neutron [None req-ab7b3da0-3a01-44d4-9dfd-3e5d2ad27432 tempest-TenantUsagesTestJSON-871253541 tempest-TenantUsagesTestJSON-871253541-project-member] [instance: b6712878-3270-43f4-9518-d41eb77124b7] Building network info cache for instance {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 744.941093] env[61923]: DEBUG nova.network.neutron [-] [instance: 62d8a170-e7ad-4d14-90d8-6f6ce32b7cf8] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 744.985625] env[61923]: DEBUG oslo_concurrency.lockutils [None req-5138192d-6349-432b-b7de-64ec5d391f8d tempest-ServerActionsTestOtherB-987045973 tempest-ServerActionsTestOtherB-987045973-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.483s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 744.986173] env[61923]: DEBUG nova.compute.manager [None req-5138192d-6349-432b-b7de-64ec5d391f8d tempest-ServerActionsTestOtherB-987045973 tempest-ServerActionsTestOtherB-987045973-project-member] [instance: b0bc08c1-38d5-45ee-b521-1cffda41c77c] Start building networks asynchronously for instance. {{(pid=61923) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 744.988749] env[61923]: DEBUG oslo_concurrency.lockutils [None req-40b011d4-39ab-4684-a124-b3ba87d489df tempest-ServerPasswordTestJSON-595643156 tempest-ServerPasswordTestJSON-595643156-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.428s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 745.388038] env[61923]: DEBUG nova.network.neutron [None req-ab7b3da0-3a01-44d4-9dfd-3e5d2ad27432 tempest-TenantUsagesTestJSON-871253541 tempest-TenantUsagesTestJSON-871253541-project-member] [instance: b6712878-3270-43f4-9518-d41eb77124b7] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 745.444287] env[61923]: INFO nova.compute.manager [-] [instance: 62d8a170-e7ad-4d14-90d8-6f6ce32b7cf8] Took 1.04 seconds to deallocate network for instance. [ 745.446951] env[61923]: DEBUG nova.compute.claims [None req-0708225f-80bb-41ec-b3ed-7dd45d589b95 tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] [instance: 62d8a170-e7ad-4d14-90d8-6f6ce32b7cf8] Aborting claim: {{(pid=61923) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 745.447140] env[61923]: DEBUG oslo_concurrency.lockutils [None req-0708225f-80bb-41ec-b3ed-7dd45d589b95 tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 745.448778] env[61923]: DEBUG nova.network.neutron [None req-ab7b3da0-3a01-44d4-9dfd-3e5d2ad27432 tempest-TenantUsagesTestJSON-871253541 tempest-TenantUsagesTestJSON-871253541-project-member] [instance: b6712878-3270-43f4-9518-d41eb77124b7] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 745.492608] env[61923]: DEBUG nova.compute.utils [None req-5138192d-6349-432b-b7de-64ec5d391f8d tempest-ServerActionsTestOtherB-987045973 tempest-ServerActionsTestOtherB-987045973-project-member] Using /dev/sd instead of None {{(pid=61923) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 745.496710] env[61923]: DEBUG nova.compute.manager [None req-5138192d-6349-432b-b7de-64ec5d391f8d tempest-ServerActionsTestOtherB-987045973 tempest-ServerActionsTestOtherB-987045973-project-member] [instance: b0bc08c1-38d5-45ee-b521-1cffda41c77c] Allocating IP information in the background. {{(pid=61923) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 745.496872] env[61923]: DEBUG nova.network.neutron [None req-5138192d-6349-432b-b7de-64ec5d391f8d tempest-ServerActionsTestOtherB-987045973 tempest-ServerActionsTestOtherB-987045973-project-member] [instance: b0bc08c1-38d5-45ee-b521-1cffda41c77c] allocate_for_instance() {{(pid=61923) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 745.559418] env[61923]: DEBUG nova.policy [None req-5138192d-6349-432b-b7de-64ec5d391f8d tempest-ServerActionsTestOtherB-987045973 tempest-ServerActionsTestOtherB-987045973-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '36d57836b49746c39b0c64aea6c9c9cb', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8edace86a5fb433e9070d38ea405d12e', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61923) authorize /opt/stack/nova/nova/policy.py:201}} [ 745.797024] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7d41608-b38b-4faa-bcf1-7ffe9ba8fd7f {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.804297] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32ff8a7c-826f-4900-86c6-55450b7d9ceb {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.838571] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2720ca25-b524-4064-b1f9-042e96708ce1 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.846315] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dccc5ec1-53ad-4bc9-b3b4-3df080c9cfae {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.860227] env[61923]: DEBUG nova.compute.provider_tree [None req-40b011d4-39ab-4684-a124-b3ba87d489df tempest-ServerPasswordTestJSON-595643156 tempest-ServerPasswordTestJSON-595643156-project-member] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 745.905880] env[61923]: DEBUG nova.network.neutron [None req-5138192d-6349-432b-b7de-64ec5d391f8d tempest-ServerActionsTestOtherB-987045973 tempest-ServerActionsTestOtherB-987045973-project-member] [instance: b0bc08c1-38d5-45ee-b521-1cffda41c77c] Successfully created port: 129cb36c-cc2b-4205-9d7b-0b4d553b568b {{(pid=61923) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 745.951015] env[61923]: DEBUG oslo_concurrency.lockutils [None req-ab7b3da0-3a01-44d4-9dfd-3e5d2ad27432 tempest-TenantUsagesTestJSON-871253541 tempest-TenantUsagesTestJSON-871253541-project-member] Releasing lock "refresh_cache-b6712878-3270-43f4-9518-d41eb77124b7" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 745.951464] env[61923]: DEBUG nova.compute.manager [None req-ab7b3da0-3a01-44d4-9dfd-3e5d2ad27432 tempest-TenantUsagesTestJSON-871253541 tempest-TenantUsagesTestJSON-871253541-project-member] [instance: b6712878-3270-43f4-9518-d41eb77124b7] Start destroying the instance on the hypervisor. {{(pid=61923) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 745.951653] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-ab7b3da0-3a01-44d4-9dfd-3e5d2ad27432 tempest-TenantUsagesTestJSON-871253541 tempest-TenantUsagesTestJSON-871253541-project-member] [instance: b6712878-3270-43f4-9518-d41eb77124b7] Destroying instance {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 745.951950] env[61923]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-641381dc-71b0-4b6f-b7cf-39c06923d4a5 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.961203] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b7b292f-6377-4ef7-9a56-ad9ca9cb0533 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.982727] env[61923]: WARNING nova.virt.vmwareapi.vmops [None req-ab7b3da0-3a01-44d4-9dfd-3e5d2ad27432 tempest-TenantUsagesTestJSON-871253541 tempest-TenantUsagesTestJSON-871253541-project-member] [instance: b6712878-3270-43f4-9518-d41eb77124b7] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance b6712878-3270-43f4-9518-d41eb77124b7 could not be found. [ 745.982956] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-ab7b3da0-3a01-44d4-9dfd-3e5d2ad27432 tempest-TenantUsagesTestJSON-871253541 tempest-TenantUsagesTestJSON-871253541-project-member] [instance: b6712878-3270-43f4-9518-d41eb77124b7] Instance destroyed {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 745.983195] env[61923]: INFO nova.compute.manager [None req-ab7b3da0-3a01-44d4-9dfd-3e5d2ad27432 tempest-TenantUsagesTestJSON-871253541 tempest-TenantUsagesTestJSON-871253541-project-member] [instance: b6712878-3270-43f4-9518-d41eb77124b7] Took 0.03 seconds to destroy the instance on the hypervisor. [ 745.983461] env[61923]: DEBUG oslo.service.loopingcall [None req-ab7b3da0-3a01-44d4-9dfd-3e5d2ad27432 tempest-TenantUsagesTestJSON-871253541 tempest-TenantUsagesTestJSON-871253541-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61923) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 745.983660] env[61923]: DEBUG nova.compute.manager [-] [instance: b6712878-3270-43f4-9518-d41eb77124b7] Deallocating network for instance {{(pid=61923) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 745.983755] env[61923]: DEBUG nova.network.neutron [-] [instance: b6712878-3270-43f4-9518-d41eb77124b7] deallocate_for_instance() {{(pid=61923) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 745.996993] env[61923]: DEBUG nova.compute.manager [None req-5138192d-6349-432b-b7de-64ec5d391f8d tempest-ServerActionsTestOtherB-987045973 tempest-ServerActionsTestOtherB-987045973-project-member] [instance: b0bc08c1-38d5-45ee-b521-1cffda41c77c] Start building block device mappings for instance. {{(pid=61923) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 746.015531] env[61923]: DEBUG nova.network.neutron [-] [instance: b6712878-3270-43f4-9518-d41eb77124b7] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 746.100070] env[61923]: DEBUG nova.compute.manager [req-efee9203-151c-4b0e-890b-4aebe517d7b4 req-e9686878-2ed8-4cca-975b-78e24d7a772d service nova] [instance: b6712878-3270-43f4-9518-d41eb77124b7] Received event network-changed-a4338c60-f3e2-4844-b170-c852adcfcb25 {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 746.100335] env[61923]: DEBUG nova.compute.manager [req-efee9203-151c-4b0e-890b-4aebe517d7b4 req-e9686878-2ed8-4cca-975b-78e24d7a772d service nova] [instance: b6712878-3270-43f4-9518-d41eb77124b7] Refreshing instance network info cache due to event network-changed-a4338c60-f3e2-4844-b170-c852adcfcb25. {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 746.100478] env[61923]: DEBUG oslo_concurrency.lockutils [req-efee9203-151c-4b0e-890b-4aebe517d7b4 req-e9686878-2ed8-4cca-975b-78e24d7a772d service nova] Acquiring lock "refresh_cache-b6712878-3270-43f4-9518-d41eb77124b7" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 746.100594] env[61923]: DEBUG oslo_concurrency.lockutils [req-efee9203-151c-4b0e-890b-4aebe517d7b4 req-e9686878-2ed8-4cca-975b-78e24d7a772d service nova] Acquired lock "refresh_cache-b6712878-3270-43f4-9518-d41eb77124b7" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 746.100741] env[61923]: DEBUG nova.network.neutron [req-efee9203-151c-4b0e-890b-4aebe517d7b4 req-e9686878-2ed8-4cca-975b-78e24d7a772d service nova] [instance: b6712878-3270-43f4-9518-d41eb77124b7] Refreshing network info cache for port a4338c60-f3e2-4844-b170-c852adcfcb25 {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 746.365565] env[61923]: DEBUG nova.scheduler.client.report [None req-40b011d4-39ab-4684-a124-b3ba87d489df tempest-ServerPasswordTestJSON-595643156 tempest-ServerPasswordTestJSON-595643156-project-member] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 746.521392] env[61923]: DEBUG nova.network.neutron [-] [instance: b6712878-3270-43f4-9518-d41eb77124b7] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 746.625701] env[61923]: DEBUG nova.network.neutron [req-efee9203-151c-4b0e-890b-4aebe517d7b4 req-e9686878-2ed8-4cca-975b-78e24d7a772d service nova] [instance: b6712878-3270-43f4-9518-d41eb77124b7] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 746.705514] env[61923]: DEBUG nova.network.neutron [req-efee9203-151c-4b0e-890b-4aebe517d7b4 req-e9686878-2ed8-4cca-975b-78e24d7a772d service nova] [instance: b6712878-3270-43f4-9518-d41eb77124b7] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 746.875783] env[61923]: DEBUG oslo_concurrency.lockutils [None req-40b011d4-39ab-4684-a124-b3ba87d489df tempest-ServerPasswordTestJSON-595643156 tempest-ServerPasswordTestJSON-595643156-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.887s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 746.876460] env[61923]: ERROR nova.compute.manager [None req-40b011d4-39ab-4684-a124-b3ba87d489df tempest-ServerPasswordTestJSON-595643156 tempest-ServerPasswordTestJSON-595643156-project-member] [instance: 7c0341f9-7401-42ee-9512-afd832c76940] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port e84bf904-db99-4abc-90aa-4be30a8f32e8, please check neutron logs for more information. [ 746.876460] env[61923]: ERROR nova.compute.manager [instance: 7c0341f9-7401-42ee-9512-afd832c76940] Traceback (most recent call last): [ 746.876460] env[61923]: ERROR nova.compute.manager [instance: 7c0341f9-7401-42ee-9512-afd832c76940] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 746.876460] env[61923]: ERROR nova.compute.manager [instance: 7c0341f9-7401-42ee-9512-afd832c76940] self.driver.spawn(context, instance, image_meta, [ 746.876460] env[61923]: ERROR nova.compute.manager [instance: 7c0341f9-7401-42ee-9512-afd832c76940] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 746.876460] env[61923]: ERROR nova.compute.manager [instance: 7c0341f9-7401-42ee-9512-afd832c76940] self._vmops.spawn(context, instance, image_meta, injected_files, [ 746.876460] env[61923]: ERROR nova.compute.manager [instance: 7c0341f9-7401-42ee-9512-afd832c76940] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 746.876460] env[61923]: ERROR nova.compute.manager [instance: 7c0341f9-7401-42ee-9512-afd832c76940] vm_ref = self.build_virtual_machine(instance, [ 746.876460] env[61923]: ERROR nova.compute.manager [instance: 7c0341f9-7401-42ee-9512-afd832c76940] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 746.876460] env[61923]: ERROR nova.compute.manager [instance: 7c0341f9-7401-42ee-9512-afd832c76940] vif_infos = vmwarevif.get_vif_info(self._session, [ 746.876460] env[61923]: ERROR nova.compute.manager [instance: 7c0341f9-7401-42ee-9512-afd832c76940] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 746.876836] env[61923]: ERROR nova.compute.manager [instance: 7c0341f9-7401-42ee-9512-afd832c76940] for vif in network_info: [ 746.876836] env[61923]: ERROR nova.compute.manager [instance: 7c0341f9-7401-42ee-9512-afd832c76940] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 746.876836] env[61923]: ERROR nova.compute.manager [instance: 7c0341f9-7401-42ee-9512-afd832c76940] return self._sync_wrapper(fn, *args, **kwargs) [ 746.876836] env[61923]: ERROR nova.compute.manager [instance: 7c0341f9-7401-42ee-9512-afd832c76940] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 746.876836] env[61923]: ERROR nova.compute.manager [instance: 7c0341f9-7401-42ee-9512-afd832c76940] self.wait() [ 746.876836] env[61923]: ERROR nova.compute.manager [instance: 7c0341f9-7401-42ee-9512-afd832c76940] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 746.876836] env[61923]: ERROR nova.compute.manager [instance: 7c0341f9-7401-42ee-9512-afd832c76940] self[:] = self._gt.wait() [ 746.876836] env[61923]: ERROR nova.compute.manager [instance: 7c0341f9-7401-42ee-9512-afd832c76940] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 746.876836] env[61923]: ERROR nova.compute.manager [instance: 7c0341f9-7401-42ee-9512-afd832c76940] return self._exit_event.wait() [ 746.876836] env[61923]: ERROR nova.compute.manager [instance: 7c0341f9-7401-42ee-9512-afd832c76940] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 746.876836] env[61923]: ERROR nova.compute.manager [instance: 7c0341f9-7401-42ee-9512-afd832c76940] result = hub.switch() [ 746.876836] env[61923]: ERROR nova.compute.manager [instance: 7c0341f9-7401-42ee-9512-afd832c76940] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 746.876836] env[61923]: ERROR nova.compute.manager [instance: 7c0341f9-7401-42ee-9512-afd832c76940] return self.greenlet.switch() [ 746.877208] env[61923]: ERROR nova.compute.manager [instance: 7c0341f9-7401-42ee-9512-afd832c76940] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 746.877208] env[61923]: ERROR nova.compute.manager [instance: 7c0341f9-7401-42ee-9512-afd832c76940] result = function(*args, **kwargs) [ 746.877208] env[61923]: ERROR nova.compute.manager [instance: 7c0341f9-7401-42ee-9512-afd832c76940] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 746.877208] env[61923]: ERROR nova.compute.manager [instance: 7c0341f9-7401-42ee-9512-afd832c76940] return func(*args, **kwargs) [ 746.877208] env[61923]: ERROR nova.compute.manager [instance: 7c0341f9-7401-42ee-9512-afd832c76940] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 746.877208] env[61923]: ERROR nova.compute.manager [instance: 7c0341f9-7401-42ee-9512-afd832c76940] raise e [ 746.877208] env[61923]: ERROR nova.compute.manager [instance: 7c0341f9-7401-42ee-9512-afd832c76940] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 746.877208] env[61923]: ERROR nova.compute.manager [instance: 7c0341f9-7401-42ee-9512-afd832c76940] nwinfo = self.network_api.allocate_for_instance( [ 746.877208] env[61923]: ERROR nova.compute.manager [instance: 7c0341f9-7401-42ee-9512-afd832c76940] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 746.877208] env[61923]: ERROR nova.compute.manager [instance: 7c0341f9-7401-42ee-9512-afd832c76940] created_port_ids = self._update_ports_for_instance( [ 746.877208] env[61923]: ERROR nova.compute.manager [instance: 7c0341f9-7401-42ee-9512-afd832c76940] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 746.877208] env[61923]: ERROR nova.compute.manager [instance: 7c0341f9-7401-42ee-9512-afd832c76940] with excutils.save_and_reraise_exception(): [ 746.877208] env[61923]: ERROR nova.compute.manager [instance: 7c0341f9-7401-42ee-9512-afd832c76940] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 746.877600] env[61923]: ERROR nova.compute.manager [instance: 7c0341f9-7401-42ee-9512-afd832c76940] self.force_reraise() [ 746.877600] env[61923]: ERROR nova.compute.manager [instance: 7c0341f9-7401-42ee-9512-afd832c76940] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 746.877600] env[61923]: ERROR nova.compute.manager [instance: 7c0341f9-7401-42ee-9512-afd832c76940] raise self.value [ 746.877600] env[61923]: ERROR nova.compute.manager [instance: 7c0341f9-7401-42ee-9512-afd832c76940] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 746.877600] env[61923]: ERROR nova.compute.manager [instance: 7c0341f9-7401-42ee-9512-afd832c76940] updated_port = self._update_port( [ 746.877600] env[61923]: ERROR nova.compute.manager [instance: 7c0341f9-7401-42ee-9512-afd832c76940] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 746.877600] env[61923]: ERROR nova.compute.manager [instance: 7c0341f9-7401-42ee-9512-afd832c76940] _ensure_no_port_binding_failure(port) [ 746.877600] env[61923]: ERROR nova.compute.manager [instance: 7c0341f9-7401-42ee-9512-afd832c76940] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 746.877600] env[61923]: ERROR nova.compute.manager [instance: 7c0341f9-7401-42ee-9512-afd832c76940] raise exception.PortBindingFailed(port_id=port['id']) [ 746.877600] env[61923]: ERROR nova.compute.manager [instance: 7c0341f9-7401-42ee-9512-afd832c76940] nova.exception.PortBindingFailed: Binding failed for port e84bf904-db99-4abc-90aa-4be30a8f32e8, please check neutron logs for more information. [ 746.877600] env[61923]: ERROR nova.compute.manager [instance: 7c0341f9-7401-42ee-9512-afd832c76940] [ 746.877912] env[61923]: DEBUG nova.compute.utils [None req-40b011d4-39ab-4684-a124-b3ba87d489df tempest-ServerPasswordTestJSON-595643156 tempest-ServerPasswordTestJSON-595643156-project-member] [instance: 7c0341f9-7401-42ee-9512-afd832c76940] Binding failed for port e84bf904-db99-4abc-90aa-4be30a8f32e8, please check neutron logs for more information. {{(pid=61923) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 746.878431] env[61923]: DEBUG oslo_concurrency.lockutils [None req-f1ee1de3-0f72-403f-9698-da4536b6f37f tempest-ServerDiagnosticsV248Test-1822646244 tempest-ServerDiagnosticsV248Test-1822646244-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.255s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 746.879905] env[61923]: INFO nova.compute.claims [None req-f1ee1de3-0f72-403f-9698-da4536b6f37f tempest-ServerDiagnosticsV248Test-1822646244 tempest-ServerDiagnosticsV248Test-1822646244-project-member] [instance: 655b01ae-76a5-4c09-9245-e629fec8fd4b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 746.882845] env[61923]: DEBUG nova.compute.manager [None req-40b011d4-39ab-4684-a124-b3ba87d489df tempest-ServerPasswordTestJSON-595643156 tempest-ServerPasswordTestJSON-595643156-project-member] [instance: 7c0341f9-7401-42ee-9512-afd832c76940] Build of instance 7c0341f9-7401-42ee-9512-afd832c76940 was re-scheduled: Binding failed for port e84bf904-db99-4abc-90aa-4be30a8f32e8, please check neutron logs for more information. {{(pid=61923) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 746.883292] env[61923]: DEBUG nova.compute.manager [None req-40b011d4-39ab-4684-a124-b3ba87d489df tempest-ServerPasswordTestJSON-595643156 tempest-ServerPasswordTestJSON-595643156-project-member] [instance: 7c0341f9-7401-42ee-9512-afd832c76940] Unplugging VIFs for instance {{(pid=61923) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 746.883599] env[61923]: DEBUG oslo_concurrency.lockutils [None req-40b011d4-39ab-4684-a124-b3ba87d489df tempest-ServerPasswordTestJSON-595643156 tempest-ServerPasswordTestJSON-595643156-project-member] Acquiring lock "refresh_cache-7c0341f9-7401-42ee-9512-afd832c76940" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 746.883669] env[61923]: DEBUG oslo_concurrency.lockutils [None req-40b011d4-39ab-4684-a124-b3ba87d489df tempest-ServerPasswordTestJSON-595643156 tempest-ServerPasswordTestJSON-595643156-project-member] Acquired lock "refresh_cache-7c0341f9-7401-42ee-9512-afd832c76940" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 746.883802] env[61923]: DEBUG nova.network.neutron [None req-40b011d4-39ab-4684-a124-b3ba87d489df tempest-ServerPasswordTestJSON-595643156 tempest-ServerPasswordTestJSON-595643156-project-member] [instance: 7c0341f9-7401-42ee-9512-afd832c76940] Building network info cache for instance {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 747.007171] env[61923]: DEBUG nova.compute.manager [None req-5138192d-6349-432b-b7de-64ec5d391f8d tempest-ServerActionsTestOtherB-987045973 tempest-ServerActionsTestOtherB-987045973-project-member] [instance: b0bc08c1-38d5-45ee-b521-1cffda41c77c] Start spawning the instance on the hypervisor. {{(pid=61923) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 747.023965] env[61923]: INFO nova.compute.manager [-] [instance: b6712878-3270-43f4-9518-d41eb77124b7] Took 1.04 seconds to deallocate network for instance. [ 747.028143] env[61923]: DEBUG nova.compute.claims [None req-ab7b3da0-3a01-44d4-9dfd-3e5d2ad27432 tempest-TenantUsagesTestJSON-871253541 tempest-TenantUsagesTestJSON-871253541-project-member] [instance: b6712878-3270-43f4-9518-d41eb77124b7] Aborting claim: {{(pid=61923) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 747.028348] env[61923]: DEBUG oslo_concurrency.lockutils [None req-ab7b3da0-3a01-44d4-9dfd-3e5d2ad27432 tempest-TenantUsagesTestJSON-871253541 tempest-TenantUsagesTestJSON-871253541-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 747.034439] env[61923]: DEBUG nova.virt.hardware [None req-5138192d-6349-432b-b7de-64ec5d391f8d tempest-ServerActionsTestOtherB-987045973 tempest-ServerActionsTestOtherB-987045973-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-29T20:07:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-29T20:07:35Z,direct_url=,disk_format='vmdk',id=0f153f63-ae0a-45b1-b7f5-7f9b673c947f,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='4e7b5e3b3c3443c8bd5c70f8a15739f5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-29T20:07:36Z,virtual_size=,visibility=), allow threads: False {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 747.034661] env[61923]: DEBUG nova.virt.hardware [None req-5138192d-6349-432b-b7de-64ec5d391f8d tempest-ServerActionsTestOtherB-987045973 tempest-ServerActionsTestOtherB-987045973-project-member] Flavor limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 747.034810] env[61923]: DEBUG nova.virt.hardware [None req-5138192d-6349-432b-b7de-64ec5d391f8d tempest-ServerActionsTestOtherB-987045973 tempest-ServerActionsTestOtherB-987045973-project-member] Image limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 747.034984] env[61923]: DEBUG nova.virt.hardware [None req-5138192d-6349-432b-b7de-64ec5d391f8d tempest-ServerActionsTestOtherB-987045973 tempest-ServerActionsTestOtherB-987045973-project-member] Flavor pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 747.035138] env[61923]: DEBUG nova.virt.hardware [None req-5138192d-6349-432b-b7de-64ec5d391f8d tempest-ServerActionsTestOtherB-987045973 tempest-ServerActionsTestOtherB-987045973-project-member] Image pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 747.035282] env[61923]: DEBUG nova.virt.hardware [None req-5138192d-6349-432b-b7de-64ec5d391f8d tempest-ServerActionsTestOtherB-987045973 tempest-ServerActionsTestOtherB-987045973-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 747.035483] env[61923]: DEBUG nova.virt.hardware [None req-5138192d-6349-432b-b7de-64ec5d391f8d tempest-ServerActionsTestOtherB-987045973 tempest-ServerActionsTestOtherB-987045973-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 747.035632] env[61923]: DEBUG nova.virt.hardware [None req-5138192d-6349-432b-b7de-64ec5d391f8d tempest-ServerActionsTestOtherB-987045973 tempest-ServerActionsTestOtherB-987045973-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 747.035789] env[61923]: DEBUG nova.virt.hardware [None req-5138192d-6349-432b-b7de-64ec5d391f8d tempest-ServerActionsTestOtherB-987045973 tempest-ServerActionsTestOtherB-987045973-project-member] Got 1 possible topologies {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 747.035945] env[61923]: DEBUG nova.virt.hardware [None req-5138192d-6349-432b-b7de-64ec5d391f8d tempest-ServerActionsTestOtherB-987045973 tempest-ServerActionsTestOtherB-987045973-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 747.036129] env[61923]: DEBUG nova.virt.hardware [None req-5138192d-6349-432b-b7de-64ec5d391f8d tempest-ServerActionsTestOtherB-987045973 tempest-ServerActionsTestOtherB-987045973-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 747.037203] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5305d815-e74a-4346-be7c-509ff7efd024 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.044726] env[61923]: ERROR nova.compute.manager [None req-5138192d-6349-432b-b7de-64ec5d391f8d tempest-ServerActionsTestOtherB-987045973 tempest-ServerActionsTestOtherB-987045973-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 129cb36c-cc2b-4205-9d7b-0b4d553b568b, please check neutron logs for more information. [ 747.044726] env[61923]: ERROR nova.compute.manager Traceback (most recent call last): [ 747.044726] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 747.044726] env[61923]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 747.044726] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 747.044726] env[61923]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 747.044726] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 747.044726] env[61923]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 747.044726] env[61923]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 747.044726] env[61923]: ERROR nova.compute.manager self.force_reraise() [ 747.044726] env[61923]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 747.044726] env[61923]: ERROR nova.compute.manager raise self.value [ 747.044726] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 747.044726] env[61923]: ERROR nova.compute.manager updated_port = self._update_port( [ 747.044726] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 747.044726] env[61923]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 747.045314] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 747.045314] env[61923]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 747.045314] env[61923]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 129cb36c-cc2b-4205-9d7b-0b4d553b568b, please check neutron logs for more information. [ 747.045314] env[61923]: ERROR nova.compute.manager [ 747.045314] env[61923]: Traceback (most recent call last): [ 747.045314] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 747.045314] env[61923]: listener.cb(fileno) [ 747.045314] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 747.045314] env[61923]: result = function(*args, **kwargs) [ 747.045314] env[61923]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 747.045314] env[61923]: return func(*args, **kwargs) [ 747.045314] env[61923]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 747.045314] env[61923]: raise e [ 747.045314] env[61923]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 747.045314] env[61923]: nwinfo = self.network_api.allocate_for_instance( [ 747.045314] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 747.045314] env[61923]: created_port_ids = self._update_ports_for_instance( [ 747.045314] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 747.045314] env[61923]: with excutils.save_and_reraise_exception(): [ 747.045314] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 747.045314] env[61923]: self.force_reraise() [ 747.045314] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 747.045314] env[61923]: raise self.value [ 747.045314] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 747.045314] env[61923]: updated_port = self._update_port( [ 747.045314] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 747.045314] env[61923]: _ensure_no_port_binding_failure(port) [ 747.045314] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 747.045314] env[61923]: raise exception.PortBindingFailed(port_id=port['id']) [ 747.046622] env[61923]: nova.exception.PortBindingFailed: Binding failed for port 129cb36c-cc2b-4205-9d7b-0b4d553b568b, please check neutron logs for more information. [ 747.046622] env[61923]: Removing descriptor: 14 [ 747.046622] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87c7cd2e-0591-4702-ab91-066a9f213726 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.059789] env[61923]: ERROR nova.compute.manager [None req-5138192d-6349-432b-b7de-64ec5d391f8d tempest-ServerActionsTestOtherB-987045973 tempest-ServerActionsTestOtherB-987045973-project-member] [instance: b0bc08c1-38d5-45ee-b521-1cffda41c77c] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 129cb36c-cc2b-4205-9d7b-0b4d553b568b, please check neutron logs for more information. [ 747.059789] env[61923]: ERROR nova.compute.manager [instance: b0bc08c1-38d5-45ee-b521-1cffda41c77c] Traceback (most recent call last): [ 747.059789] env[61923]: ERROR nova.compute.manager [instance: b0bc08c1-38d5-45ee-b521-1cffda41c77c] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 747.059789] env[61923]: ERROR nova.compute.manager [instance: b0bc08c1-38d5-45ee-b521-1cffda41c77c] yield resources [ 747.059789] env[61923]: ERROR nova.compute.manager [instance: b0bc08c1-38d5-45ee-b521-1cffda41c77c] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 747.059789] env[61923]: ERROR nova.compute.manager [instance: b0bc08c1-38d5-45ee-b521-1cffda41c77c] self.driver.spawn(context, instance, image_meta, [ 747.059789] env[61923]: ERROR nova.compute.manager [instance: b0bc08c1-38d5-45ee-b521-1cffda41c77c] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 747.059789] env[61923]: ERROR nova.compute.manager [instance: b0bc08c1-38d5-45ee-b521-1cffda41c77c] self._vmops.spawn(context, instance, image_meta, injected_files, [ 747.059789] env[61923]: ERROR nova.compute.manager [instance: b0bc08c1-38d5-45ee-b521-1cffda41c77c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 747.059789] env[61923]: ERROR nova.compute.manager [instance: b0bc08c1-38d5-45ee-b521-1cffda41c77c] vm_ref = self.build_virtual_machine(instance, [ 747.059789] env[61923]: ERROR nova.compute.manager [instance: b0bc08c1-38d5-45ee-b521-1cffda41c77c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 747.060339] env[61923]: ERROR nova.compute.manager [instance: b0bc08c1-38d5-45ee-b521-1cffda41c77c] vif_infos = vmwarevif.get_vif_info(self._session, [ 747.060339] env[61923]: ERROR nova.compute.manager [instance: b0bc08c1-38d5-45ee-b521-1cffda41c77c] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 747.060339] env[61923]: ERROR nova.compute.manager [instance: b0bc08c1-38d5-45ee-b521-1cffda41c77c] for vif in network_info: [ 747.060339] env[61923]: ERROR nova.compute.manager [instance: b0bc08c1-38d5-45ee-b521-1cffda41c77c] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 747.060339] env[61923]: ERROR nova.compute.manager [instance: b0bc08c1-38d5-45ee-b521-1cffda41c77c] return self._sync_wrapper(fn, *args, **kwargs) [ 747.060339] env[61923]: ERROR nova.compute.manager [instance: b0bc08c1-38d5-45ee-b521-1cffda41c77c] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 747.060339] env[61923]: ERROR nova.compute.manager [instance: b0bc08c1-38d5-45ee-b521-1cffda41c77c] self.wait() [ 747.060339] env[61923]: ERROR nova.compute.manager [instance: b0bc08c1-38d5-45ee-b521-1cffda41c77c] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 747.060339] env[61923]: ERROR nova.compute.manager [instance: b0bc08c1-38d5-45ee-b521-1cffda41c77c] self[:] = self._gt.wait() [ 747.060339] env[61923]: ERROR nova.compute.manager [instance: b0bc08c1-38d5-45ee-b521-1cffda41c77c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 747.060339] env[61923]: ERROR nova.compute.manager [instance: b0bc08c1-38d5-45ee-b521-1cffda41c77c] return self._exit_event.wait() [ 747.060339] env[61923]: ERROR nova.compute.manager [instance: b0bc08c1-38d5-45ee-b521-1cffda41c77c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 747.060339] env[61923]: ERROR nova.compute.manager [instance: b0bc08c1-38d5-45ee-b521-1cffda41c77c] current.throw(*self._exc) [ 747.060927] env[61923]: ERROR nova.compute.manager [instance: b0bc08c1-38d5-45ee-b521-1cffda41c77c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 747.060927] env[61923]: ERROR nova.compute.manager [instance: b0bc08c1-38d5-45ee-b521-1cffda41c77c] result = function(*args, **kwargs) [ 747.060927] env[61923]: ERROR nova.compute.manager [instance: b0bc08c1-38d5-45ee-b521-1cffda41c77c] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 747.060927] env[61923]: ERROR nova.compute.manager [instance: b0bc08c1-38d5-45ee-b521-1cffda41c77c] return func(*args, **kwargs) [ 747.060927] env[61923]: ERROR nova.compute.manager [instance: b0bc08c1-38d5-45ee-b521-1cffda41c77c] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 747.060927] env[61923]: ERROR nova.compute.manager [instance: b0bc08c1-38d5-45ee-b521-1cffda41c77c] raise e [ 747.060927] env[61923]: ERROR nova.compute.manager [instance: b0bc08c1-38d5-45ee-b521-1cffda41c77c] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 747.060927] env[61923]: ERROR nova.compute.manager [instance: b0bc08c1-38d5-45ee-b521-1cffda41c77c] nwinfo = self.network_api.allocate_for_instance( [ 747.060927] env[61923]: ERROR nova.compute.manager [instance: b0bc08c1-38d5-45ee-b521-1cffda41c77c] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 747.060927] env[61923]: ERROR nova.compute.manager [instance: b0bc08c1-38d5-45ee-b521-1cffda41c77c] created_port_ids = self._update_ports_for_instance( [ 747.060927] env[61923]: ERROR nova.compute.manager [instance: b0bc08c1-38d5-45ee-b521-1cffda41c77c] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 747.060927] env[61923]: ERROR nova.compute.manager [instance: b0bc08c1-38d5-45ee-b521-1cffda41c77c] with excutils.save_and_reraise_exception(): [ 747.060927] env[61923]: ERROR nova.compute.manager [instance: b0bc08c1-38d5-45ee-b521-1cffda41c77c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 747.061440] env[61923]: ERROR nova.compute.manager [instance: b0bc08c1-38d5-45ee-b521-1cffda41c77c] self.force_reraise() [ 747.061440] env[61923]: ERROR nova.compute.manager [instance: b0bc08c1-38d5-45ee-b521-1cffda41c77c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 747.061440] env[61923]: ERROR nova.compute.manager [instance: b0bc08c1-38d5-45ee-b521-1cffda41c77c] raise self.value [ 747.061440] env[61923]: ERROR nova.compute.manager [instance: b0bc08c1-38d5-45ee-b521-1cffda41c77c] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 747.061440] env[61923]: ERROR nova.compute.manager [instance: b0bc08c1-38d5-45ee-b521-1cffda41c77c] updated_port = self._update_port( [ 747.061440] env[61923]: ERROR nova.compute.manager [instance: b0bc08c1-38d5-45ee-b521-1cffda41c77c] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 747.061440] env[61923]: ERROR nova.compute.manager [instance: b0bc08c1-38d5-45ee-b521-1cffda41c77c] _ensure_no_port_binding_failure(port) [ 747.061440] env[61923]: ERROR nova.compute.manager [instance: b0bc08c1-38d5-45ee-b521-1cffda41c77c] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 747.061440] env[61923]: ERROR nova.compute.manager [instance: b0bc08c1-38d5-45ee-b521-1cffda41c77c] raise exception.PortBindingFailed(port_id=port['id']) [ 747.061440] env[61923]: ERROR nova.compute.manager [instance: b0bc08c1-38d5-45ee-b521-1cffda41c77c] nova.exception.PortBindingFailed: Binding failed for port 129cb36c-cc2b-4205-9d7b-0b4d553b568b, please check neutron logs for more information. [ 747.061440] env[61923]: ERROR nova.compute.manager [instance: b0bc08c1-38d5-45ee-b521-1cffda41c77c] [ 747.061440] env[61923]: INFO nova.compute.manager [None req-5138192d-6349-432b-b7de-64ec5d391f8d tempest-ServerActionsTestOtherB-987045973 tempest-ServerActionsTestOtherB-987045973-project-member] [instance: b0bc08c1-38d5-45ee-b521-1cffda41c77c] Terminating instance [ 747.062756] env[61923]: DEBUG oslo_concurrency.lockutils [None req-5138192d-6349-432b-b7de-64ec5d391f8d tempest-ServerActionsTestOtherB-987045973 tempest-ServerActionsTestOtherB-987045973-project-member] Acquiring lock "refresh_cache-b0bc08c1-38d5-45ee-b521-1cffda41c77c" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 747.062915] env[61923]: DEBUG oslo_concurrency.lockutils [None req-5138192d-6349-432b-b7de-64ec5d391f8d tempest-ServerActionsTestOtherB-987045973 tempest-ServerActionsTestOtherB-987045973-project-member] Acquired lock "refresh_cache-b0bc08c1-38d5-45ee-b521-1cffda41c77c" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 747.063093] env[61923]: DEBUG nova.network.neutron [None req-5138192d-6349-432b-b7de-64ec5d391f8d tempest-ServerActionsTestOtherB-987045973 tempest-ServerActionsTestOtherB-987045973-project-member] [instance: b0bc08c1-38d5-45ee-b521-1cffda41c77c] Building network info cache for instance {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 747.207840] env[61923]: DEBUG oslo_concurrency.lockutils [req-efee9203-151c-4b0e-890b-4aebe517d7b4 req-e9686878-2ed8-4cca-975b-78e24d7a772d service nova] Releasing lock "refresh_cache-b6712878-3270-43f4-9518-d41eb77124b7" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 747.208214] env[61923]: DEBUG nova.compute.manager [req-efee9203-151c-4b0e-890b-4aebe517d7b4 req-e9686878-2ed8-4cca-975b-78e24d7a772d service nova] [instance: b6712878-3270-43f4-9518-d41eb77124b7] Received event network-vif-deleted-a4338c60-f3e2-4844-b170-c852adcfcb25 {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 747.402652] env[61923]: DEBUG nova.network.neutron [None req-40b011d4-39ab-4684-a124-b3ba87d489df tempest-ServerPasswordTestJSON-595643156 tempest-ServerPasswordTestJSON-595643156-project-member] [instance: 7c0341f9-7401-42ee-9512-afd832c76940] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 747.497309] env[61923]: DEBUG nova.network.neutron [None req-40b011d4-39ab-4684-a124-b3ba87d489df tempest-ServerPasswordTestJSON-595643156 tempest-ServerPasswordTestJSON-595643156-project-member] [instance: 7c0341f9-7401-42ee-9512-afd832c76940] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 747.577850] env[61923]: DEBUG nova.network.neutron [None req-5138192d-6349-432b-b7de-64ec5d391f8d tempest-ServerActionsTestOtherB-987045973 tempest-ServerActionsTestOtherB-987045973-project-member] [instance: b0bc08c1-38d5-45ee-b521-1cffda41c77c] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 747.688381] env[61923]: DEBUG nova.network.neutron [None req-5138192d-6349-432b-b7de-64ec5d391f8d tempest-ServerActionsTestOtherB-987045973 tempest-ServerActionsTestOtherB-987045973-project-member] [instance: b0bc08c1-38d5-45ee-b521-1cffda41c77c] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 747.999921] env[61923]: DEBUG oslo_concurrency.lockutils [None req-40b011d4-39ab-4684-a124-b3ba87d489df tempest-ServerPasswordTestJSON-595643156 tempest-ServerPasswordTestJSON-595643156-project-member] Releasing lock "refresh_cache-7c0341f9-7401-42ee-9512-afd832c76940" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 748.000237] env[61923]: DEBUG nova.compute.manager [None req-40b011d4-39ab-4684-a124-b3ba87d489df tempest-ServerPasswordTestJSON-595643156 tempest-ServerPasswordTestJSON-595643156-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61923) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 748.000474] env[61923]: DEBUG nova.compute.manager [None req-40b011d4-39ab-4684-a124-b3ba87d489df tempest-ServerPasswordTestJSON-595643156 tempest-ServerPasswordTestJSON-595643156-project-member] [instance: 7c0341f9-7401-42ee-9512-afd832c76940] Deallocating network for instance {{(pid=61923) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 748.000670] env[61923]: DEBUG nova.network.neutron [None req-40b011d4-39ab-4684-a124-b3ba87d489df tempest-ServerPasswordTestJSON-595643156 tempest-ServerPasswordTestJSON-595643156-project-member] [instance: 7c0341f9-7401-42ee-9512-afd832c76940] deallocate_for_instance() {{(pid=61923) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 748.014063] env[61923]: DEBUG nova.network.neutron [None req-40b011d4-39ab-4684-a124-b3ba87d489df tempest-ServerPasswordTestJSON-595643156 tempest-ServerPasswordTestJSON-595643156-project-member] [instance: 7c0341f9-7401-42ee-9512-afd832c76940] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 748.125790] env[61923]: DEBUG nova.compute.manager [req-de3349a9-e320-4ddc-8620-127eea36761a req-15de71a8-01d9-418f-b6bd-daaa4957b6d5 service nova] [instance: b0bc08c1-38d5-45ee-b521-1cffda41c77c] Received event network-changed-129cb36c-cc2b-4205-9d7b-0b4d553b568b {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 748.125934] env[61923]: DEBUG nova.compute.manager [req-de3349a9-e320-4ddc-8620-127eea36761a req-15de71a8-01d9-418f-b6bd-daaa4957b6d5 service nova] [instance: b0bc08c1-38d5-45ee-b521-1cffda41c77c] Refreshing instance network info cache due to event network-changed-129cb36c-cc2b-4205-9d7b-0b4d553b568b. {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 748.126159] env[61923]: DEBUG oslo_concurrency.lockutils [req-de3349a9-e320-4ddc-8620-127eea36761a req-15de71a8-01d9-418f-b6bd-daaa4957b6d5 service nova] Acquiring lock "refresh_cache-b0bc08c1-38d5-45ee-b521-1cffda41c77c" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 748.149751] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6bd5b092-c0e0-41cf-b761-963bb793fb1a {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.157437] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8971f4f3-d299-4d99-9774-6719af8a6ea0 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.187386] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1fbbefd8-ed71-49bd-a59b-e5caef458590 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.191244] env[61923]: DEBUG oslo_concurrency.lockutils [None req-5138192d-6349-432b-b7de-64ec5d391f8d tempest-ServerActionsTestOtherB-987045973 tempest-ServerActionsTestOtherB-987045973-project-member] Releasing lock "refresh_cache-b0bc08c1-38d5-45ee-b521-1cffda41c77c" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 748.191609] env[61923]: DEBUG nova.compute.manager [None req-5138192d-6349-432b-b7de-64ec5d391f8d tempest-ServerActionsTestOtherB-987045973 tempest-ServerActionsTestOtherB-987045973-project-member] [instance: b0bc08c1-38d5-45ee-b521-1cffda41c77c] Start destroying the instance on the hypervisor. {{(pid=61923) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 748.191790] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-5138192d-6349-432b-b7de-64ec5d391f8d tempest-ServerActionsTestOtherB-987045973 tempest-ServerActionsTestOtherB-987045973-project-member] [instance: b0bc08c1-38d5-45ee-b521-1cffda41c77c] Destroying instance {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 748.193733] env[61923]: DEBUG oslo_concurrency.lockutils [req-de3349a9-e320-4ddc-8620-127eea36761a req-15de71a8-01d9-418f-b6bd-daaa4957b6d5 service nova] Acquired lock "refresh_cache-b0bc08c1-38d5-45ee-b521-1cffda41c77c" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 748.193913] env[61923]: DEBUG nova.network.neutron [req-de3349a9-e320-4ddc-8620-127eea36761a req-15de71a8-01d9-418f-b6bd-daaa4957b6d5 service nova] [instance: b0bc08c1-38d5-45ee-b521-1cffda41c77c] Refreshing network info cache for port 129cb36c-cc2b-4205-9d7b-0b4d553b568b {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 748.194934] env[61923]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-9959d7ba-0ff2-4d5e-8047-df279c1b9090 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.197825] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5334bfb8-220a-48fa-8313-8e17da0600a6 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.214126] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc6242fe-2ce4-4b88-8a5c-da3f76bc88b3 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.224039] env[61923]: DEBUG nova.compute.provider_tree [None req-f1ee1de3-0f72-403f-9698-da4536b6f37f tempest-ServerDiagnosticsV248Test-1822646244 tempest-ServerDiagnosticsV248Test-1822646244-project-member] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 748.236385] env[61923]: WARNING nova.virt.vmwareapi.vmops [None req-5138192d-6349-432b-b7de-64ec5d391f8d tempest-ServerActionsTestOtherB-987045973 tempest-ServerActionsTestOtherB-987045973-project-member] [instance: b0bc08c1-38d5-45ee-b521-1cffda41c77c] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance b0bc08c1-38d5-45ee-b521-1cffda41c77c could not be found. [ 748.236592] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-5138192d-6349-432b-b7de-64ec5d391f8d tempest-ServerActionsTestOtherB-987045973 tempest-ServerActionsTestOtherB-987045973-project-member] [instance: b0bc08c1-38d5-45ee-b521-1cffda41c77c] Instance destroyed {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 748.236764] env[61923]: INFO nova.compute.manager [None req-5138192d-6349-432b-b7de-64ec5d391f8d tempest-ServerActionsTestOtherB-987045973 tempest-ServerActionsTestOtherB-987045973-project-member] [instance: b0bc08c1-38d5-45ee-b521-1cffda41c77c] Took 0.04 seconds to destroy the instance on the hypervisor. [ 748.236985] env[61923]: DEBUG oslo.service.loopingcall [None req-5138192d-6349-432b-b7de-64ec5d391f8d tempest-ServerActionsTestOtherB-987045973 tempest-ServerActionsTestOtherB-987045973-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61923) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 748.237211] env[61923]: DEBUG nova.compute.manager [-] [instance: b0bc08c1-38d5-45ee-b521-1cffda41c77c] Deallocating network for instance {{(pid=61923) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 748.237363] env[61923]: DEBUG nova.network.neutron [-] [instance: b0bc08c1-38d5-45ee-b521-1cffda41c77c] deallocate_for_instance() {{(pid=61923) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 748.251967] env[61923]: DEBUG nova.network.neutron [-] [instance: b0bc08c1-38d5-45ee-b521-1cffda41c77c] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 748.517184] env[61923]: DEBUG nova.network.neutron [None req-40b011d4-39ab-4684-a124-b3ba87d489df tempest-ServerPasswordTestJSON-595643156 tempest-ServerPasswordTestJSON-595643156-project-member] [instance: 7c0341f9-7401-42ee-9512-afd832c76940] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 748.720389] env[61923]: DEBUG nova.network.neutron [req-de3349a9-e320-4ddc-8620-127eea36761a req-15de71a8-01d9-418f-b6bd-daaa4957b6d5 service nova] [instance: b0bc08c1-38d5-45ee-b521-1cffda41c77c] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 748.727310] env[61923]: DEBUG nova.scheduler.client.report [None req-f1ee1de3-0f72-403f-9698-da4536b6f37f tempest-ServerDiagnosticsV248Test-1822646244 tempest-ServerDiagnosticsV248Test-1822646244-project-member] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 748.753788] env[61923]: DEBUG nova.network.neutron [-] [instance: b0bc08c1-38d5-45ee-b521-1cffda41c77c] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 748.800756] env[61923]: DEBUG nova.network.neutron [req-de3349a9-e320-4ddc-8620-127eea36761a req-15de71a8-01d9-418f-b6bd-daaa4957b6d5 service nova] [instance: b0bc08c1-38d5-45ee-b521-1cffda41c77c] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 749.020743] env[61923]: INFO nova.compute.manager [None req-40b011d4-39ab-4684-a124-b3ba87d489df tempest-ServerPasswordTestJSON-595643156 tempest-ServerPasswordTestJSON-595643156-project-member] [instance: 7c0341f9-7401-42ee-9512-afd832c76940] Took 1.02 seconds to deallocate network for instance. [ 749.232183] env[61923]: DEBUG oslo_concurrency.lockutils [None req-f1ee1de3-0f72-403f-9698-da4536b6f37f tempest-ServerDiagnosticsV248Test-1822646244 tempest-ServerDiagnosticsV248Test-1822646244-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.354s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 749.232717] env[61923]: DEBUG nova.compute.manager [None req-f1ee1de3-0f72-403f-9698-da4536b6f37f tempest-ServerDiagnosticsV248Test-1822646244 tempest-ServerDiagnosticsV248Test-1822646244-project-member] [instance: 655b01ae-76a5-4c09-9245-e629fec8fd4b] Start building networks asynchronously for instance. {{(pid=61923) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 749.235576] env[61923]: DEBUG oslo_concurrency.lockutils [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 16.162s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 749.235778] env[61923]: DEBUG oslo_concurrency.lockutils [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 749.235935] env[61923]: DEBUG nova.compute.resource_tracker [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61923) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 749.236232] env[61923]: DEBUG oslo_concurrency.lockutils [None req-13202777-73e8-40d3-a6ad-7bb49e61267a tempest-ImagesNegativeTestJSON-1202128281 tempest-ImagesNegativeTestJSON-1202128281-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.118s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 749.239495] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0bae373f-3049-4190-90d4-ffab01502a1a {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.247902] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eea5fdeb-af68-47a6-b59e-8c187c984373 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.263284] env[61923]: INFO nova.compute.manager [-] [instance: b0bc08c1-38d5-45ee-b521-1cffda41c77c] Took 1.03 seconds to deallocate network for instance. [ 749.265995] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65e03183-d1f7-4c19-bd76-28a04be7f66d {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.268605] env[61923]: DEBUG nova.compute.claims [None req-5138192d-6349-432b-b7de-64ec5d391f8d tempest-ServerActionsTestOtherB-987045973 tempest-ServerActionsTestOtherB-987045973-project-member] [instance: b0bc08c1-38d5-45ee-b521-1cffda41c77c] Aborting claim: {{(pid=61923) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 749.268774] env[61923]: DEBUG oslo_concurrency.lockutils [None req-5138192d-6349-432b-b7de-64ec5d391f8d tempest-ServerActionsTestOtherB-987045973 tempest-ServerActionsTestOtherB-987045973-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 749.273517] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dfc5aa06-7208-45d2-aeb9-4c5943717318 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.302064] env[61923]: DEBUG nova.compute.resource_tracker [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181384MB free_disk=178GB free_vcpus=48 pci_devices=None {{(pid=61923) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 749.302209] env[61923]: DEBUG oslo_concurrency.lockutils [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 749.302675] env[61923]: DEBUG oslo_concurrency.lockutils [req-de3349a9-e320-4ddc-8620-127eea36761a req-15de71a8-01d9-418f-b6bd-daaa4957b6d5 service nova] Releasing lock "refresh_cache-b0bc08c1-38d5-45ee-b521-1cffda41c77c" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 749.302891] env[61923]: DEBUG nova.compute.manager [req-de3349a9-e320-4ddc-8620-127eea36761a req-15de71a8-01d9-418f-b6bd-daaa4957b6d5 service nova] [instance: b0bc08c1-38d5-45ee-b521-1cffda41c77c] Received event network-vif-deleted-129cb36c-cc2b-4205-9d7b-0b4d553b568b {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 749.741059] env[61923]: DEBUG nova.compute.utils [None req-f1ee1de3-0f72-403f-9698-da4536b6f37f tempest-ServerDiagnosticsV248Test-1822646244 tempest-ServerDiagnosticsV248Test-1822646244-project-member] Using /dev/sd instead of None {{(pid=61923) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 749.742464] env[61923]: DEBUG nova.compute.manager [None req-f1ee1de3-0f72-403f-9698-da4536b6f37f tempest-ServerDiagnosticsV248Test-1822646244 tempest-ServerDiagnosticsV248Test-1822646244-project-member] [instance: 655b01ae-76a5-4c09-9245-e629fec8fd4b] Not allocating networking since 'none' was specified. {{(pid=61923) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 749.997386] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09111c15-6b20-4693-addb-2ad21465e2d3 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.004683] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1226124e-6a84-4155-a241-025cfea90f49 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.038227] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b875ef71-4cde-46e4-afa5-a0bf8f0a8535 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.045095] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48987dfb-7b41-482e-a5cc-f951daed0904 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.058277] env[61923]: DEBUG nova.compute.provider_tree [None req-13202777-73e8-40d3-a6ad-7bb49e61267a tempest-ImagesNegativeTestJSON-1202128281 tempest-ImagesNegativeTestJSON-1202128281-project-member] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 750.060608] env[61923]: INFO nova.scheduler.client.report [None req-40b011d4-39ab-4684-a124-b3ba87d489df tempest-ServerPasswordTestJSON-595643156 tempest-ServerPasswordTestJSON-595643156-project-member] Deleted allocations for instance 7c0341f9-7401-42ee-9512-afd832c76940 [ 750.246544] env[61923]: DEBUG nova.compute.manager [None req-f1ee1de3-0f72-403f-9698-da4536b6f37f tempest-ServerDiagnosticsV248Test-1822646244 tempest-ServerDiagnosticsV248Test-1822646244-project-member] [instance: 655b01ae-76a5-4c09-9245-e629fec8fd4b] Start building block device mappings for instance. {{(pid=61923) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 750.568765] env[61923]: DEBUG nova.scheduler.client.report [None req-13202777-73e8-40d3-a6ad-7bb49e61267a tempest-ImagesNegativeTestJSON-1202128281 tempest-ImagesNegativeTestJSON-1202128281-project-member] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 750.572077] env[61923]: DEBUG oslo_concurrency.lockutils [None req-40b011d4-39ab-4684-a124-b3ba87d489df tempest-ServerPasswordTestJSON-595643156 tempest-ServerPasswordTestJSON-595643156-project-member] Lock "7c0341f9-7401-42ee-9512-afd832c76940" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 146.077s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 751.073588] env[61923]: DEBUG oslo_concurrency.lockutils [None req-13202777-73e8-40d3-a6ad-7bb49e61267a tempest-ImagesNegativeTestJSON-1202128281 tempest-ImagesNegativeTestJSON-1202128281-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.837s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 751.074341] env[61923]: ERROR nova.compute.manager [None req-13202777-73e8-40d3-a6ad-7bb49e61267a tempest-ImagesNegativeTestJSON-1202128281 tempest-ImagesNegativeTestJSON-1202128281-project-member] [instance: f7e1d517-3e39-4381-83be-0152368b8120] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 66c485f3-bd28-41e5-a69c-fc82fa725d19, please check neutron logs for more information. [ 751.074341] env[61923]: ERROR nova.compute.manager [instance: f7e1d517-3e39-4381-83be-0152368b8120] Traceback (most recent call last): [ 751.074341] env[61923]: ERROR nova.compute.manager [instance: f7e1d517-3e39-4381-83be-0152368b8120] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 751.074341] env[61923]: ERROR nova.compute.manager [instance: f7e1d517-3e39-4381-83be-0152368b8120] self.driver.spawn(context, instance, image_meta, [ 751.074341] env[61923]: ERROR nova.compute.manager [instance: f7e1d517-3e39-4381-83be-0152368b8120] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 751.074341] env[61923]: ERROR nova.compute.manager [instance: f7e1d517-3e39-4381-83be-0152368b8120] self._vmops.spawn(context, instance, image_meta, injected_files, [ 751.074341] env[61923]: ERROR nova.compute.manager [instance: f7e1d517-3e39-4381-83be-0152368b8120] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 751.074341] env[61923]: ERROR nova.compute.manager [instance: f7e1d517-3e39-4381-83be-0152368b8120] vm_ref = self.build_virtual_machine(instance, [ 751.074341] env[61923]: ERROR nova.compute.manager [instance: f7e1d517-3e39-4381-83be-0152368b8120] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 751.074341] env[61923]: ERROR nova.compute.manager [instance: f7e1d517-3e39-4381-83be-0152368b8120] vif_infos = vmwarevif.get_vif_info(self._session, [ 751.074341] env[61923]: ERROR nova.compute.manager [instance: f7e1d517-3e39-4381-83be-0152368b8120] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 751.074758] env[61923]: ERROR nova.compute.manager [instance: f7e1d517-3e39-4381-83be-0152368b8120] for vif in network_info: [ 751.074758] env[61923]: ERROR nova.compute.manager [instance: f7e1d517-3e39-4381-83be-0152368b8120] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 751.074758] env[61923]: ERROR nova.compute.manager [instance: f7e1d517-3e39-4381-83be-0152368b8120] return self._sync_wrapper(fn, *args, **kwargs) [ 751.074758] env[61923]: ERROR nova.compute.manager [instance: f7e1d517-3e39-4381-83be-0152368b8120] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 751.074758] env[61923]: ERROR nova.compute.manager [instance: f7e1d517-3e39-4381-83be-0152368b8120] self.wait() [ 751.074758] env[61923]: ERROR nova.compute.manager [instance: f7e1d517-3e39-4381-83be-0152368b8120] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 751.074758] env[61923]: ERROR nova.compute.manager [instance: f7e1d517-3e39-4381-83be-0152368b8120] self[:] = self._gt.wait() [ 751.074758] env[61923]: ERROR nova.compute.manager [instance: f7e1d517-3e39-4381-83be-0152368b8120] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 751.074758] env[61923]: ERROR nova.compute.manager [instance: f7e1d517-3e39-4381-83be-0152368b8120] return self._exit_event.wait() [ 751.074758] env[61923]: ERROR nova.compute.manager [instance: f7e1d517-3e39-4381-83be-0152368b8120] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 751.074758] env[61923]: ERROR nova.compute.manager [instance: f7e1d517-3e39-4381-83be-0152368b8120] result = hub.switch() [ 751.074758] env[61923]: ERROR nova.compute.manager [instance: f7e1d517-3e39-4381-83be-0152368b8120] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 751.074758] env[61923]: ERROR nova.compute.manager [instance: f7e1d517-3e39-4381-83be-0152368b8120] return self.greenlet.switch() [ 751.075190] env[61923]: ERROR nova.compute.manager [instance: f7e1d517-3e39-4381-83be-0152368b8120] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 751.075190] env[61923]: ERROR nova.compute.manager [instance: f7e1d517-3e39-4381-83be-0152368b8120] result = function(*args, **kwargs) [ 751.075190] env[61923]: ERROR nova.compute.manager [instance: f7e1d517-3e39-4381-83be-0152368b8120] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 751.075190] env[61923]: ERROR nova.compute.manager [instance: f7e1d517-3e39-4381-83be-0152368b8120] return func(*args, **kwargs) [ 751.075190] env[61923]: ERROR nova.compute.manager [instance: f7e1d517-3e39-4381-83be-0152368b8120] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 751.075190] env[61923]: ERROR nova.compute.manager [instance: f7e1d517-3e39-4381-83be-0152368b8120] raise e [ 751.075190] env[61923]: ERROR nova.compute.manager [instance: f7e1d517-3e39-4381-83be-0152368b8120] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 751.075190] env[61923]: ERROR nova.compute.manager [instance: f7e1d517-3e39-4381-83be-0152368b8120] nwinfo = self.network_api.allocate_for_instance( [ 751.075190] env[61923]: ERROR nova.compute.manager [instance: f7e1d517-3e39-4381-83be-0152368b8120] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 751.075190] env[61923]: ERROR nova.compute.manager [instance: f7e1d517-3e39-4381-83be-0152368b8120] created_port_ids = self._update_ports_for_instance( [ 751.075190] env[61923]: ERROR nova.compute.manager [instance: f7e1d517-3e39-4381-83be-0152368b8120] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 751.075190] env[61923]: ERROR nova.compute.manager [instance: f7e1d517-3e39-4381-83be-0152368b8120] with excutils.save_and_reraise_exception(): [ 751.075190] env[61923]: ERROR nova.compute.manager [instance: f7e1d517-3e39-4381-83be-0152368b8120] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 751.075646] env[61923]: ERROR nova.compute.manager [instance: f7e1d517-3e39-4381-83be-0152368b8120] self.force_reraise() [ 751.075646] env[61923]: ERROR nova.compute.manager [instance: f7e1d517-3e39-4381-83be-0152368b8120] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 751.075646] env[61923]: ERROR nova.compute.manager [instance: f7e1d517-3e39-4381-83be-0152368b8120] raise self.value [ 751.075646] env[61923]: ERROR nova.compute.manager [instance: f7e1d517-3e39-4381-83be-0152368b8120] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 751.075646] env[61923]: ERROR nova.compute.manager [instance: f7e1d517-3e39-4381-83be-0152368b8120] updated_port = self._update_port( [ 751.075646] env[61923]: ERROR nova.compute.manager [instance: f7e1d517-3e39-4381-83be-0152368b8120] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 751.075646] env[61923]: ERROR nova.compute.manager [instance: f7e1d517-3e39-4381-83be-0152368b8120] _ensure_no_port_binding_failure(port) [ 751.075646] env[61923]: ERROR nova.compute.manager [instance: f7e1d517-3e39-4381-83be-0152368b8120] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 751.075646] env[61923]: ERROR nova.compute.manager [instance: f7e1d517-3e39-4381-83be-0152368b8120] raise exception.PortBindingFailed(port_id=port['id']) [ 751.075646] env[61923]: ERROR nova.compute.manager [instance: f7e1d517-3e39-4381-83be-0152368b8120] nova.exception.PortBindingFailed: Binding failed for port 66c485f3-bd28-41e5-a69c-fc82fa725d19, please check neutron logs for more information. [ 751.075646] env[61923]: ERROR nova.compute.manager [instance: f7e1d517-3e39-4381-83be-0152368b8120] [ 751.076022] env[61923]: DEBUG nova.compute.utils [None req-13202777-73e8-40d3-a6ad-7bb49e61267a tempest-ImagesNegativeTestJSON-1202128281 tempest-ImagesNegativeTestJSON-1202128281-project-member] [instance: f7e1d517-3e39-4381-83be-0152368b8120] Binding failed for port 66c485f3-bd28-41e5-a69c-fc82fa725d19, please check neutron logs for more information. {{(pid=61923) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 751.076582] env[61923]: DEBUG oslo_concurrency.lockutils [None req-f7b52f4c-b2e3-49f8-b056-b7254f72bb20 tempest-AttachInterfacesUnderV243Test-498638506 tempest-AttachInterfacesUnderV243Test-498638506-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 13.157s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 751.079728] env[61923]: DEBUG nova.compute.manager [None req-d3d04fc8-2bde-4c80-b2d6-dd8057edd757 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] [instance: b497f3f3-001f-4e5e-b6b7-8e2006116856] Starting instance... {{(pid=61923) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 751.082372] env[61923]: DEBUG nova.compute.manager [None req-13202777-73e8-40d3-a6ad-7bb49e61267a tempest-ImagesNegativeTestJSON-1202128281 tempest-ImagesNegativeTestJSON-1202128281-project-member] [instance: f7e1d517-3e39-4381-83be-0152368b8120] Build of instance f7e1d517-3e39-4381-83be-0152368b8120 was re-scheduled: Binding failed for port 66c485f3-bd28-41e5-a69c-fc82fa725d19, please check neutron logs for more information. {{(pid=61923) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 751.083082] env[61923]: DEBUG nova.compute.manager [None req-13202777-73e8-40d3-a6ad-7bb49e61267a tempest-ImagesNegativeTestJSON-1202128281 tempest-ImagesNegativeTestJSON-1202128281-project-member] [instance: f7e1d517-3e39-4381-83be-0152368b8120] Unplugging VIFs for instance {{(pid=61923) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 751.083082] env[61923]: DEBUG oslo_concurrency.lockutils [None req-13202777-73e8-40d3-a6ad-7bb49e61267a tempest-ImagesNegativeTestJSON-1202128281 tempest-ImagesNegativeTestJSON-1202128281-project-member] Acquiring lock "refresh_cache-f7e1d517-3e39-4381-83be-0152368b8120" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 751.083082] env[61923]: DEBUG oslo_concurrency.lockutils [None req-13202777-73e8-40d3-a6ad-7bb49e61267a tempest-ImagesNegativeTestJSON-1202128281 tempest-ImagesNegativeTestJSON-1202128281-project-member] Acquired lock "refresh_cache-f7e1d517-3e39-4381-83be-0152368b8120" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 751.083257] env[61923]: DEBUG nova.network.neutron [None req-13202777-73e8-40d3-a6ad-7bb49e61267a tempest-ImagesNegativeTestJSON-1202128281 tempest-ImagesNegativeTestJSON-1202128281-project-member] [instance: f7e1d517-3e39-4381-83be-0152368b8120] Building network info cache for instance {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 751.256937] env[61923]: DEBUG nova.compute.manager [None req-f1ee1de3-0f72-403f-9698-da4536b6f37f tempest-ServerDiagnosticsV248Test-1822646244 tempest-ServerDiagnosticsV248Test-1822646244-project-member] [instance: 655b01ae-76a5-4c09-9245-e629fec8fd4b] Start spawning the instance on the hypervisor. {{(pid=61923) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 751.281699] env[61923]: DEBUG nova.virt.hardware [None req-f1ee1de3-0f72-403f-9698-da4536b6f37f tempest-ServerDiagnosticsV248Test-1822646244 tempest-ServerDiagnosticsV248Test-1822646244-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-29T20:07:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-29T20:07:35Z,direct_url=,disk_format='vmdk',id=0f153f63-ae0a-45b1-b7f5-7f9b673c947f,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='4e7b5e3b3c3443c8bd5c70f8a15739f5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-29T20:07:36Z,virtual_size=,visibility=), allow threads: False {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 751.281978] env[61923]: DEBUG nova.virt.hardware [None req-f1ee1de3-0f72-403f-9698-da4536b6f37f tempest-ServerDiagnosticsV248Test-1822646244 tempest-ServerDiagnosticsV248Test-1822646244-project-member] Flavor limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 751.282114] env[61923]: DEBUG nova.virt.hardware [None req-f1ee1de3-0f72-403f-9698-da4536b6f37f tempest-ServerDiagnosticsV248Test-1822646244 tempest-ServerDiagnosticsV248Test-1822646244-project-member] Image limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 751.282297] env[61923]: DEBUG nova.virt.hardware [None req-f1ee1de3-0f72-403f-9698-da4536b6f37f tempest-ServerDiagnosticsV248Test-1822646244 tempest-ServerDiagnosticsV248Test-1822646244-project-member] Flavor pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 751.282441] env[61923]: DEBUG nova.virt.hardware [None req-f1ee1de3-0f72-403f-9698-da4536b6f37f tempest-ServerDiagnosticsV248Test-1822646244 tempest-ServerDiagnosticsV248Test-1822646244-project-member] Image pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 751.282583] env[61923]: DEBUG nova.virt.hardware [None req-f1ee1de3-0f72-403f-9698-da4536b6f37f tempest-ServerDiagnosticsV248Test-1822646244 tempest-ServerDiagnosticsV248Test-1822646244-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 751.282783] env[61923]: DEBUG nova.virt.hardware [None req-f1ee1de3-0f72-403f-9698-da4536b6f37f tempest-ServerDiagnosticsV248Test-1822646244 tempest-ServerDiagnosticsV248Test-1822646244-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 751.282939] env[61923]: DEBUG nova.virt.hardware [None req-f1ee1de3-0f72-403f-9698-da4536b6f37f tempest-ServerDiagnosticsV248Test-1822646244 tempest-ServerDiagnosticsV248Test-1822646244-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 751.283117] env[61923]: DEBUG nova.virt.hardware [None req-f1ee1de3-0f72-403f-9698-da4536b6f37f tempest-ServerDiagnosticsV248Test-1822646244 tempest-ServerDiagnosticsV248Test-1822646244-project-member] Got 1 possible topologies {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 751.283274] env[61923]: DEBUG nova.virt.hardware [None req-f1ee1de3-0f72-403f-9698-da4536b6f37f tempest-ServerDiagnosticsV248Test-1822646244 tempest-ServerDiagnosticsV248Test-1822646244-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 751.283437] env[61923]: DEBUG nova.virt.hardware [None req-f1ee1de3-0f72-403f-9698-da4536b6f37f tempest-ServerDiagnosticsV248Test-1822646244 tempest-ServerDiagnosticsV248Test-1822646244-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 751.284314] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1663bcc-6b4f-4b94-872d-9e0ed0d5c833 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.292392] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd947346-5b5b-430d-b27b-80bb290fc4db {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.305339] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-f1ee1de3-0f72-403f-9698-da4536b6f37f tempest-ServerDiagnosticsV248Test-1822646244 tempest-ServerDiagnosticsV248Test-1822646244-project-member] [instance: 655b01ae-76a5-4c09-9245-e629fec8fd4b] Instance VIF info [] {{(pid=61923) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 751.310778] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-f1ee1de3-0f72-403f-9698-da4536b6f37f tempest-ServerDiagnosticsV248Test-1822646244 tempest-ServerDiagnosticsV248Test-1822646244-project-member] Creating folder: Project (2728d17ff4704edcb91d8c7384a7a727). Parent ref: group-v292629. {{(pid=61923) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 751.311043] env[61923]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-e13d8ae2-c5a8-42c3-99f8-3859b2d12f30 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.321858] env[61923]: INFO nova.virt.vmwareapi.vm_util [None req-f1ee1de3-0f72-403f-9698-da4536b6f37f tempest-ServerDiagnosticsV248Test-1822646244 tempest-ServerDiagnosticsV248Test-1822646244-project-member] Created folder: Project (2728d17ff4704edcb91d8c7384a7a727) in parent group-v292629. [ 751.322055] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-f1ee1de3-0f72-403f-9698-da4536b6f37f tempest-ServerDiagnosticsV248Test-1822646244 tempest-ServerDiagnosticsV248Test-1822646244-project-member] Creating folder: Instances. Parent ref: group-v292643. {{(pid=61923) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 751.322275] env[61923]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-1239bd29-f3c1-4818-bf13-1fe325881cee {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.330708] env[61923]: INFO nova.virt.vmwareapi.vm_util [None req-f1ee1de3-0f72-403f-9698-da4536b6f37f tempest-ServerDiagnosticsV248Test-1822646244 tempest-ServerDiagnosticsV248Test-1822646244-project-member] Created folder: Instances in parent group-v292643. [ 751.330933] env[61923]: DEBUG oslo.service.loopingcall [None req-f1ee1de3-0f72-403f-9698-da4536b6f37f tempest-ServerDiagnosticsV248Test-1822646244 tempest-ServerDiagnosticsV248Test-1822646244-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61923) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 751.331128] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 655b01ae-76a5-4c09-9245-e629fec8fd4b] Creating VM on the ESX host {{(pid=61923) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 751.331313] env[61923]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-1438b1e6-dee0-4951-a835-dd14cc1a9aae {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.348277] env[61923]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 751.348277] env[61923]: value = "task-1377485" [ 751.348277] env[61923]: _type = "Task" [ 751.348277] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 751.355903] env[61923]: DEBUG oslo_vmware.api [-] Task: {'id': task-1377485, 'name': CreateVM_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 751.612200] env[61923]: DEBUG oslo_concurrency.lockutils [None req-d3d04fc8-2bde-4c80-b2d6-dd8057edd757 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 751.618972] env[61923]: DEBUG nova.network.neutron [None req-13202777-73e8-40d3-a6ad-7bb49e61267a tempest-ImagesNegativeTestJSON-1202128281 tempest-ImagesNegativeTestJSON-1202128281-project-member] [instance: f7e1d517-3e39-4381-83be-0152368b8120] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 751.791740] env[61923]: DEBUG nova.network.neutron [None req-13202777-73e8-40d3-a6ad-7bb49e61267a tempest-ImagesNegativeTestJSON-1202128281 tempest-ImagesNegativeTestJSON-1202128281-project-member] [instance: f7e1d517-3e39-4381-83be-0152368b8120] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 751.860474] env[61923]: DEBUG oslo_vmware.api [-] Task: {'id': task-1377485, 'name': CreateVM_Task, 'duration_secs': 0.23063} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 751.860631] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 655b01ae-76a5-4c09-9245-e629fec8fd4b] Created VM on the ESX host {{(pid=61923) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 751.863155] env[61923]: DEBUG oslo_concurrency.lockutils [None req-f1ee1de3-0f72-403f-9698-da4536b6f37f tempest-ServerDiagnosticsV248Test-1822646244 tempest-ServerDiagnosticsV248Test-1822646244-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 751.863402] env[61923]: DEBUG oslo_concurrency.lockutils [None req-f1ee1de3-0f72-403f-9698-da4536b6f37f tempest-ServerDiagnosticsV248Test-1822646244 tempest-ServerDiagnosticsV248Test-1822646244-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 751.863759] env[61923]: DEBUG oslo_concurrency.lockutils [None req-f1ee1de3-0f72-403f-9698-da4536b6f37f tempest-ServerDiagnosticsV248Test-1822646244 tempest-ServerDiagnosticsV248Test-1822646244-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 751.866244] env[61923]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2593009b-23cf-46ac-9167-4ba3e9e3cc19 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.871053] env[61923]: DEBUG oslo_vmware.api [None req-f1ee1de3-0f72-403f-9698-da4536b6f37f tempest-ServerDiagnosticsV248Test-1822646244 tempest-ServerDiagnosticsV248Test-1822646244-project-member] Waiting for the task: (returnval){ [ 751.871053] env[61923]: value = "session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52637a08-e735-b258-480d-f52e205e69a3" [ 751.871053] env[61923]: _type = "Task" [ 751.871053] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 751.879166] env[61923]: DEBUG oslo_vmware.api [None req-f1ee1de3-0f72-403f-9698-da4536b6f37f tempest-ServerDiagnosticsV248Test-1822646244 tempest-ServerDiagnosticsV248Test-1822646244-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52637a08-e735-b258-480d-f52e205e69a3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 751.929515] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc42d053-4af5-4c88-812b-ea1c3cd8a24f {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.937181] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-113d1683-ac30-4421-ad33-29bca055a066 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.967535] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf5bcc17-2814-4d50-9af5-3c33c3c775dd {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.974791] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24202886-35dc-4c67-a295-e47e8595647d {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.987973] env[61923]: DEBUG nova.compute.provider_tree [None req-f7b52f4c-b2e3-49f8-b056-b7254f72bb20 tempest-AttachInterfacesUnderV243Test-498638506 tempest-AttachInterfacesUnderV243Test-498638506-project-member] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 752.296067] env[61923]: DEBUG oslo_concurrency.lockutils [None req-13202777-73e8-40d3-a6ad-7bb49e61267a tempest-ImagesNegativeTestJSON-1202128281 tempest-ImagesNegativeTestJSON-1202128281-project-member] Releasing lock "refresh_cache-f7e1d517-3e39-4381-83be-0152368b8120" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 752.296395] env[61923]: DEBUG nova.compute.manager [None req-13202777-73e8-40d3-a6ad-7bb49e61267a tempest-ImagesNegativeTestJSON-1202128281 tempest-ImagesNegativeTestJSON-1202128281-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61923) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 752.296618] env[61923]: DEBUG nova.compute.manager [None req-13202777-73e8-40d3-a6ad-7bb49e61267a tempest-ImagesNegativeTestJSON-1202128281 tempest-ImagesNegativeTestJSON-1202128281-project-member] [instance: f7e1d517-3e39-4381-83be-0152368b8120] Deallocating network for instance {{(pid=61923) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 752.296822] env[61923]: DEBUG nova.network.neutron [None req-13202777-73e8-40d3-a6ad-7bb49e61267a tempest-ImagesNegativeTestJSON-1202128281 tempest-ImagesNegativeTestJSON-1202128281-project-member] [instance: f7e1d517-3e39-4381-83be-0152368b8120] deallocate_for_instance() {{(pid=61923) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 752.334903] env[61923]: DEBUG nova.network.neutron [None req-13202777-73e8-40d3-a6ad-7bb49e61267a tempest-ImagesNegativeTestJSON-1202128281 tempest-ImagesNegativeTestJSON-1202128281-project-member] [instance: f7e1d517-3e39-4381-83be-0152368b8120] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 752.388922] env[61923]: DEBUG oslo_vmware.api [None req-f1ee1de3-0f72-403f-9698-da4536b6f37f tempest-ServerDiagnosticsV248Test-1822646244 tempest-ServerDiagnosticsV248Test-1822646244-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52637a08-e735-b258-480d-f52e205e69a3, 'name': SearchDatastore_Task, 'duration_secs': 0.012216} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 752.389047] env[61923]: DEBUG oslo_concurrency.lockutils [None req-f1ee1de3-0f72-403f-9698-da4536b6f37f tempest-ServerDiagnosticsV248Test-1822646244 tempest-ServerDiagnosticsV248Test-1822646244-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 752.389188] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-f1ee1de3-0f72-403f-9698-da4536b6f37f tempest-ServerDiagnosticsV248Test-1822646244 tempest-ServerDiagnosticsV248Test-1822646244-project-member] [instance: 655b01ae-76a5-4c09-9245-e629fec8fd4b] Processing image 0f153f63-ae0a-45b1-b7f5-7f9b673c947f {{(pid=61923) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 752.389411] env[61923]: DEBUG oslo_concurrency.lockutils [None req-f1ee1de3-0f72-403f-9698-da4536b6f37f tempest-ServerDiagnosticsV248Test-1822646244 tempest-ServerDiagnosticsV248Test-1822646244-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 752.389550] env[61923]: DEBUG oslo_concurrency.lockutils [None req-f1ee1de3-0f72-403f-9698-da4536b6f37f tempest-ServerDiagnosticsV248Test-1822646244 tempest-ServerDiagnosticsV248Test-1822646244-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 752.389935] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-f1ee1de3-0f72-403f-9698-da4536b6f37f tempest-ServerDiagnosticsV248Test-1822646244 tempest-ServerDiagnosticsV248Test-1822646244-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61923) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 752.390037] env[61923]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-35ce0e85-8378-46db-972c-c39f5acf5d2c {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.397820] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-f1ee1de3-0f72-403f-9698-da4536b6f37f tempest-ServerDiagnosticsV248Test-1822646244 tempest-ServerDiagnosticsV248Test-1822646244-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61923) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 752.397820] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-f1ee1de3-0f72-403f-9698-da4536b6f37f tempest-ServerDiagnosticsV248Test-1822646244 tempest-ServerDiagnosticsV248Test-1822646244-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61923) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 752.398595] env[61923]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-286e3702-eda3-4e62-93db-ff8ff6519d3a {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.403610] env[61923]: DEBUG oslo_vmware.api [None req-f1ee1de3-0f72-403f-9698-da4536b6f37f tempest-ServerDiagnosticsV248Test-1822646244 tempest-ServerDiagnosticsV248Test-1822646244-project-member] Waiting for the task: (returnval){ [ 752.403610] env[61923]: value = "session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]523f2676-b013-cb87-0142-af01bb922c8a" [ 752.403610] env[61923]: _type = "Task" [ 752.403610] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 752.411180] env[61923]: DEBUG oslo_vmware.api [None req-f1ee1de3-0f72-403f-9698-da4536b6f37f tempest-ServerDiagnosticsV248Test-1822646244 tempest-ServerDiagnosticsV248Test-1822646244-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]523f2676-b013-cb87-0142-af01bb922c8a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 752.490871] env[61923]: DEBUG nova.scheduler.client.report [None req-f7b52f4c-b2e3-49f8-b056-b7254f72bb20 tempest-AttachInterfacesUnderV243Test-498638506 tempest-AttachInterfacesUnderV243Test-498638506-project-member] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 752.837349] env[61923]: DEBUG nova.network.neutron [None req-13202777-73e8-40d3-a6ad-7bb49e61267a tempest-ImagesNegativeTestJSON-1202128281 tempest-ImagesNegativeTestJSON-1202128281-project-member] [instance: f7e1d517-3e39-4381-83be-0152368b8120] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 752.914811] env[61923]: DEBUG oslo_vmware.api [None req-f1ee1de3-0f72-403f-9698-da4536b6f37f tempest-ServerDiagnosticsV248Test-1822646244 tempest-ServerDiagnosticsV248Test-1822646244-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]523f2676-b013-cb87-0142-af01bb922c8a, 'name': SearchDatastore_Task, 'duration_secs': 0.007801} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 752.915754] env[61923]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-256a9b42-cd44-403a-aeaf-661a368ae984 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.920914] env[61923]: DEBUG oslo_vmware.api [None req-f1ee1de3-0f72-403f-9698-da4536b6f37f tempest-ServerDiagnosticsV248Test-1822646244 tempest-ServerDiagnosticsV248Test-1822646244-project-member] Waiting for the task: (returnval){ [ 752.920914] env[61923]: value = "session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]5233f2a7-7fb8-e7e6-6e1f-7397d17d3ce1" [ 752.920914] env[61923]: _type = "Task" [ 752.920914] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 752.929047] env[61923]: DEBUG oslo_vmware.api [None req-f1ee1de3-0f72-403f-9698-da4536b6f37f tempest-ServerDiagnosticsV248Test-1822646244 tempest-ServerDiagnosticsV248Test-1822646244-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]5233f2a7-7fb8-e7e6-6e1f-7397d17d3ce1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 752.995648] env[61923]: DEBUG oslo_concurrency.lockutils [None req-f7b52f4c-b2e3-49f8-b056-b7254f72bb20 tempest-AttachInterfacesUnderV243Test-498638506 tempest-AttachInterfacesUnderV243Test-498638506-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.919s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 752.996361] env[61923]: ERROR nova.compute.manager [None req-f7b52f4c-b2e3-49f8-b056-b7254f72bb20 tempest-AttachInterfacesUnderV243Test-498638506 tempest-AttachInterfacesUnderV243Test-498638506-project-member] [instance: 0eab2200-2528-4a89-821a-cdd6a3c99297] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 4c6d06bd-7574-4546-acd9-9d362bc379b1, please check neutron logs for more information. [ 752.996361] env[61923]: ERROR nova.compute.manager [instance: 0eab2200-2528-4a89-821a-cdd6a3c99297] Traceback (most recent call last): [ 752.996361] env[61923]: ERROR nova.compute.manager [instance: 0eab2200-2528-4a89-821a-cdd6a3c99297] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 752.996361] env[61923]: ERROR nova.compute.manager [instance: 0eab2200-2528-4a89-821a-cdd6a3c99297] self.driver.spawn(context, instance, image_meta, [ 752.996361] env[61923]: ERROR nova.compute.manager [instance: 0eab2200-2528-4a89-821a-cdd6a3c99297] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 752.996361] env[61923]: ERROR nova.compute.manager [instance: 0eab2200-2528-4a89-821a-cdd6a3c99297] self._vmops.spawn(context, instance, image_meta, injected_files, [ 752.996361] env[61923]: ERROR nova.compute.manager [instance: 0eab2200-2528-4a89-821a-cdd6a3c99297] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 752.996361] env[61923]: ERROR nova.compute.manager [instance: 0eab2200-2528-4a89-821a-cdd6a3c99297] vm_ref = self.build_virtual_machine(instance, [ 752.996361] env[61923]: ERROR nova.compute.manager [instance: 0eab2200-2528-4a89-821a-cdd6a3c99297] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 752.996361] env[61923]: ERROR nova.compute.manager [instance: 0eab2200-2528-4a89-821a-cdd6a3c99297] vif_infos = vmwarevif.get_vif_info(self._session, [ 752.996361] env[61923]: ERROR nova.compute.manager [instance: 0eab2200-2528-4a89-821a-cdd6a3c99297] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 752.996773] env[61923]: ERROR nova.compute.manager [instance: 0eab2200-2528-4a89-821a-cdd6a3c99297] for vif in network_info: [ 752.996773] env[61923]: ERROR nova.compute.manager [instance: 0eab2200-2528-4a89-821a-cdd6a3c99297] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 752.996773] env[61923]: ERROR nova.compute.manager [instance: 0eab2200-2528-4a89-821a-cdd6a3c99297] return self._sync_wrapper(fn, *args, **kwargs) [ 752.996773] env[61923]: ERROR nova.compute.manager [instance: 0eab2200-2528-4a89-821a-cdd6a3c99297] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 752.996773] env[61923]: ERROR nova.compute.manager [instance: 0eab2200-2528-4a89-821a-cdd6a3c99297] self.wait() [ 752.996773] env[61923]: ERROR nova.compute.manager [instance: 0eab2200-2528-4a89-821a-cdd6a3c99297] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 752.996773] env[61923]: ERROR nova.compute.manager [instance: 0eab2200-2528-4a89-821a-cdd6a3c99297] self[:] = self._gt.wait() [ 752.996773] env[61923]: ERROR nova.compute.manager [instance: 0eab2200-2528-4a89-821a-cdd6a3c99297] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 752.996773] env[61923]: ERROR nova.compute.manager [instance: 0eab2200-2528-4a89-821a-cdd6a3c99297] return self._exit_event.wait() [ 752.996773] env[61923]: ERROR nova.compute.manager [instance: 0eab2200-2528-4a89-821a-cdd6a3c99297] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 752.996773] env[61923]: ERROR nova.compute.manager [instance: 0eab2200-2528-4a89-821a-cdd6a3c99297] result = hub.switch() [ 752.996773] env[61923]: ERROR nova.compute.manager [instance: 0eab2200-2528-4a89-821a-cdd6a3c99297] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 752.996773] env[61923]: ERROR nova.compute.manager [instance: 0eab2200-2528-4a89-821a-cdd6a3c99297] return self.greenlet.switch() [ 752.997196] env[61923]: ERROR nova.compute.manager [instance: 0eab2200-2528-4a89-821a-cdd6a3c99297] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 752.997196] env[61923]: ERROR nova.compute.manager [instance: 0eab2200-2528-4a89-821a-cdd6a3c99297] result = function(*args, **kwargs) [ 752.997196] env[61923]: ERROR nova.compute.manager [instance: 0eab2200-2528-4a89-821a-cdd6a3c99297] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 752.997196] env[61923]: ERROR nova.compute.manager [instance: 0eab2200-2528-4a89-821a-cdd6a3c99297] return func(*args, **kwargs) [ 752.997196] env[61923]: ERROR nova.compute.manager [instance: 0eab2200-2528-4a89-821a-cdd6a3c99297] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 752.997196] env[61923]: ERROR nova.compute.manager [instance: 0eab2200-2528-4a89-821a-cdd6a3c99297] raise e [ 752.997196] env[61923]: ERROR nova.compute.manager [instance: 0eab2200-2528-4a89-821a-cdd6a3c99297] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 752.997196] env[61923]: ERROR nova.compute.manager [instance: 0eab2200-2528-4a89-821a-cdd6a3c99297] nwinfo = self.network_api.allocate_for_instance( [ 752.997196] env[61923]: ERROR nova.compute.manager [instance: 0eab2200-2528-4a89-821a-cdd6a3c99297] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 752.997196] env[61923]: ERROR nova.compute.manager [instance: 0eab2200-2528-4a89-821a-cdd6a3c99297] created_port_ids = self._update_ports_for_instance( [ 752.997196] env[61923]: ERROR nova.compute.manager [instance: 0eab2200-2528-4a89-821a-cdd6a3c99297] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 752.997196] env[61923]: ERROR nova.compute.manager [instance: 0eab2200-2528-4a89-821a-cdd6a3c99297] with excutils.save_and_reraise_exception(): [ 752.997196] env[61923]: ERROR nova.compute.manager [instance: 0eab2200-2528-4a89-821a-cdd6a3c99297] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 752.997623] env[61923]: ERROR nova.compute.manager [instance: 0eab2200-2528-4a89-821a-cdd6a3c99297] self.force_reraise() [ 752.997623] env[61923]: ERROR nova.compute.manager [instance: 0eab2200-2528-4a89-821a-cdd6a3c99297] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 752.997623] env[61923]: ERROR nova.compute.manager [instance: 0eab2200-2528-4a89-821a-cdd6a3c99297] raise self.value [ 752.997623] env[61923]: ERROR nova.compute.manager [instance: 0eab2200-2528-4a89-821a-cdd6a3c99297] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 752.997623] env[61923]: ERROR nova.compute.manager [instance: 0eab2200-2528-4a89-821a-cdd6a3c99297] updated_port = self._update_port( [ 752.997623] env[61923]: ERROR nova.compute.manager [instance: 0eab2200-2528-4a89-821a-cdd6a3c99297] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 752.997623] env[61923]: ERROR nova.compute.manager [instance: 0eab2200-2528-4a89-821a-cdd6a3c99297] _ensure_no_port_binding_failure(port) [ 752.997623] env[61923]: ERROR nova.compute.manager [instance: 0eab2200-2528-4a89-821a-cdd6a3c99297] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 752.997623] env[61923]: ERROR nova.compute.manager [instance: 0eab2200-2528-4a89-821a-cdd6a3c99297] raise exception.PortBindingFailed(port_id=port['id']) [ 752.997623] env[61923]: ERROR nova.compute.manager [instance: 0eab2200-2528-4a89-821a-cdd6a3c99297] nova.exception.PortBindingFailed: Binding failed for port 4c6d06bd-7574-4546-acd9-9d362bc379b1, please check neutron logs for more information. [ 752.997623] env[61923]: ERROR nova.compute.manager [instance: 0eab2200-2528-4a89-821a-cdd6a3c99297] [ 752.997977] env[61923]: DEBUG nova.compute.utils [None req-f7b52f4c-b2e3-49f8-b056-b7254f72bb20 tempest-AttachInterfacesUnderV243Test-498638506 tempest-AttachInterfacesUnderV243Test-498638506-project-member] [instance: 0eab2200-2528-4a89-821a-cdd6a3c99297] Binding failed for port 4c6d06bd-7574-4546-acd9-9d362bc379b1, please check neutron logs for more information. {{(pid=61923) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 752.998380] env[61923]: DEBUG oslo_concurrency.lockutils [None req-4a9f7dc2-7221-4911-9bc8-b39b7a22ddda tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 13.216s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 753.002644] env[61923]: DEBUG nova.compute.manager [None req-f7b52f4c-b2e3-49f8-b056-b7254f72bb20 tempest-AttachInterfacesUnderV243Test-498638506 tempest-AttachInterfacesUnderV243Test-498638506-project-member] [instance: 0eab2200-2528-4a89-821a-cdd6a3c99297] Build of instance 0eab2200-2528-4a89-821a-cdd6a3c99297 was re-scheduled: Binding failed for port 4c6d06bd-7574-4546-acd9-9d362bc379b1, please check neutron logs for more information. {{(pid=61923) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 753.002644] env[61923]: DEBUG nova.compute.manager [None req-f7b52f4c-b2e3-49f8-b056-b7254f72bb20 tempest-AttachInterfacesUnderV243Test-498638506 tempest-AttachInterfacesUnderV243Test-498638506-project-member] [instance: 0eab2200-2528-4a89-821a-cdd6a3c99297] Unplugging VIFs for instance {{(pid=61923) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 753.002832] env[61923]: DEBUG oslo_concurrency.lockutils [None req-f7b52f4c-b2e3-49f8-b056-b7254f72bb20 tempest-AttachInterfacesUnderV243Test-498638506 tempest-AttachInterfacesUnderV243Test-498638506-project-member] Acquiring lock "refresh_cache-0eab2200-2528-4a89-821a-cdd6a3c99297" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 753.002988] env[61923]: DEBUG oslo_concurrency.lockutils [None req-f7b52f4c-b2e3-49f8-b056-b7254f72bb20 tempest-AttachInterfacesUnderV243Test-498638506 tempest-AttachInterfacesUnderV243Test-498638506-project-member] Acquired lock "refresh_cache-0eab2200-2528-4a89-821a-cdd6a3c99297" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 753.003167] env[61923]: DEBUG nova.network.neutron [None req-f7b52f4c-b2e3-49f8-b056-b7254f72bb20 tempest-AttachInterfacesUnderV243Test-498638506 tempest-AttachInterfacesUnderV243Test-498638506-project-member] [instance: 0eab2200-2528-4a89-821a-cdd6a3c99297] Building network info cache for instance {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 753.340320] env[61923]: INFO nova.compute.manager [None req-13202777-73e8-40d3-a6ad-7bb49e61267a tempest-ImagesNegativeTestJSON-1202128281 tempest-ImagesNegativeTestJSON-1202128281-project-member] [instance: f7e1d517-3e39-4381-83be-0152368b8120] Took 1.04 seconds to deallocate network for instance. [ 753.431810] env[61923]: DEBUG oslo_vmware.api [None req-f1ee1de3-0f72-403f-9698-da4536b6f37f tempest-ServerDiagnosticsV248Test-1822646244 tempest-ServerDiagnosticsV248Test-1822646244-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]5233f2a7-7fb8-e7e6-6e1f-7397d17d3ce1, 'name': SearchDatastore_Task, 'duration_secs': 0.009523} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 753.432087] env[61923]: DEBUG oslo_concurrency.lockutils [None req-f1ee1de3-0f72-403f-9698-da4536b6f37f tempest-ServerDiagnosticsV248Test-1822646244 tempest-ServerDiagnosticsV248Test-1822646244-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 753.432357] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-f1ee1de3-0f72-403f-9698-da4536b6f37f tempest-ServerDiagnosticsV248Test-1822646244 tempest-ServerDiagnosticsV248Test-1822646244-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk to [datastore1] 655b01ae-76a5-4c09-9245-e629fec8fd4b/655b01ae-76a5-4c09-9245-e629fec8fd4b.vmdk {{(pid=61923) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 753.432606] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-cb505a9a-555c-4bdf-8be0-e4102e5d1c8e {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.440496] env[61923]: DEBUG oslo_vmware.api [None req-f1ee1de3-0f72-403f-9698-da4536b6f37f tempest-ServerDiagnosticsV248Test-1822646244 tempest-ServerDiagnosticsV248Test-1822646244-project-member] Waiting for the task: (returnval){ [ 753.440496] env[61923]: value = "task-1377486" [ 753.440496] env[61923]: _type = "Task" [ 753.440496] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 753.448878] env[61923]: DEBUG oslo_vmware.api [None req-f1ee1de3-0f72-403f-9698-da4536b6f37f tempest-ServerDiagnosticsV248Test-1822646244 tempest-ServerDiagnosticsV248Test-1822646244-project-member] Task: {'id': task-1377486, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 753.523963] env[61923]: DEBUG nova.network.neutron [None req-f7b52f4c-b2e3-49f8-b056-b7254f72bb20 tempest-AttachInterfacesUnderV243Test-498638506 tempest-AttachInterfacesUnderV243Test-498638506-project-member] [instance: 0eab2200-2528-4a89-821a-cdd6a3c99297] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 753.607021] env[61923]: DEBUG nova.network.neutron [None req-f7b52f4c-b2e3-49f8-b056-b7254f72bb20 tempest-AttachInterfacesUnderV243Test-498638506 tempest-AttachInterfacesUnderV243Test-498638506-project-member] [instance: 0eab2200-2528-4a89-821a-cdd6a3c99297] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 753.813077] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bfc71739-e87c-4f85-bea9-0fd77aaa41a2 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.821049] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7323be25-8d0b-4716-b2bc-a1b3c36395a3 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.858658] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c6eb894-7909-49ba-90f2-599f1306d45f {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.867915] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf79e884-1e25-4dcd-8300-fec47b6c5986 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.883024] env[61923]: DEBUG nova.compute.provider_tree [None req-4a9f7dc2-7221-4911-9bc8-b39b7a22ddda tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 753.951741] env[61923]: DEBUG oslo_vmware.api [None req-f1ee1de3-0f72-403f-9698-da4536b6f37f tempest-ServerDiagnosticsV248Test-1822646244 tempest-ServerDiagnosticsV248Test-1822646244-project-member] Task: {'id': task-1377486, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 754.106986] env[61923]: DEBUG oslo_concurrency.lockutils [None req-f7b52f4c-b2e3-49f8-b056-b7254f72bb20 tempest-AttachInterfacesUnderV243Test-498638506 tempest-AttachInterfacesUnderV243Test-498638506-project-member] Releasing lock "refresh_cache-0eab2200-2528-4a89-821a-cdd6a3c99297" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 754.107316] env[61923]: DEBUG nova.compute.manager [None req-f7b52f4c-b2e3-49f8-b056-b7254f72bb20 tempest-AttachInterfacesUnderV243Test-498638506 tempest-AttachInterfacesUnderV243Test-498638506-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61923) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 754.107527] env[61923]: DEBUG nova.compute.manager [None req-f7b52f4c-b2e3-49f8-b056-b7254f72bb20 tempest-AttachInterfacesUnderV243Test-498638506 tempest-AttachInterfacesUnderV243Test-498638506-project-member] [instance: 0eab2200-2528-4a89-821a-cdd6a3c99297] Deallocating network for instance {{(pid=61923) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 754.107723] env[61923]: DEBUG nova.network.neutron [None req-f7b52f4c-b2e3-49f8-b056-b7254f72bb20 tempest-AttachInterfacesUnderV243Test-498638506 tempest-AttachInterfacesUnderV243Test-498638506-project-member] [instance: 0eab2200-2528-4a89-821a-cdd6a3c99297] deallocate_for_instance() {{(pid=61923) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 754.136013] env[61923]: DEBUG nova.network.neutron [None req-f7b52f4c-b2e3-49f8-b056-b7254f72bb20 tempest-AttachInterfacesUnderV243Test-498638506 tempest-AttachInterfacesUnderV243Test-498638506-project-member] [instance: 0eab2200-2528-4a89-821a-cdd6a3c99297] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 754.382846] env[61923]: INFO nova.scheduler.client.report [None req-13202777-73e8-40d3-a6ad-7bb49e61267a tempest-ImagesNegativeTestJSON-1202128281 tempest-ImagesNegativeTestJSON-1202128281-project-member] Deleted allocations for instance f7e1d517-3e39-4381-83be-0152368b8120 [ 754.392022] env[61923]: DEBUG nova.scheduler.client.report [None req-4a9f7dc2-7221-4911-9bc8-b39b7a22ddda tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 754.451743] env[61923]: DEBUG oslo_vmware.api [None req-f1ee1de3-0f72-403f-9698-da4536b6f37f tempest-ServerDiagnosticsV248Test-1822646244 tempest-ServerDiagnosticsV248Test-1822646244-project-member] Task: {'id': task-1377486, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.692643} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 754.451988] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-f1ee1de3-0f72-403f-9698-da4536b6f37f tempest-ServerDiagnosticsV248Test-1822646244 tempest-ServerDiagnosticsV248Test-1822646244-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk to [datastore1] 655b01ae-76a5-4c09-9245-e629fec8fd4b/655b01ae-76a5-4c09-9245-e629fec8fd4b.vmdk {{(pid=61923) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 754.452701] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-f1ee1de3-0f72-403f-9698-da4536b6f37f tempest-ServerDiagnosticsV248Test-1822646244 tempest-ServerDiagnosticsV248Test-1822646244-project-member] [instance: 655b01ae-76a5-4c09-9245-e629fec8fd4b] Extending root virtual disk to 1048576 {{(pid=61923) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 754.452955] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-4b873626-2877-4453-a906-742734eb588b {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.460125] env[61923]: DEBUG oslo_vmware.api [None req-f1ee1de3-0f72-403f-9698-da4536b6f37f tempest-ServerDiagnosticsV248Test-1822646244 tempest-ServerDiagnosticsV248Test-1822646244-project-member] Waiting for the task: (returnval){ [ 754.460125] env[61923]: value = "task-1377487" [ 754.460125] env[61923]: _type = "Task" [ 754.460125] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 754.468720] env[61923]: DEBUG oslo_vmware.api [None req-f1ee1de3-0f72-403f-9698-da4536b6f37f tempest-ServerDiagnosticsV248Test-1822646244 tempest-ServerDiagnosticsV248Test-1822646244-project-member] Task: {'id': task-1377487, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 754.638729] env[61923]: DEBUG nova.network.neutron [None req-f7b52f4c-b2e3-49f8-b056-b7254f72bb20 tempest-AttachInterfacesUnderV243Test-498638506 tempest-AttachInterfacesUnderV243Test-498638506-project-member] [instance: 0eab2200-2528-4a89-821a-cdd6a3c99297] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 754.893924] env[61923]: DEBUG oslo_concurrency.lockutils [None req-4a9f7dc2-7221-4911-9bc8-b39b7a22ddda tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.895s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 754.894577] env[61923]: ERROR nova.compute.manager [None req-4a9f7dc2-7221-4911-9bc8-b39b7a22ddda tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] [instance: 22d1e8ed-75c5-42e9-ae55-12d59e565914] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 32585aee-c6bc-4a95-9516-646bb26bf0a0, please check neutron logs for more information. [ 754.894577] env[61923]: ERROR nova.compute.manager [instance: 22d1e8ed-75c5-42e9-ae55-12d59e565914] Traceback (most recent call last): [ 754.894577] env[61923]: ERROR nova.compute.manager [instance: 22d1e8ed-75c5-42e9-ae55-12d59e565914] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 754.894577] env[61923]: ERROR nova.compute.manager [instance: 22d1e8ed-75c5-42e9-ae55-12d59e565914] self.driver.spawn(context, instance, image_meta, [ 754.894577] env[61923]: ERROR nova.compute.manager [instance: 22d1e8ed-75c5-42e9-ae55-12d59e565914] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 754.894577] env[61923]: ERROR nova.compute.manager [instance: 22d1e8ed-75c5-42e9-ae55-12d59e565914] self._vmops.spawn(context, instance, image_meta, injected_files, [ 754.894577] env[61923]: ERROR nova.compute.manager [instance: 22d1e8ed-75c5-42e9-ae55-12d59e565914] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 754.894577] env[61923]: ERROR nova.compute.manager [instance: 22d1e8ed-75c5-42e9-ae55-12d59e565914] vm_ref = self.build_virtual_machine(instance, [ 754.894577] env[61923]: ERROR nova.compute.manager [instance: 22d1e8ed-75c5-42e9-ae55-12d59e565914] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 754.894577] env[61923]: ERROR nova.compute.manager [instance: 22d1e8ed-75c5-42e9-ae55-12d59e565914] vif_infos = vmwarevif.get_vif_info(self._session, [ 754.894577] env[61923]: ERROR nova.compute.manager [instance: 22d1e8ed-75c5-42e9-ae55-12d59e565914] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 754.895089] env[61923]: ERROR nova.compute.manager [instance: 22d1e8ed-75c5-42e9-ae55-12d59e565914] for vif in network_info: [ 754.895089] env[61923]: ERROR nova.compute.manager [instance: 22d1e8ed-75c5-42e9-ae55-12d59e565914] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 754.895089] env[61923]: ERROR nova.compute.manager [instance: 22d1e8ed-75c5-42e9-ae55-12d59e565914] return self._sync_wrapper(fn, *args, **kwargs) [ 754.895089] env[61923]: ERROR nova.compute.manager [instance: 22d1e8ed-75c5-42e9-ae55-12d59e565914] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 754.895089] env[61923]: ERROR nova.compute.manager [instance: 22d1e8ed-75c5-42e9-ae55-12d59e565914] self.wait() [ 754.895089] env[61923]: ERROR nova.compute.manager [instance: 22d1e8ed-75c5-42e9-ae55-12d59e565914] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 754.895089] env[61923]: ERROR nova.compute.manager [instance: 22d1e8ed-75c5-42e9-ae55-12d59e565914] self[:] = self._gt.wait() [ 754.895089] env[61923]: ERROR nova.compute.manager [instance: 22d1e8ed-75c5-42e9-ae55-12d59e565914] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 754.895089] env[61923]: ERROR nova.compute.manager [instance: 22d1e8ed-75c5-42e9-ae55-12d59e565914] return self._exit_event.wait() [ 754.895089] env[61923]: ERROR nova.compute.manager [instance: 22d1e8ed-75c5-42e9-ae55-12d59e565914] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 754.895089] env[61923]: ERROR nova.compute.manager [instance: 22d1e8ed-75c5-42e9-ae55-12d59e565914] result = hub.switch() [ 754.895089] env[61923]: ERROR nova.compute.manager [instance: 22d1e8ed-75c5-42e9-ae55-12d59e565914] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 754.895089] env[61923]: ERROR nova.compute.manager [instance: 22d1e8ed-75c5-42e9-ae55-12d59e565914] return self.greenlet.switch() [ 754.895602] env[61923]: ERROR nova.compute.manager [instance: 22d1e8ed-75c5-42e9-ae55-12d59e565914] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 754.895602] env[61923]: ERROR nova.compute.manager [instance: 22d1e8ed-75c5-42e9-ae55-12d59e565914] result = function(*args, **kwargs) [ 754.895602] env[61923]: ERROR nova.compute.manager [instance: 22d1e8ed-75c5-42e9-ae55-12d59e565914] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 754.895602] env[61923]: ERROR nova.compute.manager [instance: 22d1e8ed-75c5-42e9-ae55-12d59e565914] return func(*args, **kwargs) [ 754.895602] env[61923]: ERROR nova.compute.manager [instance: 22d1e8ed-75c5-42e9-ae55-12d59e565914] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 754.895602] env[61923]: ERROR nova.compute.manager [instance: 22d1e8ed-75c5-42e9-ae55-12d59e565914] raise e [ 754.895602] env[61923]: ERROR nova.compute.manager [instance: 22d1e8ed-75c5-42e9-ae55-12d59e565914] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 754.895602] env[61923]: ERROR nova.compute.manager [instance: 22d1e8ed-75c5-42e9-ae55-12d59e565914] nwinfo = self.network_api.allocate_for_instance( [ 754.895602] env[61923]: ERROR nova.compute.manager [instance: 22d1e8ed-75c5-42e9-ae55-12d59e565914] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 754.895602] env[61923]: ERROR nova.compute.manager [instance: 22d1e8ed-75c5-42e9-ae55-12d59e565914] created_port_ids = self._update_ports_for_instance( [ 754.895602] env[61923]: ERROR nova.compute.manager [instance: 22d1e8ed-75c5-42e9-ae55-12d59e565914] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 754.895602] env[61923]: ERROR nova.compute.manager [instance: 22d1e8ed-75c5-42e9-ae55-12d59e565914] with excutils.save_and_reraise_exception(): [ 754.895602] env[61923]: ERROR nova.compute.manager [instance: 22d1e8ed-75c5-42e9-ae55-12d59e565914] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 754.896145] env[61923]: ERROR nova.compute.manager [instance: 22d1e8ed-75c5-42e9-ae55-12d59e565914] self.force_reraise() [ 754.896145] env[61923]: ERROR nova.compute.manager [instance: 22d1e8ed-75c5-42e9-ae55-12d59e565914] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 754.896145] env[61923]: ERROR nova.compute.manager [instance: 22d1e8ed-75c5-42e9-ae55-12d59e565914] raise self.value [ 754.896145] env[61923]: ERROR nova.compute.manager [instance: 22d1e8ed-75c5-42e9-ae55-12d59e565914] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 754.896145] env[61923]: ERROR nova.compute.manager [instance: 22d1e8ed-75c5-42e9-ae55-12d59e565914] updated_port = self._update_port( [ 754.896145] env[61923]: ERROR nova.compute.manager [instance: 22d1e8ed-75c5-42e9-ae55-12d59e565914] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 754.896145] env[61923]: ERROR nova.compute.manager [instance: 22d1e8ed-75c5-42e9-ae55-12d59e565914] _ensure_no_port_binding_failure(port) [ 754.896145] env[61923]: ERROR nova.compute.manager [instance: 22d1e8ed-75c5-42e9-ae55-12d59e565914] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 754.896145] env[61923]: ERROR nova.compute.manager [instance: 22d1e8ed-75c5-42e9-ae55-12d59e565914] raise exception.PortBindingFailed(port_id=port['id']) [ 754.896145] env[61923]: ERROR nova.compute.manager [instance: 22d1e8ed-75c5-42e9-ae55-12d59e565914] nova.exception.PortBindingFailed: Binding failed for port 32585aee-c6bc-4a95-9516-646bb26bf0a0, please check neutron logs for more information. [ 754.896145] env[61923]: ERROR nova.compute.manager [instance: 22d1e8ed-75c5-42e9-ae55-12d59e565914] [ 754.896616] env[61923]: DEBUG nova.compute.utils [None req-4a9f7dc2-7221-4911-9bc8-b39b7a22ddda tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] [instance: 22d1e8ed-75c5-42e9-ae55-12d59e565914] Binding failed for port 32585aee-c6bc-4a95-9516-646bb26bf0a0, please check neutron logs for more information. {{(pid=61923) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 754.897521] env[61923]: DEBUG oslo_concurrency.lockutils [None req-13202777-73e8-40d3-a6ad-7bb49e61267a tempest-ImagesNegativeTestJSON-1202128281 tempest-ImagesNegativeTestJSON-1202128281-project-member] Lock "f7e1d517-3e39-4381-83be-0152368b8120" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 144.009s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 754.898151] env[61923]: DEBUG nova.compute.manager [None req-4a9f7dc2-7221-4911-9bc8-b39b7a22ddda tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] [instance: 22d1e8ed-75c5-42e9-ae55-12d59e565914] Build of instance 22d1e8ed-75c5-42e9-ae55-12d59e565914 was re-scheduled: Binding failed for port 32585aee-c6bc-4a95-9516-646bb26bf0a0, please check neutron logs for more information. {{(pid=61923) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 754.898641] env[61923]: DEBUG nova.compute.manager [None req-4a9f7dc2-7221-4911-9bc8-b39b7a22ddda tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] [instance: 22d1e8ed-75c5-42e9-ae55-12d59e565914] Unplugging VIFs for instance {{(pid=61923) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 754.898871] env[61923]: DEBUG oslo_concurrency.lockutils [None req-4a9f7dc2-7221-4911-9bc8-b39b7a22ddda tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Acquiring lock "refresh_cache-22d1e8ed-75c5-42e9-ae55-12d59e565914" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 754.899024] env[61923]: DEBUG oslo_concurrency.lockutils [None req-4a9f7dc2-7221-4911-9bc8-b39b7a22ddda tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Acquired lock "refresh_cache-22d1e8ed-75c5-42e9-ae55-12d59e565914" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 754.899287] env[61923]: DEBUG nova.network.neutron [None req-4a9f7dc2-7221-4911-9bc8-b39b7a22ddda tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] [instance: 22d1e8ed-75c5-42e9-ae55-12d59e565914] Building network info cache for instance {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 754.900339] env[61923]: DEBUG oslo_concurrency.lockutils [None req-5bb1ca17-5442-4564-86ab-53f866a97e1d tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.007s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 754.970170] env[61923]: DEBUG oslo_vmware.api [None req-f1ee1de3-0f72-403f-9698-da4536b6f37f tempest-ServerDiagnosticsV248Test-1822646244 tempest-ServerDiagnosticsV248Test-1822646244-project-member] Task: {'id': task-1377487, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.057908} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 754.971093] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-f1ee1de3-0f72-403f-9698-da4536b6f37f tempest-ServerDiagnosticsV248Test-1822646244 tempest-ServerDiagnosticsV248Test-1822646244-project-member] [instance: 655b01ae-76a5-4c09-9245-e629fec8fd4b] Extended root virtual disk {{(pid=61923) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 754.971976] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-455982be-09b2-4c90-96b8-5da31fbe917f {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.996220] env[61923]: DEBUG nova.virt.vmwareapi.volumeops [None req-f1ee1de3-0f72-403f-9698-da4536b6f37f tempest-ServerDiagnosticsV248Test-1822646244 tempest-ServerDiagnosticsV248Test-1822646244-project-member] [instance: 655b01ae-76a5-4c09-9245-e629fec8fd4b] Reconfiguring VM instance instance-00000033 to attach disk [datastore1] 655b01ae-76a5-4c09-9245-e629fec8fd4b/655b01ae-76a5-4c09-9245-e629fec8fd4b.vmdk or device None with type sparse {{(pid=61923) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 754.996940] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ea876f0c-53a3-40d3-bb85-9ee72a5bd087 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.018121] env[61923]: DEBUG oslo_vmware.api [None req-f1ee1de3-0f72-403f-9698-da4536b6f37f tempest-ServerDiagnosticsV248Test-1822646244 tempest-ServerDiagnosticsV248Test-1822646244-project-member] Waiting for the task: (returnval){ [ 755.018121] env[61923]: value = "task-1377488" [ 755.018121] env[61923]: _type = "Task" [ 755.018121] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 755.033019] env[61923]: DEBUG oslo_vmware.api [None req-f1ee1de3-0f72-403f-9698-da4536b6f37f tempest-ServerDiagnosticsV248Test-1822646244 tempest-ServerDiagnosticsV248Test-1822646244-project-member] Task: {'id': task-1377488, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 755.142046] env[61923]: INFO nova.compute.manager [None req-f7b52f4c-b2e3-49f8-b056-b7254f72bb20 tempest-AttachInterfacesUnderV243Test-498638506 tempest-AttachInterfacesUnderV243Test-498638506-project-member] [instance: 0eab2200-2528-4a89-821a-cdd6a3c99297] Took 1.03 seconds to deallocate network for instance. [ 755.411278] env[61923]: DEBUG nova.compute.manager [None req-1d78b71c-3449-4286-a17a-dc5396f589fe tempest-ServersAdminTestJSON-678434436 tempest-ServersAdminTestJSON-678434436-project-member] [instance: faff83a1-3532-4396-ac3f-600f86653749] Starting instance... {{(pid=61923) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 755.434996] env[61923]: DEBUG nova.network.neutron [None req-4a9f7dc2-7221-4911-9bc8-b39b7a22ddda tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] [instance: 22d1e8ed-75c5-42e9-ae55-12d59e565914] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 755.534050] env[61923]: DEBUG oslo_vmware.api [None req-f1ee1de3-0f72-403f-9698-da4536b6f37f tempest-ServerDiagnosticsV248Test-1822646244 tempest-ServerDiagnosticsV248Test-1822646244-project-member] Task: {'id': task-1377488, 'name': ReconfigVM_Task, 'duration_secs': 0.319812} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 755.534334] env[61923]: DEBUG nova.virt.vmwareapi.volumeops [None req-f1ee1de3-0f72-403f-9698-da4536b6f37f tempest-ServerDiagnosticsV248Test-1822646244 tempest-ServerDiagnosticsV248Test-1822646244-project-member] [instance: 655b01ae-76a5-4c09-9245-e629fec8fd4b] Reconfigured VM instance instance-00000033 to attach disk [datastore1] 655b01ae-76a5-4c09-9245-e629fec8fd4b/655b01ae-76a5-4c09-9245-e629fec8fd4b.vmdk or device None with type sparse {{(pid=61923) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 755.534933] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-10308329-6486-4613-86a0-f4e8dda83cc4 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.542338] env[61923]: DEBUG oslo_vmware.api [None req-f1ee1de3-0f72-403f-9698-da4536b6f37f tempest-ServerDiagnosticsV248Test-1822646244 tempest-ServerDiagnosticsV248Test-1822646244-project-member] Waiting for the task: (returnval){ [ 755.542338] env[61923]: value = "task-1377489" [ 755.542338] env[61923]: _type = "Task" [ 755.542338] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 755.559347] env[61923]: DEBUG oslo_vmware.api [None req-f1ee1de3-0f72-403f-9698-da4536b6f37f tempest-ServerDiagnosticsV248Test-1822646244 tempest-ServerDiagnosticsV248Test-1822646244-project-member] Task: {'id': task-1377489, 'name': Rename_Task} progress is 6%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 755.586531] env[61923]: DEBUG nova.network.neutron [None req-4a9f7dc2-7221-4911-9bc8-b39b7a22ddda tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] [instance: 22d1e8ed-75c5-42e9-ae55-12d59e565914] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 755.776721] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63584b8f-ae20-404a-9929-74cb4db3c225 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.784739] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c449d22-d29a-4f18-be08-b4395513919d {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.820458] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b3f7fa0-ead7-4574-b2ad-ab859dff4364 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.828358] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23c131fa-f5d1-4fe9-a08c-e94441acbe76 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.843941] env[61923]: DEBUG nova.compute.provider_tree [None req-5bb1ca17-5442-4564-86ab-53f866a97e1d tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 755.933181] env[61923]: DEBUG oslo_concurrency.lockutils [None req-1d78b71c-3449-4286-a17a-dc5396f589fe tempest-ServersAdminTestJSON-678434436 tempest-ServersAdminTestJSON-678434436-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 756.056487] env[61923]: DEBUG oslo_vmware.api [None req-f1ee1de3-0f72-403f-9698-da4536b6f37f tempest-ServerDiagnosticsV248Test-1822646244 tempest-ServerDiagnosticsV248Test-1822646244-project-member] Task: {'id': task-1377489, 'name': Rename_Task, 'duration_secs': 0.134591} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 756.056765] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-f1ee1de3-0f72-403f-9698-da4536b6f37f tempest-ServerDiagnosticsV248Test-1822646244 tempest-ServerDiagnosticsV248Test-1822646244-project-member] [instance: 655b01ae-76a5-4c09-9245-e629fec8fd4b] Powering on the VM {{(pid=61923) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 756.057018] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6b0bab88-41ed-403a-afca-5550e0ca70fa {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.067607] env[61923]: DEBUG oslo_vmware.api [None req-f1ee1de3-0f72-403f-9698-da4536b6f37f tempest-ServerDiagnosticsV248Test-1822646244 tempest-ServerDiagnosticsV248Test-1822646244-project-member] Waiting for the task: (returnval){ [ 756.067607] env[61923]: value = "task-1377490" [ 756.067607] env[61923]: _type = "Task" [ 756.067607] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 756.075829] env[61923]: DEBUG oslo_vmware.api [None req-f1ee1de3-0f72-403f-9698-da4536b6f37f tempest-ServerDiagnosticsV248Test-1822646244 tempest-ServerDiagnosticsV248Test-1822646244-project-member] Task: {'id': task-1377490, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 756.092572] env[61923]: DEBUG oslo_concurrency.lockutils [None req-4a9f7dc2-7221-4911-9bc8-b39b7a22ddda tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Releasing lock "refresh_cache-22d1e8ed-75c5-42e9-ae55-12d59e565914" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 756.092572] env[61923]: DEBUG nova.compute.manager [None req-4a9f7dc2-7221-4911-9bc8-b39b7a22ddda tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61923) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 756.092572] env[61923]: DEBUG nova.compute.manager [None req-4a9f7dc2-7221-4911-9bc8-b39b7a22ddda tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] [instance: 22d1e8ed-75c5-42e9-ae55-12d59e565914] Deallocating network for instance {{(pid=61923) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 756.092572] env[61923]: DEBUG nova.network.neutron [None req-4a9f7dc2-7221-4911-9bc8-b39b7a22ddda tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] [instance: 22d1e8ed-75c5-42e9-ae55-12d59e565914] deallocate_for_instance() {{(pid=61923) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 756.120599] env[61923]: DEBUG nova.network.neutron [None req-4a9f7dc2-7221-4911-9bc8-b39b7a22ddda tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] [instance: 22d1e8ed-75c5-42e9-ae55-12d59e565914] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 756.181650] env[61923]: INFO nova.scheduler.client.report [None req-f7b52f4c-b2e3-49f8-b056-b7254f72bb20 tempest-AttachInterfacesUnderV243Test-498638506 tempest-AttachInterfacesUnderV243Test-498638506-project-member] Deleted allocations for instance 0eab2200-2528-4a89-821a-cdd6a3c99297 [ 756.304168] env[61923]: DEBUG oslo_concurrency.lockutils [None req-9ff8b121-368d-4886-9882-962e7ef81bcf tempest-ServerTagsTestJSON-19215696 tempest-ServerTagsTestJSON-19215696-project-member] Acquiring lock "48825592-dcdc-41d5-9fbf-500d1f31017b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 756.304499] env[61923]: DEBUG oslo_concurrency.lockutils [None req-9ff8b121-368d-4886-9882-962e7ef81bcf tempest-ServerTagsTestJSON-19215696 tempest-ServerTagsTestJSON-19215696-project-member] Lock "48825592-dcdc-41d5-9fbf-500d1f31017b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 756.347463] env[61923]: DEBUG nova.scheduler.client.report [None req-5bb1ca17-5442-4564-86ab-53f866a97e1d tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 756.577379] env[61923]: DEBUG oslo_vmware.api [None req-f1ee1de3-0f72-403f-9698-da4536b6f37f tempest-ServerDiagnosticsV248Test-1822646244 tempest-ServerDiagnosticsV248Test-1822646244-project-member] Task: {'id': task-1377490, 'name': PowerOnVM_Task, 'duration_secs': 0.47734} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 756.577681] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-f1ee1de3-0f72-403f-9698-da4536b6f37f tempest-ServerDiagnosticsV248Test-1822646244 tempest-ServerDiagnosticsV248Test-1822646244-project-member] [instance: 655b01ae-76a5-4c09-9245-e629fec8fd4b] Powered on the VM {{(pid=61923) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 756.577839] env[61923]: INFO nova.compute.manager [None req-f1ee1de3-0f72-403f-9698-da4536b6f37f tempest-ServerDiagnosticsV248Test-1822646244 tempest-ServerDiagnosticsV248Test-1822646244-project-member] [instance: 655b01ae-76a5-4c09-9245-e629fec8fd4b] Took 5.32 seconds to spawn the instance on the hypervisor. [ 756.578028] env[61923]: DEBUG nova.compute.manager [None req-f1ee1de3-0f72-403f-9698-da4536b6f37f tempest-ServerDiagnosticsV248Test-1822646244 tempest-ServerDiagnosticsV248Test-1822646244-project-member] [instance: 655b01ae-76a5-4c09-9245-e629fec8fd4b] Checking state {{(pid=61923) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 756.578850] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c3e4782-80a1-4b29-9b13-745e5f166403 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.626091] env[61923]: DEBUG nova.network.neutron [None req-4a9f7dc2-7221-4911-9bc8-b39b7a22ddda tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] [instance: 22d1e8ed-75c5-42e9-ae55-12d59e565914] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 756.694872] env[61923]: DEBUG oslo_concurrency.lockutils [None req-f7b52f4c-b2e3-49f8-b056-b7254f72bb20 tempest-AttachInterfacesUnderV243Test-498638506 tempest-AttachInterfacesUnderV243Test-498638506-project-member] Lock "0eab2200-2528-4a89-821a-cdd6a3c99297" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 141.733s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 756.852598] env[61923]: DEBUG oslo_concurrency.lockutils [None req-5bb1ca17-5442-4564-86ab-53f866a97e1d tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.952s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 756.853274] env[61923]: ERROR nova.compute.manager [None req-5bb1ca17-5442-4564-86ab-53f866a97e1d tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: 15eaaa7c-77be-4b8b-89d0-0b2b11a0c49e] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 439b85a4-1206-4dc7-9225-473a7f7704d9, please check neutron logs for more information. [ 756.853274] env[61923]: ERROR nova.compute.manager [instance: 15eaaa7c-77be-4b8b-89d0-0b2b11a0c49e] Traceback (most recent call last): [ 756.853274] env[61923]: ERROR nova.compute.manager [instance: 15eaaa7c-77be-4b8b-89d0-0b2b11a0c49e] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 756.853274] env[61923]: ERROR nova.compute.manager [instance: 15eaaa7c-77be-4b8b-89d0-0b2b11a0c49e] self.driver.spawn(context, instance, image_meta, [ 756.853274] env[61923]: ERROR nova.compute.manager [instance: 15eaaa7c-77be-4b8b-89d0-0b2b11a0c49e] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 756.853274] env[61923]: ERROR nova.compute.manager [instance: 15eaaa7c-77be-4b8b-89d0-0b2b11a0c49e] self._vmops.spawn(context, instance, image_meta, injected_files, [ 756.853274] env[61923]: ERROR nova.compute.manager [instance: 15eaaa7c-77be-4b8b-89d0-0b2b11a0c49e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 756.853274] env[61923]: ERROR nova.compute.manager [instance: 15eaaa7c-77be-4b8b-89d0-0b2b11a0c49e] vm_ref = self.build_virtual_machine(instance, [ 756.853274] env[61923]: ERROR nova.compute.manager [instance: 15eaaa7c-77be-4b8b-89d0-0b2b11a0c49e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 756.853274] env[61923]: ERROR nova.compute.manager [instance: 15eaaa7c-77be-4b8b-89d0-0b2b11a0c49e] vif_infos = vmwarevif.get_vif_info(self._session, [ 756.853274] env[61923]: ERROR nova.compute.manager [instance: 15eaaa7c-77be-4b8b-89d0-0b2b11a0c49e] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 756.853652] env[61923]: ERROR nova.compute.manager [instance: 15eaaa7c-77be-4b8b-89d0-0b2b11a0c49e] for vif in network_info: [ 756.853652] env[61923]: ERROR nova.compute.manager [instance: 15eaaa7c-77be-4b8b-89d0-0b2b11a0c49e] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 756.853652] env[61923]: ERROR nova.compute.manager [instance: 15eaaa7c-77be-4b8b-89d0-0b2b11a0c49e] return self._sync_wrapper(fn, *args, **kwargs) [ 756.853652] env[61923]: ERROR nova.compute.manager [instance: 15eaaa7c-77be-4b8b-89d0-0b2b11a0c49e] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 756.853652] env[61923]: ERROR nova.compute.manager [instance: 15eaaa7c-77be-4b8b-89d0-0b2b11a0c49e] self.wait() [ 756.853652] env[61923]: ERROR nova.compute.manager [instance: 15eaaa7c-77be-4b8b-89d0-0b2b11a0c49e] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 756.853652] env[61923]: ERROR nova.compute.manager [instance: 15eaaa7c-77be-4b8b-89d0-0b2b11a0c49e] self[:] = self._gt.wait() [ 756.853652] env[61923]: ERROR nova.compute.manager [instance: 15eaaa7c-77be-4b8b-89d0-0b2b11a0c49e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 756.853652] env[61923]: ERROR nova.compute.manager [instance: 15eaaa7c-77be-4b8b-89d0-0b2b11a0c49e] return self._exit_event.wait() [ 756.853652] env[61923]: ERROR nova.compute.manager [instance: 15eaaa7c-77be-4b8b-89d0-0b2b11a0c49e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 756.853652] env[61923]: ERROR nova.compute.manager [instance: 15eaaa7c-77be-4b8b-89d0-0b2b11a0c49e] result = hub.switch() [ 756.853652] env[61923]: ERROR nova.compute.manager [instance: 15eaaa7c-77be-4b8b-89d0-0b2b11a0c49e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 756.853652] env[61923]: ERROR nova.compute.manager [instance: 15eaaa7c-77be-4b8b-89d0-0b2b11a0c49e] return self.greenlet.switch() [ 756.854032] env[61923]: ERROR nova.compute.manager [instance: 15eaaa7c-77be-4b8b-89d0-0b2b11a0c49e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 756.854032] env[61923]: ERROR nova.compute.manager [instance: 15eaaa7c-77be-4b8b-89d0-0b2b11a0c49e] result = function(*args, **kwargs) [ 756.854032] env[61923]: ERROR nova.compute.manager [instance: 15eaaa7c-77be-4b8b-89d0-0b2b11a0c49e] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 756.854032] env[61923]: ERROR nova.compute.manager [instance: 15eaaa7c-77be-4b8b-89d0-0b2b11a0c49e] return func(*args, **kwargs) [ 756.854032] env[61923]: ERROR nova.compute.manager [instance: 15eaaa7c-77be-4b8b-89d0-0b2b11a0c49e] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 756.854032] env[61923]: ERROR nova.compute.manager [instance: 15eaaa7c-77be-4b8b-89d0-0b2b11a0c49e] raise e [ 756.854032] env[61923]: ERROR nova.compute.manager [instance: 15eaaa7c-77be-4b8b-89d0-0b2b11a0c49e] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 756.854032] env[61923]: ERROR nova.compute.manager [instance: 15eaaa7c-77be-4b8b-89d0-0b2b11a0c49e] nwinfo = self.network_api.allocate_for_instance( [ 756.854032] env[61923]: ERROR nova.compute.manager [instance: 15eaaa7c-77be-4b8b-89d0-0b2b11a0c49e] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 756.854032] env[61923]: ERROR nova.compute.manager [instance: 15eaaa7c-77be-4b8b-89d0-0b2b11a0c49e] created_port_ids = self._update_ports_for_instance( [ 756.854032] env[61923]: ERROR nova.compute.manager [instance: 15eaaa7c-77be-4b8b-89d0-0b2b11a0c49e] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 756.854032] env[61923]: ERROR nova.compute.manager [instance: 15eaaa7c-77be-4b8b-89d0-0b2b11a0c49e] with excutils.save_and_reraise_exception(): [ 756.854032] env[61923]: ERROR nova.compute.manager [instance: 15eaaa7c-77be-4b8b-89d0-0b2b11a0c49e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 756.854415] env[61923]: ERROR nova.compute.manager [instance: 15eaaa7c-77be-4b8b-89d0-0b2b11a0c49e] self.force_reraise() [ 756.854415] env[61923]: ERROR nova.compute.manager [instance: 15eaaa7c-77be-4b8b-89d0-0b2b11a0c49e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 756.854415] env[61923]: ERROR nova.compute.manager [instance: 15eaaa7c-77be-4b8b-89d0-0b2b11a0c49e] raise self.value [ 756.854415] env[61923]: ERROR nova.compute.manager [instance: 15eaaa7c-77be-4b8b-89d0-0b2b11a0c49e] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 756.854415] env[61923]: ERROR nova.compute.manager [instance: 15eaaa7c-77be-4b8b-89d0-0b2b11a0c49e] updated_port = self._update_port( [ 756.854415] env[61923]: ERROR nova.compute.manager [instance: 15eaaa7c-77be-4b8b-89d0-0b2b11a0c49e] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 756.854415] env[61923]: ERROR nova.compute.manager [instance: 15eaaa7c-77be-4b8b-89d0-0b2b11a0c49e] _ensure_no_port_binding_failure(port) [ 756.854415] env[61923]: ERROR nova.compute.manager [instance: 15eaaa7c-77be-4b8b-89d0-0b2b11a0c49e] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 756.854415] env[61923]: ERROR nova.compute.manager [instance: 15eaaa7c-77be-4b8b-89d0-0b2b11a0c49e] raise exception.PortBindingFailed(port_id=port['id']) [ 756.854415] env[61923]: ERROR nova.compute.manager [instance: 15eaaa7c-77be-4b8b-89d0-0b2b11a0c49e] nova.exception.PortBindingFailed: Binding failed for port 439b85a4-1206-4dc7-9225-473a7f7704d9, please check neutron logs for more information. [ 756.854415] env[61923]: ERROR nova.compute.manager [instance: 15eaaa7c-77be-4b8b-89d0-0b2b11a0c49e] [ 756.854802] env[61923]: DEBUG nova.compute.utils [None req-5bb1ca17-5442-4564-86ab-53f866a97e1d tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: 15eaaa7c-77be-4b8b-89d0-0b2b11a0c49e] Binding failed for port 439b85a4-1206-4dc7-9225-473a7f7704d9, please check neutron logs for more information. {{(pid=61923) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 756.855465] env[61923]: DEBUG oslo_concurrency.lockutils [None req-347b83bf-b4ed-4904-9bd3-d6021183ae1c tempest-ServerShowV254Test-921228709 tempest-ServerShowV254Test-921228709-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.486s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 756.857964] env[61923]: INFO nova.compute.claims [None req-347b83bf-b4ed-4904-9bd3-d6021183ae1c tempest-ServerShowV254Test-921228709 tempest-ServerShowV254Test-921228709-project-member] [instance: ecedb77f-4a2a-4047-964f-80796954d847] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 756.861555] env[61923]: DEBUG nova.compute.manager [None req-5bb1ca17-5442-4564-86ab-53f866a97e1d tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: 15eaaa7c-77be-4b8b-89d0-0b2b11a0c49e] Build of instance 15eaaa7c-77be-4b8b-89d0-0b2b11a0c49e was re-scheduled: Binding failed for port 439b85a4-1206-4dc7-9225-473a7f7704d9, please check neutron logs for more information. {{(pid=61923) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 756.862008] env[61923]: DEBUG nova.compute.manager [None req-5bb1ca17-5442-4564-86ab-53f866a97e1d tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: 15eaaa7c-77be-4b8b-89d0-0b2b11a0c49e] Unplugging VIFs for instance {{(pid=61923) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 756.862238] env[61923]: DEBUG oslo_concurrency.lockutils [None req-5bb1ca17-5442-4564-86ab-53f866a97e1d tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Acquiring lock "refresh_cache-15eaaa7c-77be-4b8b-89d0-0b2b11a0c49e" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 756.862383] env[61923]: DEBUG oslo_concurrency.lockutils [None req-5bb1ca17-5442-4564-86ab-53f866a97e1d tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Acquired lock "refresh_cache-15eaaa7c-77be-4b8b-89d0-0b2b11a0c49e" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 756.862539] env[61923]: DEBUG nova.network.neutron [None req-5bb1ca17-5442-4564-86ab-53f866a97e1d tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: 15eaaa7c-77be-4b8b-89d0-0b2b11a0c49e] Building network info cache for instance {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 757.070466] env[61923]: DEBUG nova.compute.manager [None req-1c640f71-1156-46a1-b656-4d12500e035e tempest-ServerDiagnosticsV248Test-889872339 tempest-ServerDiagnosticsV248Test-889872339-project-admin] [instance: 655b01ae-76a5-4c09-9245-e629fec8fd4b] Checking state {{(pid=61923) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 757.072029] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2785c51d-568f-4392-85b7-24d9a9a7ac8e {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.079602] env[61923]: INFO nova.compute.manager [None req-1c640f71-1156-46a1-b656-4d12500e035e tempest-ServerDiagnosticsV248Test-889872339 tempest-ServerDiagnosticsV248Test-889872339-project-admin] [instance: 655b01ae-76a5-4c09-9245-e629fec8fd4b] Retrieving diagnostics [ 757.080694] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb1de8c9-4de6-4ea0-b699-3c5b8d4b20ec {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.123088] env[61923]: INFO nova.compute.manager [None req-f1ee1de3-0f72-403f-9698-da4536b6f37f tempest-ServerDiagnosticsV248Test-1822646244 tempest-ServerDiagnosticsV248Test-1822646244-project-member] [instance: 655b01ae-76a5-4c09-9245-e629fec8fd4b] Took 26.53 seconds to build instance. [ 757.130363] env[61923]: INFO nova.compute.manager [None req-4a9f7dc2-7221-4911-9bc8-b39b7a22ddda tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] [instance: 22d1e8ed-75c5-42e9-ae55-12d59e565914] Took 1.04 seconds to deallocate network for instance. [ 757.197628] env[61923]: DEBUG nova.compute.manager [None req-6e6344ac-0cd2-410e-863e-3289b0c1fe3f tempest-ServersAdminTestJSON-678434436 tempest-ServersAdminTestJSON-678434436-project-member] [instance: 3f298e29-69ff-46ac-8018-ceb3990b9848] Starting instance... {{(pid=61923) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 757.385308] env[61923]: DEBUG nova.network.neutron [None req-5bb1ca17-5442-4564-86ab-53f866a97e1d tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: 15eaaa7c-77be-4b8b-89d0-0b2b11a0c49e] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 757.505663] env[61923]: DEBUG nova.network.neutron [None req-5bb1ca17-5442-4564-86ab-53f866a97e1d tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: 15eaaa7c-77be-4b8b-89d0-0b2b11a0c49e] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 757.625700] env[61923]: DEBUG oslo_concurrency.lockutils [None req-f1ee1de3-0f72-403f-9698-da4536b6f37f tempest-ServerDiagnosticsV248Test-1822646244 tempest-ServerDiagnosticsV248Test-1822646244-project-member] Lock "655b01ae-76a5-4c09-9245-e629fec8fd4b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 98.086s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 757.725845] env[61923]: DEBUG oslo_concurrency.lockutils [None req-6e6344ac-0cd2-410e-863e-3289b0c1fe3f tempest-ServersAdminTestJSON-678434436 tempest-ServersAdminTestJSON-678434436-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 758.007728] env[61923]: DEBUG oslo_concurrency.lockutils [None req-5bb1ca17-5442-4564-86ab-53f866a97e1d tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Releasing lock "refresh_cache-15eaaa7c-77be-4b8b-89d0-0b2b11a0c49e" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 758.008309] env[61923]: DEBUG nova.compute.manager [None req-5bb1ca17-5442-4564-86ab-53f866a97e1d tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61923) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 758.008309] env[61923]: DEBUG nova.compute.manager [None req-5bb1ca17-5442-4564-86ab-53f866a97e1d tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: 15eaaa7c-77be-4b8b-89d0-0b2b11a0c49e] Deallocating network for instance {{(pid=61923) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 758.008309] env[61923]: DEBUG nova.network.neutron [None req-5bb1ca17-5442-4564-86ab-53f866a97e1d tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: 15eaaa7c-77be-4b8b-89d0-0b2b11a0c49e] deallocate_for_instance() {{(pid=61923) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 758.023495] env[61923]: DEBUG nova.network.neutron [None req-5bb1ca17-5442-4564-86ab-53f866a97e1d tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: 15eaaa7c-77be-4b8b-89d0-0b2b11a0c49e] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 758.129675] env[61923]: DEBUG nova.compute.manager [None req-37409b66-7d98-4871-81c4-a83dd0296328 tempest-ServerAddressesNegativeTestJSON-675891034 tempest-ServerAddressesNegativeTestJSON-675891034-project-member] [instance: 6fe76341-d578-4cb3-9dce-c025f0d798b8] Starting instance... {{(pid=61923) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 758.157169] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0731bb7-83ce-4410-99b6-56e5008259c3 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.168499] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-964929c1-1262-48fe-b3d1-8a222bad2763 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.175718] env[61923]: INFO nova.scheduler.client.report [None req-4a9f7dc2-7221-4911-9bc8-b39b7a22ddda tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Deleted allocations for instance 22d1e8ed-75c5-42e9-ae55-12d59e565914 [ 758.214218] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb73142d-47e5-4605-a5ed-0bbcca552838 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.223306] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-159ded84-c4d7-4989-a985-cef293ee9569 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.238076] env[61923]: DEBUG nova.compute.provider_tree [None req-347b83bf-b4ed-4904-9bd3-d6021183ae1c tempest-ServerShowV254Test-921228709 tempest-ServerShowV254Test-921228709-project-member] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 758.526325] env[61923]: DEBUG nova.network.neutron [None req-5bb1ca17-5442-4564-86ab-53f866a97e1d tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: 15eaaa7c-77be-4b8b-89d0-0b2b11a0c49e] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 758.656433] env[61923]: DEBUG oslo_concurrency.lockutils [None req-37409b66-7d98-4871-81c4-a83dd0296328 tempest-ServerAddressesNegativeTestJSON-675891034 tempest-ServerAddressesNegativeTestJSON-675891034-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 758.686101] env[61923]: DEBUG oslo_concurrency.lockutils [None req-4a9f7dc2-7221-4911-9bc8-b39b7a22ddda tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Lock "22d1e8ed-75c5-42e9-ae55-12d59e565914" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 139.029s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 758.742876] env[61923]: DEBUG nova.scheduler.client.report [None req-347b83bf-b4ed-4904-9bd3-d6021183ae1c tempest-ServerShowV254Test-921228709 tempest-ServerShowV254Test-921228709-project-member] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 759.031601] env[61923]: INFO nova.compute.manager [None req-5bb1ca17-5442-4564-86ab-53f866a97e1d tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: 15eaaa7c-77be-4b8b-89d0-0b2b11a0c49e] Took 1.02 seconds to deallocate network for instance. [ 759.189463] env[61923]: DEBUG nova.compute.manager [None req-aed8d24a-43d8-4fbe-ac67-0dd6c37614c4 tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] [instance: 4d709a3a-65af-4adf-9ad0-4d830a9518c2] Starting instance... {{(pid=61923) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 759.249478] env[61923]: DEBUG oslo_concurrency.lockutils [None req-347b83bf-b4ed-4904-9bd3-d6021183ae1c tempest-ServerShowV254Test-921228709 tempest-ServerShowV254Test-921228709-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.394s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 759.250038] env[61923]: DEBUG nova.compute.manager [None req-347b83bf-b4ed-4904-9bd3-d6021183ae1c tempest-ServerShowV254Test-921228709 tempest-ServerShowV254Test-921228709-project-member] [instance: ecedb77f-4a2a-4047-964f-80796954d847] Start building networks asynchronously for instance. {{(pid=61923) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 759.252724] env[61923]: DEBUG oslo_concurrency.lockutils [None req-0708225f-80bb-41ec-b3ed-7dd45d589b95 tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 13.806s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 759.723300] env[61923]: DEBUG oslo_concurrency.lockutils [None req-aed8d24a-43d8-4fbe-ac67-0dd6c37614c4 tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 759.759435] env[61923]: DEBUG nova.compute.utils [None req-347b83bf-b4ed-4904-9bd3-d6021183ae1c tempest-ServerShowV254Test-921228709 tempest-ServerShowV254Test-921228709-project-member] Using /dev/sd instead of None {{(pid=61923) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 759.764483] env[61923]: DEBUG nova.compute.manager [None req-347b83bf-b4ed-4904-9bd3-d6021183ae1c tempest-ServerShowV254Test-921228709 tempest-ServerShowV254Test-921228709-project-member] [instance: ecedb77f-4a2a-4047-964f-80796954d847] Not allocating networking since 'none' was specified. {{(pid=61923) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 760.062103] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-779dc9f8-9b85-486b-85d7-50ab68a7fe08 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.065881] env[61923]: INFO nova.scheduler.client.report [None req-5bb1ca17-5442-4564-86ab-53f866a97e1d tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Deleted allocations for instance 15eaaa7c-77be-4b8b-89d0-0b2b11a0c49e [ 760.076961] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b09868e-dd01-48f3-88fa-dd420e608469 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.111446] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2efebbf8-cfc7-4fcf-968c-56388f5beb19 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.119421] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3fe6f59c-160d-41a7-b4a8-678fbae3e281 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.133626] env[61923]: DEBUG nova.compute.provider_tree [None req-0708225f-80bb-41ec-b3ed-7dd45d589b95 tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 760.265307] env[61923]: DEBUG nova.compute.manager [None req-347b83bf-b4ed-4904-9bd3-d6021183ae1c tempest-ServerShowV254Test-921228709 tempest-ServerShowV254Test-921228709-project-member] [instance: ecedb77f-4a2a-4047-964f-80796954d847] Start building block device mappings for instance. {{(pid=61923) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 760.576527] env[61923]: DEBUG oslo_concurrency.lockutils [None req-5bb1ca17-5442-4564-86ab-53f866a97e1d tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Lock "15eaaa7c-77be-4b8b-89d0-0b2b11a0c49e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 132.737s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 760.636739] env[61923]: DEBUG nova.scheduler.client.report [None req-0708225f-80bb-41ec-b3ed-7dd45d589b95 tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 761.082223] env[61923]: DEBUG nova.compute.manager [None req-aed8d24a-43d8-4fbe-ac67-0dd6c37614c4 tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] [instance: fe981195-1e2c-4c8d-8c53-327851f845c7] Starting instance... {{(pid=61923) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 761.141863] env[61923]: DEBUG oslo_concurrency.lockutils [None req-0708225f-80bb-41ec-b3ed-7dd45d589b95 tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.889s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 761.142651] env[61923]: ERROR nova.compute.manager [None req-0708225f-80bb-41ec-b3ed-7dd45d589b95 tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] [instance: 62d8a170-e7ad-4d14-90d8-6f6ce32b7cf8] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port fbdf8b6e-0d86-4121-8aeb-b977a59f1186, please check neutron logs for more information. [ 761.142651] env[61923]: ERROR nova.compute.manager [instance: 62d8a170-e7ad-4d14-90d8-6f6ce32b7cf8] Traceback (most recent call last): [ 761.142651] env[61923]: ERROR nova.compute.manager [instance: 62d8a170-e7ad-4d14-90d8-6f6ce32b7cf8] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 761.142651] env[61923]: ERROR nova.compute.manager [instance: 62d8a170-e7ad-4d14-90d8-6f6ce32b7cf8] self.driver.spawn(context, instance, image_meta, [ 761.142651] env[61923]: ERROR nova.compute.manager [instance: 62d8a170-e7ad-4d14-90d8-6f6ce32b7cf8] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 761.142651] env[61923]: ERROR nova.compute.manager [instance: 62d8a170-e7ad-4d14-90d8-6f6ce32b7cf8] self._vmops.spawn(context, instance, image_meta, injected_files, [ 761.142651] env[61923]: ERROR nova.compute.manager [instance: 62d8a170-e7ad-4d14-90d8-6f6ce32b7cf8] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 761.142651] env[61923]: ERROR nova.compute.manager [instance: 62d8a170-e7ad-4d14-90d8-6f6ce32b7cf8] vm_ref = self.build_virtual_machine(instance, [ 761.142651] env[61923]: ERROR nova.compute.manager [instance: 62d8a170-e7ad-4d14-90d8-6f6ce32b7cf8] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 761.142651] env[61923]: ERROR nova.compute.manager [instance: 62d8a170-e7ad-4d14-90d8-6f6ce32b7cf8] vif_infos = vmwarevif.get_vif_info(self._session, [ 761.142651] env[61923]: ERROR nova.compute.manager [instance: 62d8a170-e7ad-4d14-90d8-6f6ce32b7cf8] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 761.143048] env[61923]: ERROR nova.compute.manager [instance: 62d8a170-e7ad-4d14-90d8-6f6ce32b7cf8] for vif in network_info: [ 761.143048] env[61923]: ERROR nova.compute.manager [instance: 62d8a170-e7ad-4d14-90d8-6f6ce32b7cf8] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 761.143048] env[61923]: ERROR nova.compute.manager [instance: 62d8a170-e7ad-4d14-90d8-6f6ce32b7cf8] return self._sync_wrapper(fn, *args, **kwargs) [ 761.143048] env[61923]: ERROR nova.compute.manager [instance: 62d8a170-e7ad-4d14-90d8-6f6ce32b7cf8] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 761.143048] env[61923]: ERROR nova.compute.manager [instance: 62d8a170-e7ad-4d14-90d8-6f6ce32b7cf8] self.wait() [ 761.143048] env[61923]: ERROR nova.compute.manager [instance: 62d8a170-e7ad-4d14-90d8-6f6ce32b7cf8] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 761.143048] env[61923]: ERROR nova.compute.manager [instance: 62d8a170-e7ad-4d14-90d8-6f6ce32b7cf8] self[:] = self._gt.wait() [ 761.143048] env[61923]: ERROR nova.compute.manager [instance: 62d8a170-e7ad-4d14-90d8-6f6ce32b7cf8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 761.143048] env[61923]: ERROR nova.compute.manager [instance: 62d8a170-e7ad-4d14-90d8-6f6ce32b7cf8] return self._exit_event.wait() [ 761.143048] env[61923]: ERROR nova.compute.manager [instance: 62d8a170-e7ad-4d14-90d8-6f6ce32b7cf8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 761.143048] env[61923]: ERROR nova.compute.manager [instance: 62d8a170-e7ad-4d14-90d8-6f6ce32b7cf8] result = hub.switch() [ 761.143048] env[61923]: ERROR nova.compute.manager [instance: 62d8a170-e7ad-4d14-90d8-6f6ce32b7cf8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 761.143048] env[61923]: ERROR nova.compute.manager [instance: 62d8a170-e7ad-4d14-90d8-6f6ce32b7cf8] return self.greenlet.switch() [ 761.143569] env[61923]: ERROR nova.compute.manager [instance: 62d8a170-e7ad-4d14-90d8-6f6ce32b7cf8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 761.143569] env[61923]: ERROR nova.compute.manager [instance: 62d8a170-e7ad-4d14-90d8-6f6ce32b7cf8] result = function(*args, **kwargs) [ 761.143569] env[61923]: ERROR nova.compute.manager [instance: 62d8a170-e7ad-4d14-90d8-6f6ce32b7cf8] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 761.143569] env[61923]: ERROR nova.compute.manager [instance: 62d8a170-e7ad-4d14-90d8-6f6ce32b7cf8] return func(*args, **kwargs) [ 761.143569] env[61923]: ERROR nova.compute.manager [instance: 62d8a170-e7ad-4d14-90d8-6f6ce32b7cf8] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 761.143569] env[61923]: ERROR nova.compute.manager [instance: 62d8a170-e7ad-4d14-90d8-6f6ce32b7cf8] raise e [ 761.143569] env[61923]: ERROR nova.compute.manager [instance: 62d8a170-e7ad-4d14-90d8-6f6ce32b7cf8] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 761.143569] env[61923]: ERROR nova.compute.manager [instance: 62d8a170-e7ad-4d14-90d8-6f6ce32b7cf8] nwinfo = self.network_api.allocate_for_instance( [ 761.143569] env[61923]: ERROR nova.compute.manager [instance: 62d8a170-e7ad-4d14-90d8-6f6ce32b7cf8] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 761.143569] env[61923]: ERROR nova.compute.manager [instance: 62d8a170-e7ad-4d14-90d8-6f6ce32b7cf8] created_port_ids = self._update_ports_for_instance( [ 761.143569] env[61923]: ERROR nova.compute.manager [instance: 62d8a170-e7ad-4d14-90d8-6f6ce32b7cf8] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 761.143569] env[61923]: ERROR nova.compute.manager [instance: 62d8a170-e7ad-4d14-90d8-6f6ce32b7cf8] with excutils.save_and_reraise_exception(): [ 761.143569] env[61923]: ERROR nova.compute.manager [instance: 62d8a170-e7ad-4d14-90d8-6f6ce32b7cf8] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 761.143972] env[61923]: ERROR nova.compute.manager [instance: 62d8a170-e7ad-4d14-90d8-6f6ce32b7cf8] self.force_reraise() [ 761.143972] env[61923]: ERROR nova.compute.manager [instance: 62d8a170-e7ad-4d14-90d8-6f6ce32b7cf8] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 761.143972] env[61923]: ERROR nova.compute.manager [instance: 62d8a170-e7ad-4d14-90d8-6f6ce32b7cf8] raise self.value [ 761.143972] env[61923]: ERROR nova.compute.manager [instance: 62d8a170-e7ad-4d14-90d8-6f6ce32b7cf8] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 761.143972] env[61923]: ERROR nova.compute.manager [instance: 62d8a170-e7ad-4d14-90d8-6f6ce32b7cf8] updated_port = self._update_port( [ 761.143972] env[61923]: ERROR nova.compute.manager [instance: 62d8a170-e7ad-4d14-90d8-6f6ce32b7cf8] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 761.143972] env[61923]: ERROR nova.compute.manager [instance: 62d8a170-e7ad-4d14-90d8-6f6ce32b7cf8] _ensure_no_port_binding_failure(port) [ 761.143972] env[61923]: ERROR nova.compute.manager [instance: 62d8a170-e7ad-4d14-90d8-6f6ce32b7cf8] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 761.143972] env[61923]: ERROR nova.compute.manager [instance: 62d8a170-e7ad-4d14-90d8-6f6ce32b7cf8] raise exception.PortBindingFailed(port_id=port['id']) [ 761.143972] env[61923]: ERROR nova.compute.manager [instance: 62d8a170-e7ad-4d14-90d8-6f6ce32b7cf8] nova.exception.PortBindingFailed: Binding failed for port fbdf8b6e-0d86-4121-8aeb-b977a59f1186, please check neutron logs for more information. [ 761.143972] env[61923]: ERROR nova.compute.manager [instance: 62d8a170-e7ad-4d14-90d8-6f6ce32b7cf8] [ 761.144830] env[61923]: DEBUG nova.compute.utils [None req-0708225f-80bb-41ec-b3ed-7dd45d589b95 tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] [instance: 62d8a170-e7ad-4d14-90d8-6f6ce32b7cf8] Binding failed for port fbdf8b6e-0d86-4121-8aeb-b977a59f1186, please check neutron logs for more information. {{(pid=61923) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 761.145999] env[61923]: DEBUG oslo_concurrency.lockutils [None req-ab7b3da0-3a01-44d4-9dfd-3e5d2ad27432 tempest-TenantUsagesTestJSON-871253541 tempest-TenantUsagesTestJSON-871253541-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.118s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 761.148856] env[61923]: DEBUG nova.compute.manager [None req-0708225f-80bb-41ec-b3ed-7dd45d589b95 tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] [instance: 62d8a170-e7ad-4d14-90d8-6f6ce32b7cf8] Build of instance 62d8a170-e7ad-4d14-90d8-6f6ce32b7cf8 was re-scheduled: Binding failed for port fbdf8b6e-0d86-4121-8aeb-b977a59f1186, please check neutron logs for more information. {{(pid=61923) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 761.149277] env[61923]: DEBUG nova.compute.manager [None req-0708225f-80bb-41ec-b3ed-7dd45d589b95 tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] [instance: 62d8a170-e7ad-4d14-90d8-6f6ce32b7cf8] Unplugging VIFs for instance {{(pid=61923) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 761.149504] env[61923]: DEBUG oslo_concurrency.lockutils [None req-0708225f-80bb-41ec-b3ed-7dd45d589b95 tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Acquiring lock "refresh_cache-62d8a170-e7ad-4d14-90d8-6f6ce32b7cf8" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 761.149671] env[61923]: DEBUG oslo_concurrency.lockutils [None req-0708225f-80bb-41ec-b3ed-7dd45d589b95 tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Acquired lock "refresh_cache-62d8a170-e7ad-4d14-90d8-6f6ce32b7cf8" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 761.149798] env[61923]: DEBUG nova.network.neutron [None req-0708225f-80bb-41ec-b3ed-7dd45d589b95 tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] [instance: 62d8a170-e7ad-4d14-90d8-6f6ce32b7cf8] Building network info cache for instance {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 761.277399] env[61923]: DEBUG nova.compute.manager [None req-347b83bf-b4ed-4904-9bd3-d6021183ae1c tempest-ServerShowV254Test-921228709 tempest-ServerShowV254Test-921228709-project-member] [instance: ecedb77f-4a2a-4047-964f-80796954d847] Start spawning the instance on the hypervisor. {{(pid=61923) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 761.302216] env[61923]: DEBUG nova.virt.hardware [None req-347b83bf-b4ed-4904-9bd3-d6021183ae1c tempest-ServerShowV254Test-921228709 tempest-ServerShowV254Test-921228709-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-29T20:07:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-29T20:07:35Z,direct_url=,disk_format='vmdk',id=0f153f63-ae0a-45b1-b7f5-7f9b673c947f,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='4e7b5e3b3c3443c8bd5c70f8a15739f5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-29T20:07:36Z,virtual_size=,visibility=), allow threads: False {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 761.302455] env[61923]: DEBUG nova.virt.hardware [None req-347b83bf-b4ed-4904-9bd3-d6021183ae1c tempest-ServerShowV254Test-921228709 tempest-ServerShowV254Test-921228709-project-member] Flavor limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 761.302611] env[61923]: DEBUG nova.virt.hardware [None req-347b83bf-b4ed-4904-9bd3-d6021183ae1c tempest-ServerShowV254Test-921228709 tempest-ServerShowV254Test-921228709-project-member] Image limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 761.302790] env[61923]: DEBUG nova.virt.hardware [None req-347b83bf-b4ed-4904-9bd3-d6021183ae1c tempest-ServerShowV254Test-921228709 tempest-ServerShowV254Test-921228709-project-member] Flavor pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 761.302931] env[61923]: DEBUG nova.virt.hardware [None req-347b83bf-b4ed-4904-9bd3-d6021183ae1c tempest-ServerShowV254Test-921228709 tempest-ServerShowV254Test-921228709-project-member] Image pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 761.303085] env[61923]: DEBUG nova.virt.hardware [None req-347b83bf-b4ed-4904-9bd3-d6021183ae1c tempest-ServerShowV254Test-921228709 tempest-ServerShowV254Test-921228709-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 761.303400] env[61923]: DEBUG nova.virt.hardware [None req-347b83bf-b4ed-4904-9bd3-d6021183ae1c tempest-ServerShowV254Test-921228709 tempest-ServerShowV254Test-921228709-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 761.303474] env[61923]: DEBUG nova.virt.hardware [None req-347b83bf-b4ed-4904-9bd3-d6021183ae1c tempest-ServerShowV254Test-921228709 tempest-ServerShowV254Test-921228709-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 761.303607] env[61923]: DEBUG nova.virt.hardware [None req-347b83bf-b4ed-4904-9bd3-d6021183ae1c tempest-ServerShowV254Test-921228709 tempest-ServerShowV254Test-921228709-project-member] Got 1 possible topologies {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 761.303765] env[61923]: DEBUG nova.virt.hardware [None req-347b83bf-b4ed-4904-9bd3-d6021183ae1c tempest-ServerShowV254Test-921228709 tempest-ServerShowV254Test-921228709-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 761.303935] env[61923]: DEBUG nova.virt.hardware [None req-347b83bf-b4ed-4904-9bd3-d6021183ae1c tempest-ServerShowV254Test-921228709 tempest-ServerShowV254Test-921228709-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 761.304796] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97c9abea-a023-40d4-8975-edb3e4a4ca6e {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.312987] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4ae8f8b-76e4-4448-ba24-13de5c305a4b {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.326905] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-347b83bf-b4ed-4904-9bd3-d6021183ae1c tempest-ServerShowV254Test-921228709 tempest-ServerShowV254Test-921228709-project-member] [instance: ecedb77f-4a2a-4047-964f-80796954d847] Instance VIF info [] {{(pid=61923) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 761.332969] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-347b83bf-b4ed-4904-9bd3-d6021183ae1c tempest-ServerShowV254Test-921228709 tempest-ServerShowV254Test-921228709-project-member] Creating folder: Project (c29fb0e5a3f14691a41e6bdb1b3a1899). Parent ref: group-v292629. {{(pid=61923) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 761.333089] env[61923]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-946c2894-4a53-4ae7-b483-75fe664019a2 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.342297] env[61923]: INFO nova.virt.vmwareapi.vm_util [None req-347b83bf-b4ed-4904-9bd3-d6021183ae1c tempest-ServerShowV254Test-921228709 tempest-ServerShowV254Test-921228709-project-member] Created folder: Project (c29fb0e5a3f14691a41e6bdb1b3a1899) in parent group-v292629. [ 761.342498] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-347b83bf-b4ed-4904-9bd3-d6021183ae1c tempest-ServerShowV254Test-921228709 tempest-ServerShowV254Test-921228709-project-member] Creating folder: Instances. Parent ref: group-v292646. {{(pid=61923) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 761.342737] env[61923]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-ee884177-8317-41c5-8d0d-74d937ce6ade {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.349193] env[61923]: DEBUG oslo_concurrency.lockutils [None req-950e07c2-502c-43fa-8598-da5f189b07e3 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Acquiring lock "422d06d6-6932-46c3-bb25-841e6f65c028" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 761.349438] env[61923]: DEBUG oslo_concurrency.lockutils [None req-950e07c2-502c-43fa-8598-da5f189b07e3 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Lock "422d06d6-6932-46c3-bb25-841e6f65c028" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 761.354431] env[61923]: INFO nova.virt.vmwareapi.vm_util [None req-347b83bf-b4ed-4904-9bd3-d6021183ae1c tempest-ServerShowV254Test-921228709 tempest-ServerShowV254Test-921228709-project-member] Created folder: Instances in parent group-v292646. [ 761.354649] env[61923]: DEBUG oslo.service.loopingcall [None req-347b83bf-b4ed-4904-9bd3-d6021183ae1c tempest-ServerShowV254Test-921228709 tempest-ServerShowV254Test-921228709-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61923) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 761.354825] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ecedb77f-4a2a-4047-964f-80796954d847] Creating VM on the ESX host {{(pid=61923) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 761.355032] env[61923]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-4749bd55-eaec-4725-8d33-64bb06969412 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.370781] env[61923]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 761.370781] env[61923]: value = "task-1377493" [ 761.370781] env[61923]: _type = "Task" [ 761.370781] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 761.378367] env[61923]: DEBUG oslo_vmware.api [-] Task: {'id': task-1377493, 'name': CreateVM_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 761.575957] env[61923]: DEBUG oslo_concurrency.lockutils [None req-e62df7da-42b5-4263-bcf8-001a2d06427a tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Acquiring lock "73692517-1816-4e9b-ab2d-8265c683e83d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 761.576073] env[61923]: DEBUG oslo_concurrency.lockutils [None req-e62df7da-42b5-4263-bcf8-001a2d06427a tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Lock "73692517-1816-4e9b-ab2d-8265c683e83d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 761.602278] env[61923]: DEBUG oslo_concurrency.lockutils [None req-aed8d24a-43d8-4fbe-ac67-0dd6c37614c4 tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 761.673254] env[61923]: DEBUG nova.network.neutron [None req-0708225f-80bb-41ec-b3ed-7dd45d589b95 tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] [instance: 62d8a170-e7ad-4d14-90d8-6f6ce32b7cf8] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 761.760144] env[61923]: DEBUG nova.network.neutron [None req-0708225f-80bb-41ec-b3ed-7dd45d589b95 tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] [instance: 62d8a170-e7ad-4d14-90d8-6f6ce32b7cf8] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 761.881099] env[61923]: DEBUG oslo_vmware.api [-] Task: {'id': task-1377493, 'name': CreateVM_Task, 'duration_secs': 0.344615} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 761.881217] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ecedb77f-4a2a-4047-964f-80796954d847] Created VM on the ESX host {{(pid=61923) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 761.881691] env[61923]: DEBUG oslo_concurrency.lockutils [None req-347b83bf-b4ed-4904-9bd3-d6021183ae1c tempest-ServerShowV254Test-921228709 tempest-ServerShowV254Test-921228709-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 761.881880] env[61923]: DEBUG oslo_concurrency.lockutils [None req-347b83bf-b4ed-4904-9bd3-d6021183ae1c tempest-ServerShowV254Test-921228709 tempest-ServerShowV254Test-921228709-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 761.882236] env[61923]: DEBUG oslo_concurrency.lockutils [None req-347b83bf-b4ed-4904-9bd3-d6021183ae1c tempest-ServerShowV254Test-921228709 tempest-ServerShowV254Test-921228709-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 761.882516] env[61923]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-78c18af4-0770-4857-9f91-e6752e48ce3b {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.887111] env[61923]: DEBUG oslo_vmware.api [None req-347b83bf-b4ed-4904-9bd3-d6021183ae1c tempest-ServerShowV254Test-921228709 tempest-ServerShowV254Test-921228709-project-member] Waiting for the task: (returnval){ [ 761.887111] env[61923]: value = "session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]526d5266-c383-fd20-a35e-d2437f13469b" [ 761.887111] env[61923]: _type = "Task" [ 761.887111] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 761.897545] env[61923]: DEBUG oslo_vmware.api [None req-347b83bf-b4ed-4904-9bd3-d6021183ae1c tempest-ServerShowV254Test-921228709 tempest-ServerShowV254Test-921228709-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]526d5266-c383-fd20-a35e-d2437f13469b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 761.902367] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4bb28c8-b1d0-4746-828e-aa2954f699ed {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.909289] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d89fd87c-2f6d-4015-b60d-9e4fc7d6e1dd {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.938837] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a19e25cf-96b5-4dfc-9571-04fd4b0ca502 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.946320] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15074928-affd-4d26-a053-867aa1b18425 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.959199] env[61923]: DEBUG nova.compute.provider_tree [None req-ab7b3da0-3a01-44d4-9dfd-3e5d2ad27432 tempest-TenantUsagesTestJSON-871253541 tempest-TenantUsagesTestJSON-871253541-project-member] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 762.263154] env[61923]: DEBUG oslo_concurrency.lockutils [None req-0708225f-80bb-41ec-b3ed-7dd45d589b95 tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Releasing lock "refresh_cache-62d8a170-e7ad-4d14-90d8-6f6ce32b7cf8" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 762.263439] env[61923]: DEBUG nova.compute.manager [None req-0708225f-80bb-41ec-b3ed-7dd45d589b95 tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61923) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 762.263591] env[61923]: DEBUG nova.compute.manager [None req-0708225f-80bb-41ec-b3ed-7dd45d589b95 tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] [instance: 62d8a170-e7ad-4d14-90d8-6f6ce32b7cf8] Deallocating network for instance {{(pid=61923) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 762.263760] env[61923]: DEBUG nova.network.neutron [None req-0708225f-80bb-41ec-b3ed-7dd45d589b95 tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] [instance: 62d8a170-e7ad-4d14-90d8-6f6ce32b7cf8] deallocate_for_instance() {{(pid=61923) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 762.279267] env[61923]: DEBUG nova.network.neutron [None req-0708225f-80bb-41ec-b3ed-7dd45d589b95 tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] [instance: 62d8a170-e7ad-4d14-90d8-6f6ce32b7cf8] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 762.397872] env[61923]: DEBUG oslo_vmware.api [None req-347b83bf-b4ed-4904-9bd3-d6021183ae1c tempest-ServerShowV254Test-921228709 tempest-ServerShowV254Test-921228709-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]526d5266-c383-fd20-a35e-d2437f13469b, 'name': SearchDatastore_Task, 'duration_secs': 0.010221} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 762.398198] env[61923]: DEBUG oslo_concurrency.lockutils [None req-347b83bf-b4ed-4904-9bd3-d6021183ae1c tempest-ServerShowV254Test-921228709 tempest-ServerShowV254Test-921228709-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 762.398523] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-347b83bf-b4ed-4904-9bd3-d6021183ae1c tempest-ServerShowV254Test-921228709 tempest-ServerShowV254Test-921228709-project-member] [instance: ecedb77f-4a2a-4047-964f-80796954d847] Processing image 0f153f63-ae0a-45b1-b7f5-7f9b673c947f {{(pid=61923) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 762.400557] env[61923]: DEBUG oslo_concurrency.lockutils [None req-347b83bf-b4ed-4904-9bd3-d6021183ae1c tempest-ServerShowV254Test-921228709 tempest-ServerShowV254Test-921228709-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 762.400557] env[61923]: DEBUG oslo_concurrency.lockutils [None req-347b83bf-b4ed-4904-9bd3-d6021183ae1c tempest-ServerShowV254Test-921228709 tempest-ServerShowV254Test-921228709-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 762.400557] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-347b83bf-b4ed-4904-9bd3-d6021183ae1c tempest-ServerShowV254Test-921228709 tempest-ServerShowV254Test-921228709-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61923) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 762.400557] env[61923]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-7355acdb-724f-4e01-8720-1233a3063738 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.406899] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-347b83bf-b4ed-4904-9bd3-d6021183ae1c tempest-ServerShowV254Test-921228709 tempest-ServerShowV254Test-921228709-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61923) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 762.407074] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-347b83bf-b4ed-4904-9bd3-d6021183ae1c tempest-ServerShowV254Test-921228709 tempest-ServerShowV254Test-921228709-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61923) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 762.407818] env[61923]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-40061081-b3e5-435c-8a6f-201fde03c0f4 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.412662] env[61923]: DEBUG oslo_vmware.api [None req-347b83bf-b4ed-4904-9bd3-d6021183ae1c tempest-ServerShowV254Test-921228709 tempest-ServerShowV254Test-921228709-project-member] Waiting for the task: (returnval){ [ 762.412662] env[61923]: value = "session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52059821-6088-1aee-017a-ece26c617ba3" [ 762.412662] env[61923]: _type = "Task" [ 762.412662] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 762.421287] env[61923]: DEBUG oslo_vmware.api [None req-347b83bf-b4ed-4904-9bd3-d6021183ae1c tempest-ServerShowV254Test-921228709 tempest-ServerShowV254Test-921228709-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52059821-6088-1aee-017a-ece26c617ba3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 762.462364] env[61923]: DEBUG nova.scheduler.client.report [None req-ab7b3da0-3a01-44d4-9dfd-3e5d2ad27432 tempest-TenantUsagesTestJSON-871253541 tempest-TenantUsagesTestJSON-871253541-project-member] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 762.783060] env[61923]: DEBUG nova.network.neutron [None req-0708225f-80bb-41ec-b3ed-7dd45d589b95 tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] [instance: 62d8a170-e7ad-4d14-90d8-6f6ce32b7cf8] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 762.923093] env[61923]: DEBUG oslo_vmware.api [None req-347b83bf-b4ed-4904-9bd3-d6021183ae1c tempest-ServerShowV254Test-921228709 tempest-ServerShowV254Test-921228709-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52059821-6088-1aee-017a-ece26c617ba3, 'name': SearchDatastore_Task, 'duration_secs': 0.007964} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 762.923881] env[61923]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7c91ca55-96e3-4be5-bbd4-3d1f03fa1e12 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.928837] env[61923]: DEBUG oslo_vmware.api [None req-347b83bf-b4ed-4904-9bd3-d6021183ae1c tempest-ServerShowV254Test-921228709 tempest-ServerShowV254Test-921228709-project-member] Waiting for the task: (returnval){ [ 762.928837] env[61923]: value = "session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52b17950-ae16-c968-afb0-720e3490fd80" [ 762.928837] env[61923]: _type = "Task" [ 762.928837] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 762.936437] env[61923]: DEBUG oslo_vmware.api [None req-347b83bf-b4ed-4904-9bd3-d6021183ae1c tempest-ServerShowV254Test-921228709 tempest-ServerShowV254Test-921228709-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52b17950-ae16-c968-afb0-720e3490fd80, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 762.969536] env[61923]: DEBUG oslo_concurrency.lockutils [None req-ab7b3da0-3a01-44d4-9dfd-3e5d2ad27432 tempest-TenantUsagesTestJSON-871253541 tempest-TenantUsagesTestJSON-871253541-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.823s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 762.970617] env[61923]: ERROR nova.compute.manager [None req-ab7b3da0-3a01-44d4-9dfd-3e5d2ad27432 tempest-TenantUsagesTestJSON-871253541 tempest-TenantUsagesTestJSON-871253541-project-member] [instance: b6712878-3270-43f4-9518-d41eb77124b7] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port a4338c60-f3e2-4844-b170-c852adcfcb25, please check neutron logs for more information. [ 762.970617] env[61923]: ERROR nova.compute.manager [instance: b6712878-3270-43f4-9518-d41eb77124b7] Traceback (most recent call last): [ 762.970617] env[61923]: ERROR nova.compute.manager [instance: b6712878-3270-43f4-9518-d41eb77124b7] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 762.970617] env[61923]: ERROR nova.compute.manager [instance: b6712878-3270-43f4-9518-d41eb77124b7] self.driver.spawn(context, instance, image_meta, [ 762.970617] env[61923]: ERROR nova.compute.manager [instance: b6712878-3270-43f4-9518-d41eb77124b7] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 762.970617] env[61923]: ERROR nova.compute.manager [instance: b6712878-3270-43f4-9518-d41eb77124b7] self._vmops.spawn(context, instance, image_meta, injected_files, [ 762.970617] env[61923]: ERROR nova.compute.manager [instance: b6712878-3270-43f4-9518-d41eb77124b7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 762.970617] env[61923]: ERROR nova.compute.manager [instance: b6712878-3270-43f4-9518-d41eb77124b7] vm_ref = self.build_virtual_machine(instance, [ 762.970617] env[61923]: ERROR nova.compute.manager [instance: b6712878-3270-43f4-9518-d41eb77124b7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 762.970617] env[61923]: ERROR nova.compute.manager [instance: b6712878-3270-43f4-9518-d41eb77124b7] vif_infos = vmwarevif.get_vif_info(self._session, [ 762.970617] env[61923]: ERROR nova.compute.manager [instance: b6712878-3270-43f4-9518-d41eb77124b7] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 762.971077] env[61923]: ERROR nova.compute.manager [instance: b6712878-3270-43f4-9518-d41eb77124b7] for vif in network_info: [ 762.971077] env[61923]: ERROR nova.compute.manager [instance: b6712878-3270-43f4-9518-d41eb77124b7] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 762.971077] env[61923]: ERROR nova.compute.manager [instance: b6712878-3270-43f4-9518-d41eb77124b7] return self._sync_wrapper(fn, *args, **kwargs) [ 762.971077] env[61923]: ERROR nova.compute.manager [instance: b6712878-3270-43f4-9518-d41eb77124b7] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 762.971077] env[61923]: ERROR nova.compute.manager [instance: b6712878-3270-43f4-9518-d41eb77124b7] self.wait() [ 762.971077] env[61923]: ERROR nova.compute.manager [instance: b6712878-3270-43f4-9518-d41eb77124b7] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 762.971077] env[61923]: ERROR nova.compute.manager [instance: b6712878-3270-43f4-9518-d41eb77124b7] self[:] = self._gt.wait() [ 762.971077] env[61923]: ERROR nova.compute.manager [instance: b6712878-3270-43f4-9518-d41eb77124b7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 762.971077] env[61923]: ERROR nova.compute.manager [instance: b6712878-3270-43f4-9518-d41eb77124b7] return self._exit_event.wait() [ 762.971077] env[61923]: ERROR nova.compute.manager [instance: b6712878-3270-43f4-9518-d41eb77124b7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 762.971077] env[61923]: ERROR nova.compute.manager [instance: b6712878-3270-43f4-9518-d41eb77124b7] result = hub.switch() [ 762.971077] env[61923]: ERROR nova.compute.manager [instance: b6712878-3270-43f4-9518-d41eb77124b7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 762.971077] env[61923]: ERROR nova.compute.manager [instance: b6712878-3270-43f4-9518-d41eb77124b7] return self.greenlet.switch() [ 762.971454] env[61923]: ERROR nova.compute.manager [instance: b6712878-3270-43f4-9518-d41eb77124b7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 762.971454] env[61923]: ERROR nova.compute.manager [instance: b6712878-3270-43f4-9518-d41eb77124b7] result = function(*args, **kwargs) [ 762.971454] env[61923]: ERROR nova.compute.manager [instance: b6712878-3270-43f4-9518-d41eb77124b7] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 762.971454] env[61923]: ERROR nova.compute.manager [instance: b6712878-3270-43f4-9518-d41eb77124b7] return func(*args, **kwargs) [ 762.971454] env[61923]: ERROR nova.compute.manager [instance: b6712878-3270-43f4-9518-d41eb77124b7] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 762.971454] env[61923]: ERROR nova.compute.manager [instance: b6712878-3270-43f4-9518-d41eb77124b7] raise e [ 762.971454] env[61923]: ERROR nova.compute.manager [instance: b6712878-3270-43f4-9518-d41eb77124b7] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 762.971454] env[61923]: ERROR nova.compute.manager [instance: b6712878-3270-43f4-9518-d41eb77124b7] nwinfo = self.network_api.allocate_for_instance( [ 762.971454] env[61923]: ERROR nova.compute.manager [instance: b6712878-3270-43f4-9518-d41eb77124b7] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 762.971454] env[61923]: ERROR nova.compute.manager [instance: b6712878-3270-43f4-9518-d41eb77124b7] created_port_ids = self._update_ports_for_instance( [ 762.971454] env[61923]: ERROR nova.compute.manager [instance: b6712878-3270-43f4-9518-d41eb77124b7] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 762.971454] env[61923]: ERROR nova.compute.manager [instance: b6712878-3270-43f4-9518-d41eb77124b7] with excutils.save_and_reraise_exception(): [ 762.971454] env[61923]: ERROR nova.compute.manager [instance: b6712878-3270-43f4-9518-d41eb77124b7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 762.971866] env[61923]: ERROR nova.compute.manager [instance: b6712878-3270-43f4-9518-d41eb77124b7] self.force_reraise() [ 762.971866] env[61923]: ERROR nova.compute.manager [instance: b6712878-3270-43f4-9518-d41eb77124b7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 762.971866] env[61923]: ERROR nova.compute.manager [instance: b6712878-3270-43f4-9518-d41eb77124b7] raise self.value [ 762.971866] env[61923]: ERROR nova.compute.manager [instance: b6712878-3270-43f4-9518-d41eb77124b7] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 762.971866] env[61923]: ERROR nova.compute.manager [instance: b6712878-3270-43f4-9518-d41eb77124b7] updated_port = self._update_port( [ 762.971866] env[61923]: ERROR nova.compute.manager [instance: b6712878-3270-43f4-9518-d41eb77124b7] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 762.971866] env[61923]: ERROR nova.compute.manager [instance: b6712878-3270-43f4-9518-d41eb77124b7] _ensure_no_port_binding_failure(port) [ 762.971866] env[61923]: ERROR nova.compute.manager [instance: b6712878-3270-43f4-9518-d41eb77124b7] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 762.971866] env[61923]: ERROR nova.compute.manager [instance: b6712878-3270-43f4-9518-d41eb77124b7] raise exception.PortBindingFailed(port_id=port['id']) [ 762.971866] env[61923]: ERROR nova.compute.manager [instance: b6712878-3270-43f4-9518-d41eb77124b7] nova.exception.PortBindingFailed: Binding failed for port a4338c60-f3e2-4844-b170-c852adcfcb25, please check neutron logs for more information. [ 762.971866] env[61923]: ERROR nova.compute.manager [instance: b6712878-3270-43f4-9518-d41eb77124b7] [ 762.972256] env[61923]: DEBUG nova.compute.utils [None req-ab7b3da0-3a01-44d4-9dfd-3e5d2ad27432 tempest-TenantUsagesTestJSON-871253541 tempest-TenantUsagesTestJSON-871253541-project-member] [instance: b6712878-3270-43f4-9518-d41eb77124b7] Binding failed for port a4338c60-f3e2-4844-b170-c852adcfcb25, please check neutron logs for more information. {{(pid=61923) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 762.972427] env[61923]: DEBUG oslo_concurrency.lockutils [None req-5138192d-6349-432b-b7de-64ec5d391f8d tempest-ServerActionsTestOtherB-987045973 tempest-ServerActionsTestOtherB-987045973-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 13.704s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 762.975472] env[61923]: DEBUG nova.compute.manager [None req-ab7b3da0-3a01-44d4-9dfd-3e5d2ad27432 tempest-TenantUsagesTestJSON-871253541 tempest-TenantUsagesTestJSON-871253541-project-member] [instance: b6712878-3270-43f4-9518-d41eb77124b7] Build of instance b6712878-3270-43f4-9518-d41eb77124b7 was re-scheduled: Binding failed for port a4338c60-f3e2-4844-b170-c852adcfcb25, please check neutron logs for more information. {{(pid=61923) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 762.975913] env[61923]: DEBUG nova.compute.manager [None req-ab7b3da0-3a01-44d4-9dfd-3e5d2ad27432 tempest-TenantUsagesTestJSON-871253541 tempest-TenantUsagesTestJSON-871253541-project-member] [instance: b6712878-3270-43f4-9518-d41eb77124b7] Unplugging VIFs for instance {{(pid=61923) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 762.976152] env[61923]: DEBUG oslo_concurrency.lockutils [None req-ab7b3da0-3a01-44d4-9dfd-3e5d2ad27432 tempest-TenantUsagesTestJSON-871253541 tempest-TenantUsagesTestJSON-871253541-project-member] Acquiring lock "refresh_cache-b6712878-3270-43f4-9518-d41eb77124b7" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 762.976345] env[61923]: DEBUG oslo_concurrency.lockutils [None req-ab7b3da0-3a01-44d4-9dfd-3e5d2ad27432 tempest-TenantUsagesTestJSON-871253541 tempest-TenantUsagesTestJSON-871253541-project-member] Acquired lock "refresh_cache-b6712878-3270-43f4-9518-d41eb77124b7" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 762.977042] env[61923]: DEBUG nova.network.neutron [None req-ab7b3da0-3a01-44d4-9dfd-3e5d2ad27432 tempest-TenantUsagesTestJSON-871253541 tempest-TenantUsagesTestJSON-871253541-project-member] [instance: b6712878-3270-43f4-9518-d41eb77124b7] Building network info cache for instance {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 763.285341] env[61923]: INFO nova.compute.manager [None req-0708225f-80bb-41ec-b3ed-7dd45d589b95 tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] [instance: 62d8a170-e7ad-4d14-90d8-6f6ce32b7cf8] Took 1.02 seconds to deallocate network for instance. [ 763.439672] env[61923]: DEBUG oslo_vmware.api [None req-347b83bf-b4ed-4904-9bd3-d6021183ae1c tempest-ServerShowV254Test-921228709 tempest-ServerShowV254Test-921228709-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52b17950-ae16-c968-afb0-720e3490fd80, 'name': SearchDatastore_Task, 'duration_secs': 0.008756} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 763.439883] env[61923]: DEBUG oslo_concurrency.lockutils [None req-347b83bf-b4ed-4904-9bd3-d6021183ae1c tempest-ServerShowV254Test-921228709 tempest-ServerShowV254Test-921228709-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 763.440500] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-347b83bf-b4ed-4904-9bd3-d6021183ae1c tempest-ServerShowV254Test-921228709 tempest-ServerShowV254Test-921228709-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk to [datastore2] ecedb77f-4a2a-4047-964f-80796954d847/ecedb77f-4a2a-4047-964f-80796954d847.vmdk {{(pid=61923) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 763.440667] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-3473af3b-e99e-4015-a1b0-ea806721da11 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.448501] env[61923]: DEBUG oslo_vmware.api [None req-347b83bf-b4ed-4904-9bd3-d6021183ae1c tempest-ServerShowV254Test-921228709 tempest-ServerShowV254Test-921228709-project-member] Waiting for the task: (returnval){ [ 763.448501] env[61923]: value = "task-1377494" [ 763.448501] env[61923]: _type = "Task" [ 763.448501] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 763.456863] env[61923]: DEBUG oslo_vmware.api [None req-347b83bf-b4ed-4904-9bd3-d6021183ae1c tempest-ServerShowV254Test-921228709 tempest-ServerShowV254Test-921228709-project-member] Task: {'id': task-1377494, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 763.498807] env[61923]: DEBUG nova.network.neutron [None req-ab7b3da0-3a01-44d4-9dfd-3e5d2ad27432 tempest-TenantUsagesTestJSON-871253541 tempest-TenantUsagesTestJSON-871253541-project-member] [instance: b6712878-3270-43f4-9518-d41eb77124b7] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 763.560033] env[61923]: DEBUG nova.network.neutron [None req-ab7b3da0-3a01-44d4-9dfd-3e5d2ad27432 tempest-TenantUsagesTestJSON-871253541 tempest-TenantUsagesTestJSON-871253541-project-member] [instance: b6712878-3270-43f4-9518-d41eb77124b7] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 763.757244] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8bc6ad4c-d67b-4f6a-859a-291dc468372d {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.767059] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-735307a2-6ac3-43db-8d0f-4f8cd1acdd79 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.802871] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b569bac0-0b80-41e0-8d7a-b8e39623d74d {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.812697] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5c07492-4c1a-4083-b5c2-a212b5fcd0ff {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 763.827605] env[61923]: DEBUG nova.compute.provider_tree [None req-5138192d-6349-432b-b7de-64ec5d391f8d tempest-ServerActionsTestOtherB-987045973 tempest-ServerActionsTestOtherB-987045973-project-member] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 763.959798] env[61923]: DEBUG oslo_vmware.api [None req-347b83bf-b4ed-4904-9bd3-d6021183ae1c tempest-ServerShowV254Test-921228709 tempest-ServerShowV254Test-921228709-project-member] Task: {'id': task-1377494, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 764.063332] env[61923]: DEBUG oslo_concurrency.lockutils [None req-ab7b3da0-3a01-44d4-9dfd-3e5d2ad27432 tempest-TenantUsagesTestJSON-871253541 tempest-TenantUsagesTestJSON-871253541-project-member] Releasing lock "refresh_cache-b6712878-3270-43f4-9518-d41eb77124b7" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 764.063604] env[61923]: DEBUG nova.compute.manager [None req-ab7b3da0-3a01-44d4-9dfd-3e5d2ad27432 tempest-TenantUsagesTestJSON-871253541 tempest-TenantUsagesTestJSON-871253541-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61923) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 764.063784] env[61923]: DEBUG nova.compute.manager [None req-ab7b3da0-3a01-44d4-9dfd-3e5d2ad27432 tempest-TenantUsagesTestJSON-871253541 tempest-TenantUsagesTestJSON-871253541-project-member] [instance: b6712878-3270-43f4-9518-d41eb77124b7] Deallocating network for instance {{(pid=61923) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 764.063986] env[61923]: DEBUG nova.network.neutron [None req-ab7b3da0-3a01-44d4-9dfd-3e5d2ad27432 tempest-TenantUsagesTestJSON-871253541 tempest-TenantUsagesTestJSON-871253541-project-member] [instance: b6712878-3270-43f4-9518-d41eb77124b7] deallocate_for_instance() {{(pid=61923) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 764.085627] env[61923]: DEBUG nova.network.neutron [None req-ab7b3da0-3a01-44d4-9dfd-3e5d2ad27432 tempest-TenantUsagesTestJSON-871253541 tempest-TenantUsagesTestJSON-871253541-project-member] [instance: b6712878-3270-43f4-9518-d41eb77124b7] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 764.327505] env[61923]: INFO nova.scheduler.client.report [None req-0708225f-80bb-41ec-b3ed-7dd45d589b95 tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Deleted allocations for instance 62d8a170-e7ad-4d14-90d8-6f6ce32b7cf8 [ 764.333668] env[61923]: DEBUG nova.scheduler.client.report [None req-5138192d-6349-432b-b7de-64ec5d391f8d tempest-ServerActionsTestOtherB-987045973 tempest-ServerActionsTestOtherB-987045973-project-member] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 764.459702] env[61923]: DEBUG oslo_vmware.api [None req-347b83bf-b4ed-4904-9bd3-d6021183ae1c tempest-ServerShowV254Test-921228709 tempest-ServerShowV254Test-921228709-project-member] Task: {'id': task-1377494, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.51223} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 764.459961] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-347b83bf-b4ed-4904-9bd3-d6021183ae1c tempest-ServerShowV254Test-921228709 tempest-ServerShowV254Test-921228709-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk to [datastore2] ecedb77f-4a2a-4047-964f-80796954d847/ecedb77f-4a2a-4047-964f-80796954d847.vmdk {{(pid=61923) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 764.460195] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-347b83bf-b4ed-4904-9bd3-d6021183ae1c tempest-ServerShowV254Test-921228709 tempest-ServerShowV254Test-921228709-project-member] [instance: ecedb77f-4a2a-4047-964f-80796954d847] Extending root virtual disk to 1048576 {{(pid=61923) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 764.460450] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ea142969-6b0a-45fe-9e22-78c8b43cca9f {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.466617] env[61923]: DEBUG oslo_vmware.api [None req-347b83bf-b4ed-4904-9bd3-d6021183ae1c tempest-ServerShowV254Test-921228709 tempest-ServerShowV254Test-921228709-project-member] Waiting for the task: (returnval){ [ 764.466617] env[61923]: value = "task-1377495" [ 764.466617] env[61923]: _type = "Task" [ 764.466617] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 764.473761] env[61923]: DEBUG oslo_vmware.api [None req-347b83bf-b4ed-4904-9bd3-d6021183ae1c tempest-ServerShowV254Test-921228709 tempest-ServerShowV254Test-921228709-project-member] Task: {'id': task-1377495, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 764.588157] env[61923]: DEBUG nova.network.neutron [None req-ab7b3da0-3a01-44d4-9dfd-3e5d2ad27432 tempest-TenantUsagesTestJSON-871253541 tempest-TenantUsagesTestJSON-871253541-project-member] [instance: b6712878-3270-43f4-9518-d41eb77124b7] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 764.839608] env[61923]: DEBUG oslo_concurrency.lockutils [None req-5138192d-6349-432b-b7de-64ec5d391f8d tempest-ServerActionsTestOtherB-987045973 tempest-ServerActionsTestOtherB-987045973-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.867s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 764.839812] env[61923]: ERROR nova.compute.manager [None req-5138192d-6349-432b-b7de-64ec5d391f8d tempest-ServerActionsTestOtherB-987045973 tempest-ServerActionsTestOtherB-987045973-project-member] [instance: b0bc08c1-38d5-45ee-b521-1cffda41c77c] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 129cb36c-cc2b-4205-9d7b-0b4d553b568b, please check neutron logs for more information. [ 764.839812] env[61923]: ERROR nova.compute.manager [instance: b0bc08c1-38d5-45ee-b521-1cffda41c77c] Traceback (most recent call last): [ 764.839812] env[61923]: ERROR nova.compute.manager [instance: b0bc08c1-38d5-45ee-b521-1cffda41c77c] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 764.839812] env[61923]: ERROR nova.compute.manager [instance: b0bc08c1-38d5-45ee-b521-1cffda41c77c] self.driver.spawn(context, instance, image_meta, [ 764.839812] env[61923]: ERROR nova.compute.manager [instance: b0bc08c1-38d5-45ee-b521-1cffda41c77c] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 764.839812] env[61923]: ERROR nova.compute.manager [instance: b0bc08c1-38d5-45ee-b521-1cffda41c77c] self._vmops.spawn(context, instance, image_meta, injected_files, [ 764.839812] env[61923]: ERROR nova.compute.manager [instance: b0bc08c1-38d5-45ee-b521-1cffda41c77c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 764.839812] env[61923]: ERROR nova.compute.manager [instance: b0bc08c1-38d5-45ee-b521-1cffda41c77c] vm_ref = self.build_virtual_machine(instance, [ 764.839812] env[61923]: ERROR nova.compute.manager [instance: b0bc08c1-38d5-45ee-b521-1cffda41c77c] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 764.839812] env[61923]: ERROR nova.compute.manager [instance: b0bc08c1-38d5-45ee-b521-1cffda41c77c] vif_infos = vmwarevif.get_vif_info(self._session, [ 764.839812] env[61923]: ERROR nova.compute.manager [instance: b0bc08c1-38d5-45ee-b521-1cffda41c77c] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 764.840233] env[61923]: ERROR nova.compute.manager [instance: b0bc08c1-38d5-45ee-b521-1cffda41c77c] for vif in network_info: [ 764.840233] env[61923]: ERROR nova.compute.manager [instance: b0bc08c1-38d5-45ee-b521-1cffda41c77c] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 764.840233] env[61923]: ERROR nova.compute.manager [instance: b0bc08c1-38d5-45ee-b521-1cffda41c77c] return self._sync_wrapper(fn, *args, **kwargs) [ 764.840233] env[61923]: ERROR nova.compute.manager [instance: b0bc08c1-38d5-45ee-b521-1cffda41c77c] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 764.840233] env[61923]: ERROR nova.compute.manager [instance: b0bc08c1-38d5-45ee-b521-1cffda41c77c] self.wait() [ 764.840233] env[61923]: ERROR nova.compute.manager [instance: b0bc08c1-38d5-45ee-b521-1cffda41c77c] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 764.840233] env[61923]: ERROR nova.compute.manager [instance: b0bc08c1-38d5-45ee-b521-1cffda41c77c] self[:] = self._gt.wait() [ 764.840233] env[61923]: ERROR nova.compute.manager [instance: b0bc08c1-38d5-45ee-b521-1cffda41c77c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 764.840233] env[61923]: ERROR nova.compute.manager [instance: b0bc08c1-38d5-45ee-b521-1cffda41c77c] return self._exit_event.wait() [ 764.840233] env[61923]: ERROR nova.compute.manager [instance: b0bc08c1-38d5-45ee-b521-1cffda41c77c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 764.840233] env[61923]: ERROR nova.compute.manager [instance: b0bc08c1-38d5-45ee-b521-1cffda41c77c] current.throw(*self._exc) [ 764.840233] env[61923]: ERROR nova.compute.manager [instance: b0bc08c1-38d5-45ee-b521-1cffda41c77c] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 764.840233] env[61923]: ERROR nova.compute.manager [instance: b0bc08c1-38d5-45ee-b521-1cffda41c77c] result = function(*args, **kwargs) [ 764.840746] env[61923]: ERROR nova.compute.manager [instance: b0bc08c1-38d5-45ee-b521-1cffda41c77c] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 764.840746] env[61923]: ERROR nova.compute.manager [instance: b0bc08c1-38d5-45ee-b521-1cffda41c77c] return func(*args, **kwargs) [ 764.840746] env[61923]: ERROR nova.compute.manager [instance: b0bc08c1-38d5-45ee-b521-1cffda41c77c] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 764.840746] env[61923]: ERROR nova.compute.manager [instance: b0bc08c1-38d5-45ee-b521-1cffda41c77c] raise e [ 764.840746] env[61923]: ERROR nova.compute.manager [instance: b0bc08c1-38d5-45ee-b521-1cffda41c77c] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 764.840746] env[61923]: ERROR nova.compute.manager [instance: b0bc08c1-38d5-45ee-b521-1cffda41c77c] nwinfo = self.network_api.allocate_for_instance( [ 764.840746] env[61923]: ERROR nova.compute.manager [instance: b0bc08c1-38d5-45ee-b521-1cffda41c77c] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 764.840746] env[61923]: ERROR nova.compute.manager [instance: b0bc08c1-38d5-45ee-b521-1cffda41c77c] created_port_ids = self._update_ports_for_instance( [ 764.840746] env[61923]: ERROR nova.compute.manager [instance: b0bc08c1-38d5-45ee-b521-1cffda41c77c] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 764.840746] env[61923]: ERROR nova.compute.manager [instance: b0bc08c1-38d5-45ee-b521-1cffda41c77c] with excutils.save_and_reraise_exception(): [ 764.840746] env[61923]: ERROR nova.compute.manager [instance: b0bc08c1-38d5-45ee-b521-1cffda41c77c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 764.840746] env[61923]: ERROR nova.compute.manager [instance: b0bc08c1-38d5-45ee-b521-1cffda41c77c] self.force_reraise() [ 764.840746] env[61923]: ERROR nova.compute.manager [instance: b0bc08c1-38d5-45ee-b521-1cffda41c77c] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 764.841186] env[61923]: ERROR nova.compute.manager [instance: b0bc08c1-38d5-45ee-b521-1cffda41c77c] raise self.value [ 764.841186] env[61923]: ERROR nova.compute.manager [instance: b0bc08c1-38d5-45ee-b521-1cffda41c77c] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 764.841186] env[61923]: ERROR nova.compute.manager [instance: b0bc08c1-38d5-45ee-b521-1cffda41c77c] updated_port = self._update_port( [ 764.841186] env[61923]: ERROR nova.compute.manager [instance: b0bc08c1-38d5-45ee-b521-1cffda41c77c] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 764.841186] env[61923]: ERROR nova.compute.manager [instance: b0bc08c1-38d5-45ee-b521-1cffda41c77c] _ensure_no_port_binding_failure(port) [ 764.841186] env[61923]: ERROR nova.compute.manager [instance: b0bc08c1-38d5-45ee-b521-1cffda41c77c] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 764.841186] env[61923]: ERROR nova.compute.manager [instance: b0bc08c1-38d5-45ee-b521-1cffda41c77c] raise exception.PortBindingFailed(port_id=port['id']) [ 764.841186] env[61923]: ERROR nova.compute.manager [instance: b0bc08c1-38d5-45ee-b521-1cffda41c77c] nova.exception.PortBindingFailed: Binding failed for port 129cb36c-cc2b-4205-9d7b-0b4d553b568b, please check neutron logs for more information. [ 764.841186] env[61923]: ERROR nova.compute.manager [instance: b0bc08c1-38d5-45ee-b521-1cffda41c77c] [ 764.841186] env[61923]: DEBUG nova.compute.utils [None req-5138192d-6349-432b-b7de-64ec5d391f8d tempest-ServerActionsTestOtherB-987045973 tempest-ServerActionsTestOtherB-987045973-project-member] [instance: b0bc08c1-38d5-45ee-b521-1cffda41c77c] Binding failed for port 129cb36c-cc2b-4205-9d7b-0b4d553b568b, please check neutron logs for more information. {{(pid=61923) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 764.841835] env[61923]: DEBUG oslo_concurrency.lockutils [None req-0708225f-80bb-41ec-b3ed-7dd45d589b95 tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Lock "62d8a170-e7ad-4d14-90d8-6f6ce32b7cf8" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 122.251s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 764.842277] env[61923]: DEBUG nova.compute.manager [None req-5138192d-6349-432b-b7de-64ec5d391f8d tempest-ServerActionsTestOtherB-987045973 tempest-ServerActionsTestOtherB-987045973-project-member] [instance: b0bc08c1-38d5-45ee-b521-1cffda41c77c] Build of instance b0bc08c1-38d5-45ee-b521-1cffda41c77c was re-scheduled: Binding failed for port 129cb36c-cc2b-4205-9d7b-0b4d553b568b, please check neutron logs for more information. {{(pid=61923) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 764.842711] env[61923]: DEBUG nova.compute.manager [None req-5138192d-6349-432b-b7de-64ec5d391f8d tempest-ServerActionsTestOtherB-987045973 tempest-ServerActionsTestOtherB-987045973-project-member] [instance: b0bc08c1-38d5-45ee-b521-1cffda41c77c] Unplugging VIFs for instance {{(pid=61923) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 764.842934] env[61923]: DEBUG oslo_concurrency.lockutils [None req-5138192d-6349-432b-b7de-64ec5d391f8d tempest-ServerActionsTestOtherB-987045973 tempest-ServerActionsTestOtherB-987045973-project-member] Acquiring lock "refresh_cache-b0bc08c1-38d5-45ee-b521-1cffda41c77c" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 764.843093] env[61923]: DEBUG oslo_concurrency.lockutils [None req-5138192d-6349-432b-b7de-64ec5d391f8d tempest-ServerActionsTestOtherB-987045973 tempest-ServerActionsTestOtherB-987045973-project-member] Acquired lock "refresh_cache-b0bc08c1-38d5-45ee-b521-1cffda41c77c" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 764.843415] env[61923]: DEBUG nova.network.neutron [None req-5138192d-6349-432b-b7de-64ec5d391f8d tempest-ServerActionsTestOtherB-987045973 tempest-ServerActionsTestOtherB-987045973-project-member] [instance: b0bc08c1-38d5-45ee-b521-1cffda41c77c] Building network info cache for instance {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 764.844158] env[61923]: DEBUG oslo_concurrency.lockutils [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 15.542s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 764.976019] env[61923]: DEBUG oslo_vmware.api [None req-347b83bf-b4ed-4904-9bd3-d6021183ae1c tempest-ServerShowV254Test-921228709 tempest-ServerShowV254Test-921228709-project-member] Task: {'id': task-1377495, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.061629} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 764.976413] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-347b83bf-b4ed-4904-9bd3-d6021183ae1c tempest-ServerShowV254Test-921228709 tempest-ServerShowV254Test-921228709-project-member] [instance: ecedb77f-4a2a-4047-964f-80796954d847] Extended root virtual disk {{(pid=61923) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 764.977233] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e83352f3-bcf7-41df-ace2-949d673de7e3 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.997750] env[61923]: DEBUG nova.virt.vmwareapi.volumeops [None req-347b83bf-b4ed-4904-9bd3-d6021183ae1c tempest-ServerShowV254Test-921228709 tempest-ServerShowV254Test-921228709-project-member] [instance: ecedb77f-4a2a-4047-964f-80796954d847] Reconfiguring VM instance instance-00000034 to attach disk [datastore2] ecedb77f-4a2a-4047-964f-80796954d847/ecedb77f-4a2a-4047-964f-80796954d847.vmdk or device None with type sparse {{(pid=61923) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 764.998485] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-69da7311-e7d4-490f-916c-b4271b658251 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.017396] env[61923]: DEBUG oslo_vmware.api [None req-347b83bf-b4ed-4904-9bd3-d6021183ae1c tempest-ServerShowV254Test-921228709 tempest-ServerShowV254Test-921228709-project-member] Waiting for the task: (returnval){ [ 765.017396] env[61923]: value = "task-1377496" [ 765.017396] env[61923]: _type = "Task" [ 765.017396] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 765.026412] env[61923]: DEBUG oslo_vmware.api [None req-347b83bf-b4ed-4904-9bd3-d6021183ae1c tempest-ServerShowV254Test-921228709 tempest-ServerShowV254Test-921228709-project-member] Task: {'id': task-1377496, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 765.091731] env[61923]: INFO nova.compute.manager [None req-ab7b3da0-3a01-44d4-9dfd-3e5d2ad27432 tempest-TenantUsagesTestJSON-871253541 tempest-TenantUsagesTestJSON-871253541-project-member] [instance: b6712878-3270-43f4-9518-d41eb77124b7] Took 1.03 seconds to deallocate network for instance. [ 765.352573] env[61923]: DEBUG nova.compute.manager [None req-a9ea97fa-71e5-4463-a1f3-ab69743d14a4 tempest-ServerActionsTestJSON-1295690556 tempest-ServerActionsTestJSON-1295690556-project-member] [instance: ad3b3933-5bd9-444a-af7b-c81e868e0622] Starting instance... {{(pid=61923) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 765.374626] env[61923]: DEBUG nova.network.neutron [None req-5138192d-6349-432b-b7de-64ec5d391f8d tempest-ServerActionsTestOtherB-987045973 tempest-ServerActionsTestOtherB-987045973-project-member] [instance: b0bc08c1-38d5-45ee-b521-1cffda41c77c] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 765.476438] env[61923]: DEBUG nova.network.neutron [None req-5138192d-6349-432b-b7de-64ec5d391f8d tempest-ServerActionsTestOtherB-987045973 tempest-ServerActionsTestOtherB-987045973-project-member] [instance: b0bc08c1-38d5-45ee-b521-1cffda41c77c] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 765.527765] env[61923]: DEBUG oslo_vmware.api [None req-347b83bf-b4ed-4904-9bd3-d6021183ae1c tempest-ServerShowV254Test-921228709 tempest-ServerShowV254Test-921228709-project-member] Task: {'id': task-1377496, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 765.843065] env[61923]: DEBUG oslo_concurrency.lockutils [None req-513daa64-2b05-4c0d-94e9-55b95de63eec tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Acquiring lock "59198f4d-4dde-4eaf-9f6c-a962cbe53c6e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 765.843295] env[61923]: DEBUG oslo_concurrency.lockutils [None req-513daa64-2b05-4c0d-94e9-55b95de63eec tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Lock "59198f4d-4dde-4eaf-9f6c-a962cbe53c6e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 765.874414] env[61923]: DEBUG oslo_concurrency.lockutils [None req-a9ea97fa-71e5-4463-a1f3-ab69743d14a4 tempest-ServerActionsTestJSON-1295690556 tempest-ServerActionsTestJSON-1295690556-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 765.979276] env[61923]: DEBUG oslo_concurrency.lockutils [None req-5138192d-6349-432b-b7de-64ec5d391f8d tempest-ServerActionsTestOtherB-987045973 tempest-ServerActionsTestOtherB-987045973-project-member] Releasing lock "refresh_cache-b0bc08c1-38d5-45ee-b521-1cffda41c77c" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 765.979519] env[61923]: DEBUG nova.compute.manager [None req-5138192d-6349-432b-b7de-64ec5d391f8d tempest-ServerActionsTestOtherB-987045973 tempest-ServerActionsTestOtherB-987045973-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61923) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 765.979722] env[61923]: DEBUG nova.compute.manager [None req-5138192d-6349-432b-b7de-64ec5d391f8d tempest-ServerActionsTestOtherB-987045973 tempest-ServerActionsTestOtherB-987045973-project-member] [instance: b0bc08c1-38d5-45ee-b521-1cffda41c77c] Deallocating network for instance {{(pid=61923) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 765.979891] env[61923]: DEBUG nova.network.neutron [None req-5138192d-6349-432b-b7de-64ec5d391f8d tempest-ServerActionsTestOtherB-987045973 tempest-ServerActionsTestOtherB-987045973-project-member] [instance: b0bc08c1-38d5-45ee-b521-1cffda41c77c] deallocate_for_instance() {{(pid=61923) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 765.995217] env[61923]: DEBUG nova.network.neutron [None req-5138192d-6349-432b-b7de-64ec5d391f8d tempest-ServerActionsTestOtherB-987045973 tempest-ServerActionsTestOtherB-987045973-project-member] [instance: b0bc08c1-38d5-45ee-b521-1cffda41c77c] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 766.027048] env[61923]: DEBUG oslo_vmware.api [None req-347b83bf-b4ed-4904-9bd3-d6021183ae1c tempest-ServerShowV254Test-921228709 tempest-ServerShowV254Test-921228709-project-member] Task: {'id': task-1377496, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 766.119359] env[61923]: INFO nova.scheduler.client.report [None req-ab7b3da0-3a01-44d4-9dfd-3e5d2ad27432 tempest-TenantUsagesTestJSON-871253541 tempest-TenantUsagesTestJSON-871253541-project-member] Deleted allocations for instance b6712878-3270-43f4-9518-d41eb77124b7 [ 766.380643] env[61923]: DEBUG nova.compute.resource_tracker [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Instance b6712878-3270-43f4-9518-d41eb77124b7 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61923) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 766.498114] env[61923]: DEBUG nova.network.neutron [None req-5138192d-6349-432b-b7de-64ec5d391f8d tempest-ServerActionsTestOtherB-987045973 tempest-ServerActionsTestOtherB-987045973-project-member] [instance: b0bc08c1-38d5-45ee-b521-1cffda41c77c] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 766.527921] env[61923]: DEBUG oslo_vmware.api [None req-347b83bf-b4ed-4904-9bd3-d6021183ae1c tempest-ServerShowV254Test-921228709 tempest-ServerShowV254Test-921228709-project-member] Task: {'id': task-1377496, 'name': ReconfigVM_Task, 'duration_secs': 1.27672} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 766.528270] env[61923]: DEBUG nova.virt.vmwareapi.volumeops [None req-347b83bf-b4ed-4904-9bd3-d6021183ae1c tempest-ServerShowV254Test-921228709 tempest-ServerShowV254Test-921228709-project-member] [instance: ecedb77f-4a2a-4047-964f-80796954d847] Reconfigured VM instance instance-00000034 to attach disk [datastore2] ecedb77f-4a2a-4047-964f-80796954d847/ecedb77f-4a2a-4047-964f-80796954d847.vmdk or device None with type sparse {{(pid=61923) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 766.528869] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e7b251ba-e95e-4d8c-b2b6-daf789dc6837 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.535467] env[61923]: DEBUG oslo_vmware.api [None req-347b83bf-b4ed-4904-9bd3-d6021183ae1c tempest-ServerShowV254Test-921228709 tempest-ServerShowV254Test-921228709-project-member] Waiting for the task: (returnval){ [ 766.535467] env[61923]: value = "task-1377497" [ 766.535467] env[61923]: _type = "Task" [ 766.535467] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 766.543346] env[61923]: DEBUG oslo_vmware.api [None req-347b83bf-b4ed-4904-9bd3-d6021183ae1c tempest-ServerShowV254Test-921228709 tempest-ServerShowV254Test-921228709-project-member] Task: {'id': task-1377497, 'name': Rename_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 766.626953] env[61923]: DEBUG oslo_concurrency.lockutils [None req-ab7b3da0-3a01-44d4-9dfd-3e5d2ad27432 tempest-TenantUsagesTestJSON-871253541 tempest-TenantUsagesTestJSON-871253541-project-member] Lock "b6712878-3270-43f4-9518-d41eb77124b7" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 110.567s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 766.884071] env[61923]: DEBUG nova.compute.resource_tracker [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Instance b0bc08c1-38d5-45ee-b521-1cffda41c77c has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61923) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 766.884248] env[61923]: DEBUG nova.compute.resource_tracker [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Instance 655b01ae-76a5-4c09-9245-e629fec8fd4b actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61923) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 766.884375] env[61923]: DEBUG nova.compute.resource_tracker [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Instance ecedb77f-4a2a-4047-964f-80796954d847 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61923) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 767.002036] env[61923]: INFO nova.compute.manager [None req-5138192d-6349-432b-b7de-64ec5d391f8d tempest-ServerActionsTestOtherB-987045973 tempest-ServerActionsTestOtherB-987045973-project-member] [instance: b0bc08c1-38d5-45ee-b521-1cffda41c77c] Took 1.02 seconds to deallocate network for instance. [ 767.045951] env[61923]: DEBUG oslo_vmware.api [None req-347b83bf-b4ed-4904-9bd3-d6021183ae1c tempest-ServerShowV254Test-921228709 tempest-ServerShowV254Test-921228709-project-member] Task: {'id': task-1377497, 'name': Rename_Task, 'duration_secs': 0.136799} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 767.046272] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-347b83bf-b4ed-4904-9bd3-d6021183ae1c tempest-ServerShowV254Test-921228709 tempest-ServerShowV254Test-921228709-project-member] [instance: ecedb77f-4a2a-4047-964f-80796954d847] Powering on the VM {{(pid=61923) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 767.046550] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5f54ab14-9a64-49ca-9c22-1386e8aa8e03 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.052790] env[61923]: DEBUG oslo_vmware.api [None req-347b83bf-b4ed-4904-9bd3-d6021183ae1c tempest-ServerShowV254Test-921228709 tempest-ServerShowV254Test-921228709-project-member] Waiting for the task: (returnval){ [ 767.052790] env[61923]: value = "task-1377498" [ 767.052790] env[61923]: _type = "Task" [ 767.052790] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 767.061388] env[61923]: DEBUG oslo_vmware.api [None req-347b83bf-b4ed-4904-9bd3-d6021183ae1c tempest-ServerShowV254Test-921228709 tempest-ServerShowV254Test-921228709-project-member] Task: {'id': task-1377498, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 767.130120] env[61923]: DEBUG nova.compute.manager [None req-0adbd41a-bb08-4b5a-bc10-ab9778329dd8 tempest-VolumesAdminNegativeTest-1422120048 tempest-VolumesAdminNegativeTest-1422120048-project-member] [instance: 7bd3204b-5754-4ff2-b0ff-f30556dd72a1] Starting instance... {{(pid=61923) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 767.387864] env[61923]: DEBUG nova.compute.resource_tracker [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Instance b497f3f3-001f-4e5e-b6b7-8e2006116856 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61923) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 767.562954] env[61923]: DEBUG oslo_vmware.api [None req-347b83bf-b4ed-4904-9bd3-d6021183ae1c tempest-ServerShowV254Test-921228709 tempest-ServerShowV254Test-921228709-project-member] Task: {'id': task-1377498, 'name': PowerOnVM_Task, 'duration_secs': 0.402497} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 767.562954] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-347b83bf-b4ed-4904-9bd3-d6021183ae1c tempest-ServerShowV254Test-921228709 tempest-ServerShowV254Test-921228709-project-member] [instance: ecedb77f-4a2a-4047-964f-80796954d847] Powered on the VM {{(pid=61923) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 767.563180] env[61923]: INFO nova.compute.manager [None req-347b83bf-b4ed-4904-9bd3-d6021183ae1c tempest-ServerShowV254Test-921228709 tempest-ServerShowV254Test-921228709-project-member] [instance: ecedb77f-4a2a-4047-964f-80796954d847] Took 6.29 seconds to spawn the instance on the hypervisor. [ 767.563364] env[61923]: DEBUG nova.compute.manager [None req-347b83bf-b4ed-4904-9bd3-d6021183ae1c tempest-ServerShowV254Test-921228709 tempest-ServerShowV254Test-921228709-project-member] [instance: ecedb77f-4a2a-4047-964f-80796954d847] Checking state {{(pid=61923) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 767.564330] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d96519e-7239-402b-a26d-5bc49efac0d3 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.622720] env[61923]: DEBUG nova.compute.manager [None req-1a9a2566-c1c5-4bb9-9861-1c50f6af4ac8 tempest-ServerDiagnosticsV248Test-889872339 tempest-ServerDiagnosticsV248Test-889872339-project-admin] [instance: 655b01ae-76a5-4c09-9245-e629fec8fd4b] Checking state {{(pid=61923) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 767.624026] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8026856-648a-4d75-9b0f-dbdcc77e1286 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.631165] env[61923]: INFO nova.compute.manager [None req-1a9a2566-c1c5-4bb9-9861-1c50f6af4ac8 tempest-ServerDiagnosticsV248Test-889872339 tempest-ServerDiagnosticsV248Test-889872339-project-admin] [instance: 655b01ae-76a5-4c09-9245-e629fec8fd4b] Retrieving diagnostics [ 767.632182] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8428781b-a490-416d-8a64-f4c9e02a7997 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.665091] env[61923]: DEBUG oslo_concurrency.lockutils [None req-0adbd41a-bb08-4b5a-bc10-ab9778329dd8 tempest-VolumesAdminNegativeTest-1422120048 tempest-VolumesAdminNegativeTest-1422120048-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 767.891309] env[61923]: DEBUG nova.compute.resource_tracker [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Instance faff83a1-3532-4396-ac3f-600f86653749 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61923) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 768.038205] env[61923]: INFO nova.scheduler.client.report [None req-5138192d-6349-432b-b7de-64ec5d391f8d tempest-ServerActionsTestOtherB-987045973 tempest-ServerActionsTestOtherB-987045973-project-member] Deleted allocations for instance b0bc08c1-38d5-45ee-b521-1cffda41c77c [ 768.079450] env[61923]: INFO nova.compute.manager [None req-347b83bf-b4ed-4904-9bd3-d6021183ae1c tempest-ServerShowV254Test-921228709 tempest-ServerShowV254Test-921228709-project-member] [instance: ecedb77f-4a2a-4047-964f-80796954d847] Took 25.73 seconds to build instance. [ 768.393943] env[61923]: DEBUG nova.compute.resource_tracker [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Instance 3f298e29-69ff-46ac-8018-ceb3990b9848 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61923) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 768.549913] env[61923]: DEBUG oslo_concurrency.lockutils [None req-5138192d-6349-432b-b7de-64ec5d391f8d tempest-ServerActionsTestOtherB-987045973 tempest-ServerActionsTestOtherB-987045973-project-member] Lock "b0bc08c1-38d5-45ee-b521-1cffda41c77c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 111.262s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 768.583137] env[61923]: DEBUG oslo_concurrency.lockutils [None req-347b83bf-b4ed-4904-9bd3-d6021183ae1c tempest-ServerShowV254Test-921228709 tempest-ServerShowV254Test-921228709-project-member] Lock "ecedb77f-4a2a-4047-964f-80796954d847" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 105.897s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 768.716009] env[61923]: DEBUG oslo_concurrency.lockutils [None req-f8f53256-e468-4f43-98b4-c171269c77f6 tempest-ServerDiagnosticsV248Test-1822646244 tempest-ServerDiagnosticsV248Test-1822646244-project-member] Acquiring lock "655b01ae-76a5-4c09-9245-e629fec8fd4b" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 768.716332] env[61923]: DEBUG oslo_concurrency.lockutils [None req-f8f53256-e468-4f43-98b4-c171269c77f6 tempest-ServerDiagnosticsV248Test-1822646244 tempest-ServerDiagnosticsV248Test-1822646244-project-member] Lock "655b01ae-76a5-4c09-9245-e629fec8fd4b" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 768.716591] env[61923]: DEBUG oslo_concurrency.lockutils [None req-f8f53256-e468-4f43-98b4-c171269c77f6 tempest-ServerDiagnosticsV248Test-1822646244 tempest-ServerDiagnosticsV248Test-1822646244-project-member] Acquiring lock "655b01ae-76a5-4c09-9245-e629fec8fd4b-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 768.716816] env[61923]: DEBUG oslo_concurrency.lockutils [None req-f8f53256-e468-4f43-98b4-c171269c77f6 tempest-ServerDiagnosticsV248Test-1822646244 tempest-ServerDiagnosticsV248Test-1822646244-project-member] Lock "655b01ae-76a5-4c09-9245-e629fec8fd4b-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 768.717050] env[61923]: DEBUG oslo_concurrency.lockutils [None req-f8f53256-e468-4f43-98b4-c171269c77f6 tempest-ServerDiagnosticsV248Test-1822646244 tempest-ServerDiagnosticsV248Test-1822646244-project-member] Lock "655b01ae-76a5-4c09-9245-e629fec8fd4b-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 768.720951] env[61923]: INFO nova.compute.manager [None req-f8f53256-e468-4f43-98b4-c171269c77f6 tempest-ServerDiagnosticsV248Test-1822646244 tempest-ServerDiagnosticsV248Test-1822646244-project-member] [instance: 655b01ae-76a5-4c09-9245-e629fec8fd4b] Terminating instance [ 768.724417] env[61923]: DEBUG oslo_concurrency.lockutils [None req-f8f53256-e468-4f43-98b4-c171269c77f6 tempest-ServerDiagnosticsV248Test-1822646244 tempest-ServerDiagnosticsV248Test-1822646244-project-member] Acquiring lock "refresh_cache-655b01ae-76a5-4c09-9245-e629fec8fd4b" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 768.724708] env[61923]: DEBUG oslo_concurrency.lockutils [None req-f8f53256-e468-4f43-98b4-c171269c77f6 tempest-ServerDiagnosticsV248Test-1822646244 tempest-ServerDiagnosticsV248Test-1822646244-project-member] Acquired lock "refresh_cache-655b01ae-76a5-4c09-9245-e629fec8fd4b" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 768.724980] env[61923]: DEBUG nova.network.neutron [None req-f8f53256-e468-4f43-98b4-c171269c77f6 tempest-ServerDiagnosticsV248Test-1822646244 tempest-ServerDiagnosticsV248Test-1822646244-project-member] [instance: 655b01ae-76a5-4c09-9245-e629fec8fd4b] Building network info cache for instance {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 768.899447] env[61923]: DEBUG nova.compute.resource_tracker [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Instance 6fe76341-d578-4cb3-9dce-c025f0d798b8 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61923) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 769.052130] env[61923]: DEBUG nova.compute.manager [None req-28282894-8344-49f7-8696-a5bcafc57c69 tempest-ServerRescueTestJSONUnderV235-1784422495 tempest-ServerRescueTestJSONUnderV235-1784422495-project-member] [instance: b2d2b34d-827f-4f96-af8e-936135ba8700] Starting instance... {{(pid=61923) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 769.086581] env[61923]: DEBUG nova.compute.manager [None req-998fcb75-b821-4586-ae48-2fe4d02a9864 tempest-AttachVolumeShelveTestJSON-1252849694 tempest-AttachVolumeShelveTestJSON-1252849694-project-member] [instance: 870602dd-1b85-4c97-9506-f08e9f1bade6] Starting instance... {{(pid=61923) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 769.115633] env[61923]: INFO nova.compute.manager [None req-6d07cbaf-b976-4d41-9fed-82d41c8e5c61 tempest-ServerShowV254Test-921228709 tempest-ServerShowV254Test-921228709-project-member] [instance: ecedb77f-4a2a-4047-964f-80796954d847] Rebuilding instance [ 769.168661] env[61923]: DEBUG nova.compute.manager [None req-6d07cbaf-b976-4d41-9fed-82d41c8e5c61 tempest-ServerShowV254Test-921228709 tempest-ServerShowV254Test-921228709-project-member] [instance: ecedb77f-4a2a-4047-964f-80796954d847] Checking state {{(pid=61923) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 769.169851] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5cab980a-f7cd-43ae-b5a4-2f38b1dc5e52 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.248031] env[61923]: DEBUG nova.network.neutron [None req-f8f53256-e468-4f43-98b4-c171269c77f6 tempest-ServerDiagnosticsV248Test-1822646244 tempest-ServerDiagnosticsV248Test-1822646244-project-member] [instance: 655b01ae-76a5-4c09-9245-e629fec8fd4b] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 769.343259] env[61923]: DEBUG nova.network.neutron [None req-f8f53256-e468-4f43-98b4-c171269c77f6 tempest-ServerDiagnosticsV248Test-1822646244 tempest-ServerDiagnosticsV248Test-1822646244-project-member] [instance: 655b01ae-76a5-4c09-9245-e629fec8fd4b] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 769.403916] env[61923]: DEBUG nova.compute.resource_tracker [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Instance 4d709a3a-65af-4adf-9ad0-4d830a9518c2 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61923) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 769.573993] env[61923]: DEBUG oslo_concurrency.lockutils [None req-28282894-8344-49f7-8696-a5bcafc57c69 tempest-ServerRescueTestJSONUnderV235-1784422495 tempest-ServerRescueTestJSONUnderV235-1784422495-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 769.604935] env[61923]: DEBUG oslo_concurrency.lockutils [None req-998fcb75-b821-4586-ae48-2fe4d02a9864 tempest-AttachVolumeShelveTestJSON-1252849694 tempest-AttachVolumeShelveTestJSON-1252849694-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 769.682088] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-6d07cbaf-b976-4d41-9fed-82d41c8e5c61 tempest-ServerShowV254Test-921228709 tempest-ServerShowV254Test-921228709-project-member] [instance: ecedb77f-4a2a-4047-964f-80796954d847] Powering off the VM {{(pid=61923) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 769.682458] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-346b5abb-311c-4587-9abf-0fd079873be9 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.694463] env[61923]: DEBUG oslo_vmware.api [None req-6d07cbaf-b976-4d41-9fed-82d41c8e5c61 tempest-ServerShowV254Test-921228709 tempest-ServerShowV254Test-921228709-project-member] Waiting for the task: (returnval){ [ 769.694463] env[61923]: value = "task-1377499" [ 769.694463] env[61923]: _type = "Task" [ 769.694463] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 769.704765] env[61923]: DEBUG oslo_vmware.api [None req-6d07cbaf-b976-4d41-9fed-82d41c8e5c61 tempest-ServerShowV254Test-921228709 tempest-ServerShowV254Test-921228709-project-member] Task: {'id': task-1377499, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 769.847380] env[61923]: DEBUG oslo_concurrency.lockutils [None req-f8f53256-e468-4f43-98b4-c171269c77f6 tempest-ServerDiagnosticsV248Test-1822646244 tempest-ServerDiagnosticsV248Test-1822646244-project-member] Releasing lock "refresh_cache-655b01ae-76a5-4c09-9245-e629fec8fd4b" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 769.847820] env[61923]: DEBUG nova.compute.manager [None req-f8f53256-e468-4f43-98b4-c171269c77f6 tempest-ServerDiagnosticsV248Test-1822646244 tempest-ServerDiagnosticsV248Test-1822646244-project-member] [instance: 655b01ae-76a5-4c09-9245-e629fec8fd4b] Start destroying the instance on the hypervisor. {{(pid=61923) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 769.848021] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-f8f53256-e468-4f43-98b4-c171269c77f6 tempest-ServerDiagnosticsV248Test-1822646244 tempest-ServerDiagnosticsV248Test-1822646244-project-member] [instance: 655b01ae-76a5-4c09-9245-e629fec8fd4b] Destroying instance {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 769.848922] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ce9bb7c-3a4e-4fcc-92cc-6e264c4ec64a {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.858769] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-f8f53256-e468-4f43-98b4-c171269c77f6 tempest-ServerDiagnosticsV248Test-1822646244 tempest-ServerDiagnosticsV248Test-1822646244-project-member] [instance: 655b01ae-76a5-4c09-9245-e629fec8fd4b] Powering off the VM {{(pid=61923) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 769.859257] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-767660fa-9329-4c49-8497-25b87d0bf279 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.866834] env[61923]: DEBUG oslo_vmware.api [None req-f8f53256-e468-4f43-98b4-c171269c77f6 tempest-ServerDiagnosticsV248Test-1822646244 tempest-ServerDiagnosticsV248Test-1822646244-project-member] Waiting for the task: (returnval){ [ 769.866834] env[61923]: value = "task-1377500" [ 769.866834] env[61923]: _type = "Task" [ 769.866834] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 769.876756] env[61923]: DEBUG oslo_vmware.api [None req-f8f53256-e468-4f43-98b4-c171269c77f6 tempest-ServerDiagnosticsV248Test-1822646244 tempest-ServerDiagnosticsV248Test-1822646244-project-member] Task: {'id': task-1377500, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 769.907749] env[61923]: DEBUG nova.compute.resource_tracker [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Instance fe981195-1e2c-4c8d-8c53-327851f845c7 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61923) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 770.205564] env[61923]: DEBUG oslo_vmware.api [None req-6d07cbaf-b976-4d41-9fed-82d41c8e5c61 tempest-ServerShowV254Test-921228709 tempest-ServerShowV254Test-921228709-project-member] Task: {'id': task-1377499, 'name': PowerOffVM_Task, 'duration_secs': 0.193611} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 770.205817] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-6d07cbaf-b976-4d41-9fed-82d41c8e5c61 tempest-ServerShowV254Test-921228709 tempest-ServerShowV254Test-921228709-project-member] [instance: ecedb77f-4a2a-4047-964f-80796954d847] Powered off the VM {{(pid=61923) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 770.206039] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-6d07cbaf-b976-4d41-9fed-82d41c8e5c61 tempest-ServerShowV254Test-921228709 tempest-ServerShowV254Test-921228709-project-member] [instance: ecedb77f-4a2a-4047-964f-80796954d847] Destroying instance {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 770.206850] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52efe1b9-006d-42f5-b23f-08806985fa86 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.213998] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-6d07cbaf-b976-4d41-9fed-82d41c8e5c61 tempest-ServerShowV254Test-921228709 tempest-ServerShowV254Test-921228709-project-member] [instance: ecedb77f-4a2a-4047-964f-80796954d847] Unregistering the VM {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 770.214634] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-78512f7b-72a8-4a02-9a7b-b1448cd8f257 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.244019] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-6d07cbaf-b976-4d41-9fed-82d41c8e5c61 tempest-ServerShowV254Test-921228709 tempest-ServerShowV254Test-921228709-project-member] [instance: ecedb77f-4a2a-4047-964f-80796954d847] Unregistered the VM {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 770.244019] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-6d07cbaf-b976-4d41-9fed-82d41c8e5c61 tempest-ServerShowV254Test-921228709 tempest-ServerShowV254Test-921228709-project-member] [instance: ecedb77f-4a2a-4047-964f-80796954d847] Deleting contents of the VM from datastore datastore2 {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 770.244019] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-6d07cbaf-b976-4d41-9fed-82d41c8e5c61 tempest-ServerShowV254Test-921228709 tempest-ServerShowV254Test-921228709-project-member] Deleting the datastore file [datastore2] ecedb77f-4a2a-4047-964f-80796954d847 {{(pid=61923) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 770.244019] env[61923]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-de0b28fc-1b63-4978-9fc5-3283f13927d1 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.255663] env[61923]: DEBUG oslo_vmware.api [None req-6d07cbaf-b976-4d41-9fed-82d41c8e5c61 tempest-ServerShowV254Test-921228709 tempest-ServerShowV254Test-921228709-project-member] Waiting for the task: (returnval){ [ 770.255663] env[61923]: value = "task-1377502" [ 770.255663] env[61923]: _type = "Task" [ 770.255663] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 770.262910] env[61923]: DEBUG oslo_vmware.api [None req-6d07cbaf-b976-4d41-9fed-82d41c8e5c61 tempest-ServerShowV254Test-921228709 tempest-ServerShowV254Test-921228709-project-member] Task: {'id': task-1377502, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 770.376719] env[61923]: DEBUG oslo_vmware.api [None req-f8f53256-e468-4f43-98b4-c171269c77f6 tempest-ServerDiagnosticsV248Test-1822646244 tempest-ServerDiagnosticsV248Test-1822646244-project-member] Task: {'id': task-1377500, 'name': PowerOffVM_Task, 'duration_secs': 0.191257} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 770.377024] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-f8f53256-e468-4f43-98b4-c171269c77f6 tempest-ServerDiagnosticsV248Test-1822646244 tempest-ServerDiagnosticsV248Test-1822646244-project-member] [instance: 655b01ae-76a5-4c09-9245-e629fec8fd4b] Powered off the VM {{(pid=61923) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 770.377199] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-f8f53256-e468-4f43-98b4-c171269c77f6 tempest-ServerDiagnosticsV248Test-1822646244 tempest-ServerDiagnosticsV248Test-1822646244-project-member] [instance: 655b01ae-76a5-4c09-9245-e629fec8fd4b] Unregistering the VM {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 770.377446] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-5d2f6c17-aba0-43ac-b712-29fcd853a2d0 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.400907] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-f8f53256-e468-4f43-98b4-c171269c77f6 tempest-ServerDiagnosticsV248Test-1822646244 tempest-ServerDiagnosticsV248Test-1822646244-project-member] [instance: 655b01ae-76a5-4c09-9245-e629fec8fd4b] Unregistered the VM {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 770.401143] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-f8f53256-e468-4f43-98b4-c171269c77f6 tempest-ServerDiagnosticsV248Test-1822646244 tempest-ServerDiagnosticsV248Test-1822646244-project-member] [instance: 655b01ae-76a5-4c09-9245-e629fec8fd4b] Deleting contents of the VM from datastore datastore1 {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 770.401750] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-f8f53256-e468-4f43-98b4-c171269c77f6 tempest-ServerDiagnosticsV248Test-1822646244 tempest-ServerDiagnosticsV248Test-1822646244-project-member] Deleting the datastore file [datastore1] 655b01ae-76a5-4c09-9245-e629fec8fd4b {{(pid=61923) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 770.402920] env[61923]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-2d41cc22-fdeb-4bab-9a13-249428812bf8 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.410233] env[61923]: DEBUG oslo_vmware.api [None req-f8f53256-e468-4f43-98b4-c171269c77f6 tempest-ServerDiagnosticsV248Test-1822646244 tempest-ServerDiagnosticsV248Test-1822646244-project-member] Waiting for the task: (returnval){ [ 770.410233] env[61923]: value = "task-1377504" [ 770.410233] env[61923]: _type = "Task" [ 770.410233] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 770.413563] env[61923]: DEBUG nova.compute.resource_tracker [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Instance ad3b3933-5bd9-444a-af7b-c81e868e0622 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61923) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 770.424192] env[61923]: DEBUG oslo_vmware.api [None req-f8f53256-e468-4f43-98b4-c171269c77f6 tempest-ServerDiagnosticsV248Test-1822646244 tempest-ServerDiagnosticsV248Test-1822646244-project-member] Task: {'id': task-1377504, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 770.768240] env[61923]: DEBUG oslo_vmware.api [None req-6d07cbaf-b976-4d41-9fed-82d41c8e5c61 tempest-ServerShowV254Test-921228709 tempest-ServerShowV254Test-921228709-project-member] Task: {'id': task-1377502, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.095287} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 770.768644] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-6d07cbaf-b976-4d41-9fed-82d41c8e5c61 tempest-ServerShowV254Test-921228709 tempest-ServerShowV254Test-921228709-project-member] Deleted the datastore file {{(pid=61923) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 770.769034] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-6d07cbaf-b976-4d41-9fed-82d41c8e5c61 tempest-ServerShowV254Test-921228709 tempest-ServerShowV254Test-921228709-project-member] [instance: ecedb77f-4a2a-4047-964f-80796954d847] Deleted contents of the VM from datastore datastore2 {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 770.769347] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-6d07cbaf-b976-4d41-9fed-82d41c8e5c61 tempest-ServerShowV254Test-921228709 tempest-ServerShowV254Test-921228709-project-member] [instance: ecedb77f-4a2a-4047-964f-80796954d847] Instance destroyed {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 770.920422] env[61923]: DEBUG nova.compute.resource_tracker [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Instance 7bd3204b-5754-4ff2-b0ff-f30556dd72a1 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61923) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 770.921869] env[61923]: DEBUG oslo_vmware.api [None req-f8f53256-e468-4f43-98b4-c171269c77f6 tempest-ServerDiagnosticsV248Test-1822646244 tempest-ServerDiagnosticsV248Test-1822646244-project-member] Task: {'id': task-1377504, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.111152} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 770.922557] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-f8f53256-e468-4f43-98b4-c171269c77f6 tempest-ServerDiagnosticsV248Test-1822646244 tempest-ServerDiagnosticsV248Test-1822646244-project-member] Deleted the datastore file {{(pid=61923) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 770.923040] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-f8f53256-e468-4f43-98b4-c171269c77f6 tempest-ServerDiagnosticsV248Test-1822646244 tempest-ServerDiagnosticsV248Test-1822646244-project-member] [instance: 655b01ae-76a5-4c09-9245-e629fec8fd4b] Deleted contents of the VM from datastore datastore1 {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 770.923040] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-f8f53256-e468-4f43-98b4-c171269c77f6 tempest-ServerDiagnosticsV248Test-1822646244 tempest-ServerDiagnosticsV248Test-1822646244-project-member] [instance: 655b01ae-76a5-4c09-9245-e629fec8fd4b] Instance destroyed {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 770.926295] env[61923]: INFO nova.compute.manager [None req-f8f53256-e468-4f43-98b4-c171269c77f6 tempest-ServerDiagnosticsV248Test-1822646244 tempest-ServerDiagnosticsV248Test-1822646244-project-member] [instance: 655b01ae-76a5-4c09-9245-e629fec8fd4b] Took 1.08 seconds to destroy the instance on the hypervisor. [ 770.926295] env[61923]: DEBUG oslo.service.loopingcall [None req-f8f53256-e468-4f43-98b4-c171269c77f6 tempest-ServerDiagnosticsV248Test-1822646244 tempest-ServerDiagnosticsV248Test-1822646244-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61923) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 770.926295] env[61923]: DEBUG nova.compute.manager [-] [instance: 655b01ae-76a5-4c09-9245-e629fec8fd4b] Deallocating network for instance {{(pid=61923) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 770.926295] env[61923]: DEBUG nova.network.neutron [-] [instance: 655b01ae-76a5-4c09-9245-e629fec8fd4b] deallocate_for_instance() {{(pid=61923) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 770.946785] env[61923]: DEBUG nova.network.neutron [-] [instance: 655b01ae-76a5-4c09-9245-e629fec8fd4b] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 771.423837] env[61923]: DEBUG nova.compute.resource_tracker [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Instance b2d2b34d-827f-4f96-af8e-936135ba8700 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61923) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 771.449499] env[61923]: DEBUG nova.network.neutron [-] [instance: 655b01ae-76a5-4c09-9245-e629fec8fd4b] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 771.807062] env[61923]: DEBUG nova.virt.hardware [None req-6d07cbaf-b976-4d41-9fed-82d41c8e5c61 tempest-ServerShowV254Test-921228709 tempest-ServerShowV254Test-921228709-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-29T20:07:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-29T20:07:35Z,direct_url=,disk_format='vmdk',id=0f153f63-ae0a-45b1-b7f5-7f9b673c947f,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='4e7b5e3b3c3443c8bd5c70f8a15739f5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-29T20:07:36Z,virtual_size=,visibility=), allow threads: False {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 771.807324] env[61923]: DEBUG nova.virt.hardware [None req-6d07cbaf-b976-4d41-9fed-82d41c8e5c61 tempest-ServerShowV254Test-921228709 tempest-ServerShowV254Test-921228709-project-member] Flavor limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 771.807479] env[61923]: DEBUG nova.virt.hardware [None req-6d07cbaf-b976-4d41-9fed-82d41c8e5c61 tempest-ServerShowV254Test-921228709 tempest-ServerShowV254Test-921228709-project-member] Image limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 771.807656] env[61923]: DEBUG nova.virt.hardware [None req-6d07cbaf-b976-4d41-9fed-82d41c8e5c61 tempest-ServerShowV254Test-921228709 tempest-ServerShowV254Test-921228709-project-member] Flavor pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 771.807813] env[61923]: DEBUG nova.virt.hardware [None req-6d07cbaf-b976-4d41-9fed-82d41c8e5c61 tempest-ServerShowV254Test-921228709 tempest-ServerShowV254Test-921228709-project-member] Image pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 771.807957] env[61923]: DEBUG nova.virt.hardware [None req-6d07cbaf-b976-4d41-9fed-82d41c8e5c61 tempest-ServerShowV254Test-921228709 tempest-ServerShowV254Test-921228709-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 771.808380] env[61923]: DEBUG nova.virt.hardware [None req-6d07cbaf-b976-4d41-9fed-82d41c8e5c61 tempest-ServerShowV254Test-921228709 tempest-ServerShowV254Test-921228709-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 771.810942] env[61923]: DEBUG nova.virt.hardware [None req-6d07cbaf-b976-4d41-9fed-82d41c8e5c61 tempest-ServerShowV254Test-921228709 tempest-ServerShowV254Test-921228709-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 771.811244] env[61923]: DEBUG nova.virt.hardware [None req-6d07cbaf-b976-4d41-9fed-82d41c8e5c61 tempest-ServerShowV254Test-921228709 tempest-ServerShowV254Test-921228709-project-member] Got 1 possible topologies {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 771.811467] env[61923]: DEBUG nova.virt.hardware [None req-6d07cbaf-b976-4d41-9fed-82d41c8e5c61 tempest-ServerShowV254Test-921228709 tempest-ServerShowV254Test-921228709-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 771.811680] env[61923]: DEBUG nova.virt.hardware [None req-6d07cbaf-b976-4d41-9fed-82d41c8e5c61 tempest-ServerShowV254Test-921228709 tempest-ServerShowV254Test-921228709-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 771.812573] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a83a71af-5489-474c-aa07-b14c96e6dcd7 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.820893] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20546b34-af53-4e4e-9936-581d0df2f124 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.834088] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-6d07cbaf-b976-4d41-9fed-82d41c8e5c61 tempest-ServerShowV254Test-921228709 tempest-ServerShowV254Test-921228709-project-member] [instance: ecedb77f-4a2a-4047-964f-80796954d847] Instance VIF info [] {{(pid=61923) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 771.839869] env[61923]: DEBUG oslo.service.loopingcall [None req-6d07cbaf-b976-4d41-9fed-82d41c8e5c61 tempest-ServerShowV254Test-921228709 tempest-ServerShowV254Test-921228709-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61923) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 771.840165] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ecedb77f-4a2a-4047-964f-80796954d847] Creating VM on the ESX host {{(pid=61923) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 771.840416] env[61923]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-bf7f6e20-ee29-4b8a-a9e9-391724298f7b {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.860082] env[61923]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 771.860082] env[61923]: value = "task-1377505" [ 771.860082] env[61923]: _type = "Task" [ 771.860082] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 771.870574] env[61923]: DEBUG oslo_vmware.api [-] Task: {'id': task-1377505, 'name': CreateVM_Task} progress is 5%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 771.931316] env[61923]: DEBUG nova.compute.resource_tracker [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Instance 870602dd-1b85-4c97-9506-f08e9f1bade6 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61923) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 771.953052] env[61923]: INFO nova.compute.manager [-] [instance: 655b01ae-76a5-4c09-9245-e629fec8fd4b] Took 1.03 seconds to deallocate network for instance. [ 772.123330] env[61923]: DEBUG oslo_concurrency.lockutils [None req-80740a2c-9779-4afc-be30-c4469cd3724d tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] Acquiring lock "fc4fdf89-4615-4586-9dcd-ea90dc258361" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 772.123598] env[61923]: DEBUG oslo_concurrency.lockutils [None req-80740a2c-9779-4afc-be30-c4469cd3724d tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] Lock "fc4fdf89-4615-4586-9dcd-ea90dc258361" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 772.370570] env[61923]: DEBUG oslo_vmware.api [-] Task: {'id': task-1377505, 'name': CreateVM_Task, 'duration_secs': 0.240027} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 772.370752] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ecedb77f-4a2a-4047-964f-80796954d847] Created VM on the ESX host {{(pid=61923) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 772.371184] env[61923]: DEBUG oslo_concurrency.lockutils [None req-6d07cbaf-b976-4d41-9fed-82d41c8e5c61 tempest-ServerShowV254Test-921228709 tempest-ServerShowV254Test-921228709-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 772.371346] env[61923]: DEBUG oslo_concurrency.lockutils [None req-6d07cbaf-b976-4d41-9fed-82d41c8e5c61 tempest-ServerShowV254Test-921228709 tempest-ServerShowV254Test-921228709-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 772.371671] env[61923]: DEBUG oslo_concurrency.lockutils [None req-6d07cbaf-b976-4d41-9fed-82d41c8e5c61 tempest-ServerShowV254Test-921228709 tempest-ServerShowV254Test-921228709-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 772.371924] env[61923]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-659336c6-9da4-4ced-828f-25a1e650559c {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.376486] env[61923]: DEBUG oslo_vmware.api [None req-6d07cbaf-b976-4d41-9fed-82d41c8e5c61 tempest-ServerShowV254Test-921228709 tempest-ServerShowV254Test-921228709-project-member] Waiting for the task: (returnval){ [ 772.376486] env[61923]: value = "session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]5213f02c-4ab9-9441-3589-d8d82c3db963" [ 772.376486] env[61923]: _type = "Task" [ 772.376486] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 772.384255] env[61923]: DEBUG oslo_vmware.api [None req-6d07cbaf-b976-4d41-9fed-82d41c8e5c61 tempest-ServerShowV254Test-921228709 tempest-ServerShowV254Test-921228709-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]5213f02c-4ab9-9441-3589-d8d82c3db963, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 772.433282] env[61923]: DEBUG nova.compute.resource_tracker [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Instance b8c60af7-3561-468c-a9a5-1c645080fc69 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61923) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 772.459672] env[61923]: DEBUG oslo_concurrency.lockutils [None req-f8f53256-e468-4f43-98b4-c171269c77f6 tempest-ServerDiagnosticsV248Test-1822646244 tempest-ServerDiagnosticsV248Test-1822646244-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 772.887638] env[61923]: DEBUG oslo_vmware.api [None req-6d07cbaf-b976-4d41-9fed-82d41c8e5c61 tempest-ServerShowV254Test-921228709 tempest-ServerShowV254Test-921228709-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]5213f02c-4ab9-9441-3589-d8d82c3db963, 'name': SearchDatastore_Task, 'duration_secs': 0.008981} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 772.887960] env[61923]: DEBUG oslo_concurrency.lockutils [None req-6d07cbaf-b976-4d41-9fed-82d41c8e5c61 tempest-ServerShowV254Test-921228709 tempest-ServerShowV254Test-921228709-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 772.888202] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-6d07cbaf-b976-4d41-9fed-82d41c8e5c61 tempest-ServerShowV254Test-921228709 tempest-ServerShowV254Test-921228709-project-member] [instance: ecedb77f-4a2a-4047-964f-80796954d847] Processing image 0f153f63-ae0a-45b1-b7f5-7f9b673c947f {{(pid=61923) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 772.888466] env[61923]: DEBUG oslo_concurrency.lockutils [None req-6d07cbaf-b976-4d41-9fed-82d41c8e5c61 tempest-ServerShowV254Test-921228709 tempest-ServerShowV254Test-921228709-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 772.888608] env[61923]: DEBUG oslo_concurrency.lockutils [None req-6d07cbaf-b976-4d41-9fed-82d41c8e5c61 tempest-ServerShowV254Test-921228709 tempest-ServerShowV254Test-921228709-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 772.888783] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-6d07cbaf-b976-4d41-9fed-82d41c8e5c61 tempest-ServerShowV254Test-921228709 tempest-ServerShowV254Test-921228709-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61923) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 772.889072] env[61923]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-15a0bcaa-20ba-44e4-9ebc-94a10d47f736 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.898244] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-6d07cbaf-b976-4d41-9fed-82d41c8e5c61 tempest-ServerShowV254Test-921228709 tempest-ServerShowV254Test-921228709-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61923) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 772.898244] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-6d07cbaf-b976-4d41-9fed-82d41c8e5c61 tempest-ServerShowV254Test-921228709 tempest-ServerShowV254Test-921228709-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61923) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 772.898244] env[61923]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-19bd7e70-42ad-4a78-a9d5-e13097bdf82a {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.903093] env[61923]: DEBUG oslo_vmware.api [None req-6d07cbaf-b976-4d41-9fed-82d41c8e5c61 tempest-ServerShowV254Test-921228709 tempest-ServerShowV254Test-921228709-project-member] Waiting for the task: (returnval){ [ 772.903093] env[61923]: value = "session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52b3cff4-e18e-27ba-6803-2f0c3241feb1" [ 772.903093] env[61923]: _type = "Task" [ 772.903093] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 772.910378] env[61923]: DEBUG oslo_vmware.api [None req-6d07cbaf-b976-4d41-9fed-82d41c8e5c61 tempest-ServerShowV254Test-921228709 tempest-ServerShowV254Test-921228709-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52b3cff4-e18e-27ba-6803-2f0c3241feb1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 772.936225] env[61923]: DEBUG nova.compute.resource_tracker [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Instance 1d1b1dc4-c701-42b2-beba-344d622aef41 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61923) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 773.413976] env[61923]: DEBUG oslo_vmware.api [None req-6d07cbaf-b976-4d41-9fed-82d41c8e5c61 tempest-ServerShowV254Test-921228709 tempest-ServerShowV254Test-921228709-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52b3cff4-e18e-27ba-6803-2f0c3241feb1, 'name': SearchDatastore_Task, 'duration_secs': 0.007977} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 773.414398] env[61923]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-dae3649b-51c9-4f4a-bdfe-33b25160b6cd {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.419262] env[61923]: DEBUG oslo_vmware.api [None req-6d07cbaf-b976-4d41-9fed-82d41c8e5c61 tempest-ServerShowV254Test-921228709 tempest-ServerShowV254Test-921228709-project-member] Waiting for the task: (returnval){ [ 773.419262] env[61923]: value = "session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52ce7917-10ae-20de-9556-8971dfb4e3d7" [ 773.419262] env[61923]: _type = "Task" [ 773.419262] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 773.426932] env[61923]: DEBUG oslo_vmware.api [None req-6d07cbaf-b976-4d41-9fed-82d41c8e5c61 tempest-ServerShowV254Test-921228709 tempest-ServerShowV254Test-921228709-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52ce7917-10ae-20de-9556-8971dfb4e3d7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 773.439999] env[61923]: DEBUG nova.compute.resource_tracker [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Instance f1839f61-4314-48fe-8ab6-14b5e49d569d has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61923) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 773.929946] env[61923]: DEBUG oslo_vmware.api [None req-6d07cbaf-b976-4d41-9fed-82d41c8e5c61 tempest-ServerShowV254Test-921228709 tempest-ServerShowV254Test-921228709-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52ce7917-10ae-20de-9556-8971dfb4e3d7, 'name': SearchDatastore_Task, 'duration_secs': 0.008101} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 773.930160] env[61923]: DEBUG oslo_concurrency.lockutils [None req-6d07cbaf-b976-4d41-9fed-82d41c8e5c61 tempest-ServerShowV254Test-921228709 tempest-ServerShowV254Test-921228709-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 773.930413] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-6d07cbaf-b976-4d41-9fed-82d41c8e5c61 tempest-ServerShowV254Test-921228709 tempest-ServerShowV254Test-921228709-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk to [datastore1] ecedb77f-4a2a-4047-964f-80796954d847/ecedb77f-4a2a-4047-964f-80796954d847.vmdk {{(pid=61923) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 773.930663] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-937e8475-c63b-47d4-8b4b-902e37a2aa97 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.937026] env[61923]: DEBUG oslo_vmware.api [None req-6d07cbaf-b976-4d41-9fed-82d41c8e5c61 tempest-ServerShowV254Test-921228709 tempest-ServerShowV254Test-921228709-project-member] Waiting for the task: (returnval){ [ 773.937026] env[61923]: value = "task-1377506" [ 773.937026] env[61923]: _type = "Task" [ 773.937026] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 773.944955] env[61923]: DEBUG nova.compute.resource_tracker [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Instance 48825592-dcdc-41d5-9fbf-500d1f31017b has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61923) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 773.945946] env[61923]: DEBUG oslo_vmware.api [None req-6d07cbaf-b976-4d41-9fed-82d41c8e5c61 tempest-ServerShowV254Test-921228709 tempest-ServerShowV254Test-921228709-project-member] Task: {'id': task-1377506, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 774.449412] env[61923]: DEBUG nova.compute.resource_tracker [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Instance 422d06d6-6932-46c3-bb25-841e6f65c028 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61923) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 774.453013] env[61923]: DEBUG oslo_vmware.api [None req-6d07cbaf-b976-4d41-9fed-82d41c8e5c61 tempest-ServerShowV254Test-921228709 tempest-ServerShowV254Test-921228709-project-member] Task: {'id': task-1377506, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 774.947739] env[61923]: DEBUG oslo_vmware.api [None req-6d07cbaf-b976-4d41-9fed-82d41c8e5c61 tempest-ServerShowV254Test-921228709 tempest-ServerShowV254Test-921228709-project-member] Task: {'id': task-1377506, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.690439} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 774.947998] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-6d07cbaf-b976-4d41-9fed-82d41c8e5c61 tempest-ServerShowV254Test-921228709 tempest-ServerShowV254Test-921228709-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk to [datastore1] ecedb77f-4a2a-4047-964f-80796954d847/ecedb77f-4a2a-4047-964f-80796954d847.vmdk {{(pid=61923) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 774.948218] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-6d07cbaf-b976-4d41-9fed-82d41c8e5c61 tempest-ServerShowV254Test-921228709 tempest-ServerShowV254Test-921228709-project-member] [instance: ecedb77f-4a2a-4047-964f-80796954d847] Extending root virtual disk to 1048576 {{(pid=61923) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 774.948478] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-83885aa1-da68-4129-b774-fcccabfba988 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.952944] env[61923]: DEBUG nova.compute.resource_tracker [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Instance 73692517-1816-4e9b-ab2d-8265c683e83d has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61923) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 774.956057] env[61923]: DEBUG oslo_vmware.api [None req-6d07cbaf-b976-4d41-9fed-82d41c8e5c61 tempest-ServerShowV254Test-921228709 tempest-ServerShowV254Test-921228709-project-member] Waiting for the task: (returnval){ [ 774.956057] env[61923]: value = "task-1377507" [ 774.956057] env[61923]: _type = "Task" [ 774.956057] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 774.963473] env[61923]: DEBUG oslo_vmware.api [None req-6d07cbaf-b976-4d41-9fed-82d41c8e5c61 tempest-ServerShowV254Test-921228709 tempest-ServerShowV254Test-921228709-project-member] Task: {'id': task-1377507, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 775.456484] env[61923]: DEBUG nova.compute.resource_tracker [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Instance 59198f4d-4dde-4eaf-9f6c-a962cbe53c6e has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61923) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 775.456808] env[61923]: DEBUG nova.compute.resource_tracker [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Total usable vcpus: 48, total allocated vcpus: 2 {{(pid=61923) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 775.456917] env[61923]: DEBUG nova.compute.resource_tracker [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=896MB phys_disk=200GB used_disk=2GB total_vcpus=48 used_vcpus=2 pci_stats=[] {{(pid=61923) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 775.468693] env[61923]: DEBUG oslo_vmware.api [None req-6d07cbaf-b976-4d41-9fed-82d41c8e5c61 tempest-ServerShowV254Test-921228709 tempest-ServerShowV254Test-921228709-project-member] Task: {'id': task-1377507, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.062693} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 775.468962] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-6d07cbaf-b976-4d41-9fed-82d41c8e5c61 tempest-ServerShowV254Test-921228709 tempest-ServerShowV254Test-921228709-project-member] [instance: ecedb77f-4a2a-4047-964f-80796954d847] Extended root virtual disk {{(pid=61923) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 775.469761] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ebd2cbab-bd1f-4712-9f62-fab11efaaa91 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.489885] env[61923]: DEBUG nova.virt.vmwareapi.volumeops [None req-6d07cbaf-b976-4d41-9fed-82d41c8e5c61 tempest-ServerShowV254Test-921228709 tempest-ServerShowV254Test-921228709-project-member] [instance: ecedb77f-4a2a-4047-964f-80796954d847] Reconfiguring VM instance instance-00000034 to attach disk [datastore1] ecedb77f-4a2a-4047-964f-80796954d847/ecedb77f-4a2a-4047-964f-80796954d847.vmdk or device None with type sparse {{(pid=61923) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 775.492481] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-54624d98-b0ae-4498-92e7-5532561d4cf1 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.512395] env[61923]: DEBUG oslo_vmware.api [None req-6d07cbaf-b976-4d41-9fed-82d41c8e5c61 tempest-ServerShowV254Test-921228709 tempest-ServerShowV254Test-921228709-project-member] Waiting for the task: (returnval){ [ 775.512395] env[61923]: value = "task-1377508" [ 775.512395] env[61923]: _type = "Task" [ 775.512395] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 775.523182] env[61923]: DEBUG oslo_vmware.api [None req-6d07cbaf-b976-4d41-9fed-82d41c8e5c61 tempest-ServerShowV254Test-921228709 tempest-ServerShowV254Test-921228709-project-member] Task: {'id': task-1377508, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 775.719848] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d71d2b26-0bc0-49d5-b930-5ec13c8cf253 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.727880] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67a53d57-f89c-4e3e-9ef6-1a98733f218c {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.758006] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-702da676-72d7-442d-bd7b-e0a53412c2ac {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.765838] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29004a10-346c-4cf9-b0d6-43f30b49bacd {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 775.779957] env[61923]: DEBUG nova.compute.provider_tree [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 776.021404] env[61923]: DEBUG oslo_vmware.api [None req-6d07cbaf-b976-4d41-9fed-82d41c8e5c61 tempest-ServerShowV254Test-921228709 tempest-ServerShowV254Test-921228709-project-member] Task: {'id': task-1377508, 'name': ReconfigVM_Task, 'duration_secs': 0.31531} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 776.022024] env[61923]: DEBUG nova.virt.vmwareapi.volumeops [None req-6d07cbaf-b976-4d41-9fed-82d41c8e5c61 tempest-ServerShowV254Test-921228709 tempest-ServerShowV254Test-921228709-project-member] [instance: ecedb77f-4a2a-4047-964f-80796954d847] Reconfigured VM instance instance-00000034 to attach disk [datastore1] ecedb77f-4a2a-4047-964f-80796954d847/ecedb77f-4a2a-4047-964f-80796954d847.vmdk or device None with type sparse {{(pid=61923) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 776.022475] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-1f10d94c-3a1a-45a4-89d9-40bbe3c67c9d {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.028521] env[61923]: DEBUG oslo_vmware.api [None req-6d07cbaf-b976-4d41-9fed-82d41c8e5c61 tempest-ServerShowV254Test-921228709 tempest-ServerShowV254Test-921228709-project-member] Waiting for the task: (returnval){ [ 776.028521] env[61923]: value = "task-1377509" [ 776.028521] env[61923]: _type = "Task" [ 776.028521] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 776.035915] env[61923]: DEBUG oslo_vmware.api [None req-6d07cbaf-b976-4d41-9fed-82d41c8e5c61 tempest-ServerShowV254Test-921228709 tempest-ServerShowV254Test-921228709-project-member] Task: {'id': task-1377509, 'name': Rename_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 776.285626] env[61923]: DEBUG nova.scheduler.client.report [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 776.537912] env[61923]: DEBUG oslo_vmware.api [None req-6d07cbaf-b976-4d41-9fed-82d41c8e5c61 tempest-ServerShowV254Test-921228709 tempest-ServerShowV254Test-921228709-project-member] Task: {'id': task-1377509, 'name': Rename_Task, 'duration_secs': 0.153454} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 776.538209] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-6d07cbaf-b976-4d41-9fed-82d41c8e5c61 tempest-ServerShowV254Test-921228709 tempest-ServerShowV254Test-921228709-project-member] [instance: ecedb77f-4a2a-4047-964f-80796954d847] Powering on the VM {{(pid=61923) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 776.538478] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b5446f6d-4218-4559-b903-01b60e4b5775 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.544510] env[61923]: DEBUG oslo_vmware.api [None req-6d07cbaf-b976-4d41-9fed-82d41c8e5c61 tempest-ServerShowV254Test-921228709 tempest-ServerShowV254Test-921228709-project-member] Waiting for the task: (returnval){ [ 776.544510] env[61923]: value = "task-1377510" [ 776.544510] env[61923]: _type = "Task" [ 776.544510] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 776.552898] env[61923]: DEBUG oslo_vmware.api [None req-6d07cbaf-b976-4d41-9fed-82d41c8e5c61 tempest-ServerShowV254Test-921228709 tempest-ServerShowV254Test-921228709-project-member] Task: {'id': task-1377510, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 776.790786] env[61923]: DEBUG nova.compute.resource_tracker [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61923) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 776.790786] env[61923]: DEBUG oslo_concurrency.lockutils [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 11.947s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 776.791048] env[61923]: DEBUG oslo_concurrency.lockutils [None req-d3d04fc8-2bde-4c80-b2d6-dd8057edd757 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 25.179s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 776.792660] env[61923]: INFO nova.compute.claims [None req-d3d04fc8-2bde-4c80-b2d6-dd8057edd757 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] [instance: b497f3f3-001f-4e5e-b6b7-8e2006116856] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 776.795310] env[61923]: DEBUG oslo_service.periodic_task [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Running periodic task ComputeManager._run_pending_deletes {{(pid=61923) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 776.795485] env[61923]: DEBUG nova.compute.manager [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Cleaning up deleted instances {{(pid=61923) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11228}} [ 777.054986] env[61923]: DEBUG oslo_vmware.api [None req-6d07cbaf-b976-4d41-9fed-82d41c8e5c61 tempest-ServerShowV254Test-921228709 tempest-ServerShowV254Test-921228709-project-member] Task: {'id': task-1377510, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 777.300236] env[61923]: DEBUG nova.compute.manager [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] There are 1 instances to clean {{(pid=61923) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11237}} [ 777.301061] env[61923]: DEBUG nova.compute.manager [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] [instance: f7cc960c-e06a-4c58-9367-ec8771fe09d5] Instance has had 0 of 5 cleanup attempts {{(pid=61923) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 777.555488] env[61923]: DEBUG oslo_vmware.api [None req-6d07cbaf-b976-4d41-9fed-82d41c8e5c61 tempest-ServerShowV254Test-921228709 tempest-ServerShowV254Test-921228709-project-member] Task: {'id': task-1377510, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 777.806340] env[61923]: DEBUG oslo_service.periodic_task [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Running periodic task ComputeManager._cleanup_incomplete_migrations {{(pid=61923) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 777.806499] env[61923]: DEBUG nova.compute.manager [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Cleaning up deleted instances with incomplete migration {{(pid=61923) _cleanup_incomplete_migrations /opt/stack/nova/nova/compute/manager.py:11266}} [ 777.941633] env[61923]: DEBUG nova.scheduler.client.report [None req-d3d04fc8-2bde-4c80-b2d6-dd8057edd757 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] Refreshing inventories for resource provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 777.956909] env[61923]: DEBUG nova.scheduler.client.report [None req-d3d04fc8-2bde-4c80-b2d6-dd8057edd757 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] Updating ProviderTree inventory for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 777.957159] env[61923]: DEBUG nova.compute.provider_tree [None req-d3d04fc8-2bde-4c80-b2d6-dd8057edd757 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] Updating inventory in ProviderTree for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 777.969029] env[61923]: DEBUG nova.scheduler.client.report [None req-d3d04fc8-2bde-4c80-b2d6-dd8057edd757 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] Refreshing aggregate associations for resource provider f81803f2-f7f8-4939-a757-a77d34a1d0a2, aggregates: None {{(pid=61923) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 778.001947] env[61923]: DEBUG nova.scheduler.client.report [None req-d3d04fc8-2bde-4c80-b2d6-dd8057edd757 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] Refreshing trait associations for resource provider f81803f2-f7f8-4939-a757-a77d34a1d0a2, traits: COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NODE,HW_ARCH_X86_64,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_IMAGE_TYPE_ISO {{(pid=61923) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 778.059383] env[61923]: DEBUG oslo_vmware.api [None req-6d07cbaf-b976-4d41-9fed-82d41c8e5c61 tempest-ServerShowV254Test-921228709 tempest-ServerShowV254Test-921228709-project-member] Task: {'id': task-1377510, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 778.221412] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e4adda4-7608-4c8e-88d6-3c3ce36b23b0 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.228789] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd39e9df-7b10-4795-83f1-facd73a4ad15 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.259257] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04d6b857-5d5f-40f3-a0e2-e022c3352eff {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.266838] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e16a04c-cdac-4e59-95eb-19b939c5cbc1 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.282032] env[61923]: DEBUG nova.compute.provider_tree [None req-d3d04fc8-2bde-4c80-b2d6-dd8057edd757 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 778.309633] env[61923]: DEBUG oslo_service.periodic_task [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Running periodic task ComputeManager._cleanup_expired_console_auth_tokens {{(pid=61923) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 778.557405] env[61923]: DEBUG oslo_vmware.api [None req-6d07cbaf-b976-4d41-9fed-82d41c8e5c61 tempest-ServerShowV254Test-921228709 tempest-ServerShowV254Test-921228709-project-member] Task: {'id': task-1377510, 'name': PowerOnVM_Task, 'duration_secs': 1.852046} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 778.557689] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-6d07cbaf-b976-4d41-9fed-82d41c8e5c61 tempest-ServerShowV254Test-921228709 tempest-ServerShowV254Test-921228709-project-member] [instance: ecedb77f-4a2a-4047-964f-80796954d847] Powered on the VM {{(pid=61923) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 778.557873] env[61923]: DEBUG nova.compute.manager [None req-6d07cbaf-b976-4d41-9fed-82d41c8e5c61 tempest-ServerShowV254Test-921228709 tempest-ServerShowV254Test-921228709-project-member] [instance: ecedb77f-4a2a-4047-964f-80796954d847] Checking state {{(pid=61923) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 778.558682] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60052fa3-fee4-4938-bd2d-a2162ca7d506 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.784228] env[61923]: DEBUG nova.scheduler.client.report [None req-d3d04fc8-2bde-4c80-b2d6-dd8057edd757 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 779.076596] env[61923]: DEBUG oslo_concurrency.lockutils [None req-6d07cbaf-b976-4d41-9fed-82d41c8e5c61 tempest-ServerShowV254Test-921228709 tempest-ServerShowV254Test-921228709-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 779.289476] env[61923]: DEBUG oslo_concurrency.lockutils [None req-d3d04fc8-2bde-4c80-b2d6-dd8057edd757 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.498s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 779.290051] env[61923]: DEBUG nova.compute.manager [None req-d3d04fc8-2bde-4c80-b2d6-dd8057edd757 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] [instance: b497f3f3-001f-4e5e-b6b7-8e2006116856] Start building networks asynchronously for instance. {{(pid=61923) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 779.292762] env[61923]: DEBUG oslo_concurrency.lockutils [None req-1d78b71c-3449-4286-a17a-dc5396f589fe tempest-ServersAdminTestJSON-678434436 tempest-ServersAdminTestJSON-678434436-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 23.360s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 779.294173] env[61923]: INFO nova.compute.claims [None req-1d78b71c-3449-4286-a17a-dc5396f589fe tempest-ServersAdminTestJSON-678434436 tempest-ServersAdminTestJSON-678434436-project-member] [instance: faff83a1-3532-4396-ac3f-600f86653749] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 779.805229] env[61923]: DEBUG nova.compute.utils [None req-d3d04fc8-2bde-4c80-b2d6-dd8057edd757 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] Using /dev/sd instead of None {{(pid=61923) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 779.806693] env[61923]: DEBUG nova.compute.manager [None req-d3d04fc8-2bde-4c80-b2d6-dd8057edd757 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] [instance: b497f3f3-001f-4e5e-b6b7-8e2006116856] Allocating IP information in the background. {{(pid=61923) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 779.806880] env[61923]: DEBUG nova.network.neutron [None req-d3d04fc8-2bde-4c80-b2d6-dd8057edd757 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] [instance: b497f3f3-001f-4e5e-b6b7-8e2006116856] allocate_for_instance() {{(pid=61923) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 779.855510] env[61923]: DEBUG nova.policy [None req-d3d04fc8-2bde-4c80-b2d6-dd8057edd757 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '629d5a54afa44599a4c0756d997cabc1', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '87f548d252244b759e3f157af9ab875d', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61923) authorize /opt/stack/nova/nova/policy.py:201}} [ 780.193189] env[61923]: DEBUG nova.network.neutron [None req-d3d04fc8-2bde-4c80-b2d6-dd8057edd757 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] [instance: b497f3f3-001f-4e5e-b6b7-8e2006116856] Successfully created port: fc48dbbd-d086-4537-ab13-11a91a9ac0fc {{(pid=61923) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 780.213569] env[61923]: DEBUG oslo_concurrency.lockutils [None req-33807679-7c2d-4470-9448-1b4dbb76ab45 tempest-ServerShowV254Test-921228709 tempest-ServerShowV254Test-921228709-project-member] Acquiring lock "ecedb77f-4a2a-4047-964f-80796954d847" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 780.213871] env[61923]: DEBUG oslo_concurrency.lockutils [None req-33807679-7c2d-4470-9448-1b4dbb76ab45 tempest-ServerShowV254Test-921228709 tempest-ServerShowV254Test-921228709-project-member] Lock "ecedb77f-4a2a-4047-964f-80796954d847" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 780.214134] env[61923]: DEBUG oslo_concurrency.lockutils [None req-33807679-7c2d-4470-9448-1b4dbb76ab45 tempest-ServerShowV254Test-921228709 tempest-ServerShowV254Test-921228709-project-member] Acquiring lock "ecedb77f-4a2a-4047-964f-80796954d847-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 780.214361] env[61923]: DEBUG oslo_concurrency.lockutils [None req-33807679-7c2d-4470-9448-1b4dbb76ab45 tempest-ServerShowV254Test-921228709 tempest-ServerShowV254Test-921228709-project-member] Lock "ecedb77f-4a2a-4047-964f-80796954d847-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 780.214563] env[61923]: DEBUG oslo_concurrency.lockutils [None req-33807679-7c2d-4470-9448-1b4dbb76ab45 tempest-ServerShowV254Test-921228709 tempest-ServerShowV254Test-921228709-project-member] Lock "ecedb77f-4a2a-4047-964f-80796954d847-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 780.216870] env[61923]: INFO nova.compute.manager [None req-33807679-7c2d-4470-9448-1b4dbb76ab45 tempest-ServerShowV254Test-921228709 tempest-ServerShowV254Test-921228709-project-member] [instance: ecedb77f-4a2a-4047-964f-80796954d847] Terminating instance [ 780.218648] env[61923]: DEBUG oslo_concurrency.lockutils [None req-33807679-7c2d-4470-9448-1b4dbb76ab45 tempest-ServerShowV254Test-921228709 tempest-ServerShowV254Test-921228709-project-member] Acquiring lock "refresh_cache-ecedb77f-4a2a-4047-964f-80796954d847" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 780.218813] env[61923]: DEBUG oslo_concurrency.lockutils [None req-33807679-7c2d-4470-9448-1b4dbb76ab45 tempest-ServerShowV254Test-921228709 tempest-ServerShowV254Test-921228709-project-member] Acquired lock "refresh_cache-ecedb77f-4a2a-4047-964f-80796954d847" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 780.219015] env[61923]: DEBUG nova.network.neutron [None req-33807679-7c2d-4470-9448-1b4dbb76ab45 tempest-ServerShowV254Test-921228709 tempest-ServerShowV254Test-921228709-project-member] [instance: ecedb77f-4a2a-4047-964f-80796954d847] Building network info cache for instance {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 780.311086] env[61923]: DEBUG nova.compute.manager [None req-d3d04fc8-2bde-4c80-b2d6-dd8057edd757 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] [instance: b497f3f3-001f-4e5e-b6b7-8e2006116856] Start building block device mappings for instance. {{(pid=61923) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 780.600623] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87dd92c5-e5b9-460e-8906-3e09ebb73709 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.610070] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09d5e701-e2e0-4246-86d2-2e65dd5a933a {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.641180] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6bfc6be5-4dfb-45dc-a1d6-909df2d2ae52 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.649020] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51b7e9c7-cfc7-4f06-82d4-7b1d53babf89 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.662508] env[61923]: DEBUG nova.compute.provider_tree [None req-1d78b71c-3449-4286-a17a-dc5396f589fe tempest-ServersAdminTestJSON-678434436 tempest-ServersAdminTestJSON-678434436-project-member] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 780.743230] env[61923]: DEBUG nova.network.neutron [None req-33807679-7c2d-4470-9448-1b4dbb76ab45 tempest-ServerShowV254Test-921228709 tempest-ServerShowV254Test-921228709-project-member] [instance: ecedb77f-4a2a-4047-964f-80796954d847] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 780.824044] env[61923]: DEBUG nova.network.neutron [None req-33807679-7c2d-4470-9448-1b4dbb76ab45 tempest-ServerShowV254Test-921228709 tempest-ServerShowV254Test-921228709-project-member] [instance: ecedb77f-4a2a-4047-964f-80796954d847] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 781.165668] env[61923]: DEBUG nova.scheduler.client.report [None req-1d78b71c-3449-4286-a17a-dc5396f589fe tempest-ServersAdminTestJSON-678434436 tempest-ServersAdminTestJSON-678434436-project-member] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 781.255125] env[61923]: DEBUG nova.compute.manager [req-2896729d-ee43-4b3b-b56e-2c0bc12631fe req-cd6a7933-84a9-4792-8d70-600b9884ef83 service nova] [instance: b497f3f3-001f-4e5e-b6b7-8e2006116856] Received event network-changed-fc48dbbd-d086-4537-ab13-11a91a9ac0fc {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 781.255125] env[61923]: DEBUG nova.compute.manager [req-2896729d-ee43-4b3b-b56e-2c0bc12631fe req-cd6a7933-84a9-4792-8d70-600b9884ef83 service nova] [instance: b497f3f3-001f-4e5e-b6b7-8e2006116856] Refreshing instance network info cache due to event network-changed-fc48dbbd-d086-4537-ab13-11a91a9ac0fc. {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 781.255125] env[61923]: DEBUG oslo_concurrency.lockutils [req-2896729d-ee43-4b3b-b56e-2c0bc12631fe req-cd6a7933-84a9-4792-8d70-600b9884ef83 service nova] Acquiring lock "refresh_cache-b497f3f3-001f-4e5e-b6b7-8e2006116856" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 781.255125] env[61923]: DEBUG oslo_concurrency.lockutils [req-2896729d-ee43-4b3b-b56e-2c0bc12631fe req-cd6a7933-84a9-4792-8d70-600b9884ef83 service nova] Acquired lock "refresh_cache-b497f3f3-001f-4e5e-b6b7-8e2006116856" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 781.255125] env[61923]: DEBUG nova.network.neutron [req-2896729d-ee43-4b3b-b56e-2c0bc12631fe req-cd6a7933-84a9-4792-8d70-600b9884ef83 service nova] [instance: b497f3f3-001f-4e5e-b6b7-8e2006116856] Refreshing network info cache for port fc48dbbd-d086-4537-ab13-11a91a9ac0fc {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 781.330098] env[61923]: DEBUG nova.compute.manager [None req-d3d04fc8-2bde-4c80-b2d6-dd8057edd757 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] [instance: b497f3f3-001f-4e5e-b6b7-8e2006116856] Start spawning the instance on the hypervisor. {{(pid=61923) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 781.330098] env[61923]: DEBUG oslo_concurrency.lockutils [None req-33807679-7c2d-4470-9448-1b4dbb76ab45 tempest-ServerShowV254Test-921228709 tempest-ServerShowV254Test-921228709-project-member] Releasing lock "refresh_cache-ecedb77f-4a2a-4047-964f-80796954d847" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 781.330518] env[61923]: DEBUG nova.compute.manager [None req-33807679-7c2d-4470-9448-1b4dbb76ab45 tempest-ServerShowV254Test-921228709 tempest-ServerShowV254Test-921228709-project-member] [instance: ecedb77f-4a2a-4047-964f-80796954d847] Start destroying the instance on the hypervisor. {{(pid=61923) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 781.333403] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-33807679-7c2d-4470-9448-1b4dbb76ab45 tempest-ServerShowV254Test-921228709 tempest-ServerShowV254Test-921228709-project-member] [instance: ecedb77f-4a2a-4047-964f-80796954d847] Destroying instance {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 781.333403] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1d666e8-3d91-4d22-afdb-9c65fd4dbf4c {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.341552] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-33807679-7c2d-4470-9448-1b4dbb76ab45 tempest-ServerShowV254Test-921228709 tempest-ServerShowV254Test-921228709-project-member] [instance: ecedb77f-4a2a-4047-964f-80796954d847] Powering off the VM {{(pid=61923) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 781.341871] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-93f71c9a-edcc-4db2-bb4a-342bc49bb751 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.348023] env[61923]: DEBUG oslo_vmware.api [None req-33807679-7c2d-4470-9448-1b4dbb76ab45 tempest-ServerShowV254Test-921228709 tempest-ServerShowV254Test-921228709-project-member] Waiting for the task: (returnval){ [ 781.348023] env[61923]: value = "task-1377511" [ 781.348023] env[61923]: _type = "Task" [ 781.348023] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 781.356835] env[61923]: DEBUG nova.virt.hardware [None req-d3d04fc8-2bde-4c80-b2d6-dd8057edd757 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-29T20:07:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-29T20:07:35Z,direct_url=,disk_format='vmdk',id=0f153f63-ae0a-45b1-b7f5-7f9b673c947f,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='4e7b5e3b3c3443c8bd5c70f8a15739f5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-29T20:07:36Z,virtual_size=,visibility=), allow threads: False {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 781.357123] env[61923]: DEBUG nova.virt.hardware [None req-d3d04fc8-2bde-4c80-b2d6-dd8057edd757 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] Flavor limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 781.357255] env[61923]: DEBUG nova.virt.hardware [None req-d3d04fc8-2bde-4c80-b2d6-dd8057edd757 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] Image limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 781.357429] env[61923]: DEBUG nova.virt.hardware [None req-d3d04fc8-2bde-4c80-b2d6-dd8057edd757 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] Flavor pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 781.357571] env[61923]: DEBUG nova.virt.hardware [None req-d3d04fc8-2bde-4c80-b2d6-dd8057edd757 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] Image pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 781.357714] env[61923]: DEBUG nova.virt.hardware [None req-d3d04fc8-2bde-4c80-b2d6-dd8057edd757 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 781.357911] env[61923]: DEBUG nova.virt.hardware [None req-d3d04fc8-2bde-4c80-b2d6-dd8057edd757 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 781.358145] env[61923]: DEBUG nova.virt.hardware [None req-d3d04fc8-2bde-4c80-b2d6-dd8057edd757 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 781.358876] env[61923]: DEBUG nova.virt.hardware [None req-d3d04fc8-2bde-4c80-b2d6-dd8057edd757 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] Got 1 possible topologies {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 781.358876] env[61923]: DEBUG nova.virt.hardware [None req-d3d04fc8-2bde-4c80-b2d6-dd8057edd757 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 781.358876] env[61923]: DEBUG nova.virt.hardware [None req-d3d04fc8-2bde-4c80-b2d6-dd8057edd757 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 781.359625] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f200d697-8230-4fd5-ba9a-ec784370becd {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.365476] env[61923]: DEBUG oslo_vmware.api [None req-33807679-7c2d-4470-9448-1b4dbb76ab45 tempest-ServerShowV254Test-921228709 tempest-ServerShowV254Test-921228709-project-member] Task: {'id': task-1377511, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 781.370766] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4639aee5-156b-4321-b66c-6e32c20305ed {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.412990] env[61923]: ERROR nova.compute.manager [None req-d3d04fc8-2bde-4c80-b2d6-dd8057edd757 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port fc48dbbd-d086-4537-ab13-11a91a9ac0fc, please check neutron logs for more information. [ 781.412990] env[61923]: ERROR nova.compute.manager Traceback (most recent call last): [ 781.412990] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 781.412990] env[61923]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 781.412990] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 781.412990] env[61923]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 781.412990] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 781.412990] env[61923]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 781.412990] env[61923]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 781.412990] env[61923]: ERROR nova.compute.manager self.force_reraise() [ 781.412990] env[61923]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 781.412990] env[61923]: ERROR nova.compute.manager raise self.value [ 781.412990] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 781.412990] env[61923]: ERROR nova.compute.manager updated_port = self._update_port( [ 781.412990] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 781.412990] env[61923]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 781.413412] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 781.413412] env[61923]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 781.413412] env[61923]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port fc48dbbd-d086-4537-ab13-11a91a9ac0fc, please check neutron logs for more information. [ 781.413412] env[61923]: ERROR nova.compute.manager [ 781.413412] env[61923]: Traceback (most recent call last): [ 781.413412] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 781.413412] env[61923]: listener.cb(fileno) [ 781.413412] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 781.413412] env[61923]: result = function(*args, **kwargs) [ 781.413412] env[61923]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 781.413412] env[61923]: return func(*args, **kwargs) [ 781.413412] env[61923]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 781.413412] env[61923]: raise e [ 781.413412] env[61923]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 781.413412] env[61923]: nwinfo = self.network_api.allocate_for_instance( [ 781.413412] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 781.413412] env[61923]: created_port_ids = self._update_ports_for_instance( [ 781.413412] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 781.413412] env[61923]: with excutils.save_and_reraise_exception(): [ 781.413412] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 781.413412] env[61923]: self.force_reraise() [ 781.413412] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 781.413412] env[61923]: raise self.value [ 781.413412] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 781.413412] env[61923]: updated_port = self._update_port( [ 781.413412] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 781.413412] env[61923]: _ensure_no_port_binding_failure(port) [ 781.413412] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 781.413412] env[61923]: raise exception.PortBindingFailed(port_id=port['id']) [ 781.414087] env[61923]: nova.exception.PortBindingFailed: Binding failed for port fc48dbbd-d086-4537-ab13-11a91a9ac0fc, please check neutron logs for more information. [ 781.414087] env[61923]: Removing descriptor: 14 [ 781.414087] env[61923]: ERROR nova.compute.manager [None req-d3d04fc8-2bde-4c80-b2d6-dd8057edd757 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] [instance: b497f3f3-001f-4e5e-b6b7-8e2006116856] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port fc48dbbd-d086-4537-ab13-11a91a9ac0fc, please check neutron logs for more information. [ 781.414087] env[61923]: ERROR nova.compute.manager [instance: b497f3f3-001f-4e5e-b6b7-8e2006116856] Traceback (most recent call last): [ 781.414087] env[61923]: ERROR nova.compute.manager [instance: b497f3f3-001f-4e5e-b6b7-8e2006116856] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 781.414087] env[61923]: ERROR nova.compute.manager [instance: b497f3f3-001f-4e5e-b6b7-8e2006116856] yield resources [ 781.414087] env[61923]: ERROR nova.compute.manager [instance: b497f3f3-001f-4e5e-b6b7-8e2006116856] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 781.414087] env[61923]: ERROR nova.compute.manager [instance: b497f3f3-001f-4e5e-b6b7-8e2006116856] self.driver.spawn(context, instance, image_meta, [ 781.414087] env[61923]: ERROR nova.compute.manager [instance: b497f3f3-001f-4e5e-b6b7-8e2006116856] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 781.414087] env[61923]: ERROR nova.compute.manager [instance: b497f3f3-001f-4e5e-b6b7-8e2006116856] self._vmops.spawn(context, instance, image_meta, injected_files, [ 781.414087] env[61923]: ERROR nova.compute.manager [instance: b497f3f3-001f-4e5e-b6b7-8e2006116856] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 781.414087] env[61923]: ERROR nova.compute.manager [instance: b497f3f3-001f-4e5e-b6b7-8e2006116856] vm_ref = self.build_virtual_machine(instance, [ 781.414361] env[61923]: ERROR nova.compute.manager [instance: b497f3f3-001f-4e5e-b6b7-8e2006116856] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 781.414361] env[61923]: ERROR nova.compute.manager [instance: b497f3f3-001f-4e5e-b6b7-8e2006116856] vif_infos = vmwarevif.get_vif_info(self._session, [ 781.414361] env[61923]: ERROR nova.compute.manager [instance: b497f3f3-001f-4e5e-b6b7-8e2006116856] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 781.414361] env[61923]: ERROR nova.compute.manager [instance: b497f3f3-001f-4e5e-b6b7-8e2006116856] for vif in network_info: [ 781.414361] env[61923]: ERROR nova.compute.manager [instance: b497f3f3-001f-4e5e-b6b7-8e2006116856] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 781.414361] env[61923]: ERROR nova.compute.manager [instance: b497f3f3-001f-4e5e-b6b7-8e2006116856] return self._sync_wrapper(fn, *args, **kwargs) [ 781.414361] env[61923]: ERROR nova.compute.manager [instance: b497f3f3-001f-4e5e-b6b7-8e2006116856] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 781.414361] env[61923]: ERROR nova.compute.manager [instance: b497f3f3-001f-4e5e-b6b7-8e2006116856] self.wait() [ 781.414361] env[61923]: ERROR nova.compute.manager [instance: b497f3f3-001f-4e5e-b6b7-8e2006116856] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 781.414361] env[61923]: ERROR nova.compute.manager [instance: b497f3f3-001f-4e5e-b6b7-8e2006116856] self[:] = self._gt.wait() [ 781.414361] env[61923]: ERROR nova.compute.manager [instance: b497f3f3-001f-4e5e-b6b7-8e2006116856] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 781.414361] env[61923]: ERROR nova.compute.manager [instance: b497f3f3-001f-4e5e-b6b7-8e2006116856] return self._exit_event.wait() [ 781.414361] env[61923]: ERROR nova.compute.manager [instance: b497f3f3-001f-4e5e-b6b7-8e2006116856] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 781.414659] env[61923]: ERROR nova.compute.manager [instance: b497f3f3-001f-4e5e-b6b7-8e2006116856] result = hub.switch() [ 781.414659] env[61923]: ERROR nova.compute.manager [instance: b497f3f3-001f-4e5e-b6b7-8e2006116856] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 781.414659] env[61923]: ERROR nova.compute.manager [instance: b497f3f3-001f-4e5e-b6b7-8e2006116856] return self.greenlet.switch() [ 781.414659] env[61923]: ERROR nova.compute.manager [instance: b497f3f3-001f-4e5e-b6b7-8e2006116856] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 781.414659] env[61923]: ERROR nova.compute.manager [instance: b497f3f3-001f-4e5e-b6b7-8e2006116856] result = function(*args, **kwargs) [ 781.414659] env[61923]: ERROR nova.compute.manager [instance: b497f3f3-001f-4e5e-b6b7-8e2006116856] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 781.414659] env[61923]: ERROR nova.compute.manager [instance: b497f3f3-001f-4e5e-b6b7-8e2006116856] return func(*args, **kwargs) [ 781.414659] env[61923]: ERROR nova.compute.manager [instance: b497f3f3-001f-4e5e-b6b7-8e2006116856] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 781.414659] env[61923]: ERROR nova.compute.manager [instance: b497f3f3-001f-4e5e-b6b7-8e2006116856] raise e [ 781.414659] env[61923]: ERROR nova.compute.manager [instance: b497f3f3-001f-4e5e-b6b7-8e2006116856] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 781.414659] env[61923]: ERROR nova.compute.manager [instance: b497f3f3-001f-4e5e-b6b7-8e2006116856] nwinfo = self.network_api.allocate_for_instance( [ 781.414659] env[61923]: ERROR nova.compute.manager [instance: b497f3f3-001f-4e5e-b6b7-8e2006116856] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 781.414659] env[61923]: ERROR nova.compute.manager [instance: b497f3f3-001f-4e5e-b6b7-8e2006116856] created_port_ids = self._update_ports_for_instance( [ 781.414960] env[61923]: ERROR nova.compute.manager [instance: b497f3f3-001f-4e5e-b6b7-8e2006116856] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 781.414960] env[61923]: ERROR nova.compute.manager [instance: b497f3f3-001f-4e5e-b6b7-8e2006116856] with excutils.save_and_reraise_exception(): [ 781.414960] env[61923]: ERROR nova.compute.manager [instance: b497f3f3-001f-4e5e-b6b7-8e2006116856] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 781.414960] env[61923]: ERROR nova.compute.manager [instance: b497f3f3-001f-4e5e-b6b7-8e2006116856] self.force_reraise() [ 781.414960] env[61923]: ERROR nova.compute.manager [instance: b497f3f3-001f-4e5e-b6b7-8e2006116856] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 781.414960] env[61923]: ERROR nova.compute.manager [instance: b497f3f3-001f-4e5e-b6b7-8e2006116856] raise self.value [ 781.414960] env[61923]: ERROR nova.compute.manager [instance: b497f3f3-001f-4e5e-b6b7-8e2006116856] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 781.414960] env[61923]: ERROR nova.compute.manager [instance: b497f3f3-001f-4e5e-b6b7-8e2006116856] updated_port = self._update_port( [ 781.414960] env[61923]: ERROR nova.compute.manager [instance: b497f3f3-001f-4e5e-b6b7-8e2006116856] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 781.414960] env[61923]: ERROR nova.compute.manager [instance: b497f3f3-001f-4e5e-b6b7-8e2006116856] _ensure_no_port_binding_failure(port) [ 781.414960] env[61923]: ERROR nova.compute.manager [instance: b497f3f3-001f-4e5e-b6b7-8e2006116856] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 781.414960] env[61923]: ERROR nova.compute.manager [instance: b497f3f3-001f-4e5e-b6b7-8e2006116856] raise exception.PortBindingFailed(port_id=port['id']) [ 781.415275] env[61923]: ERROR nova.compute.manager [instance: b497f3f3-001f-4e5e-b6b7-8e2006116856] nova.exception.PortBindingFailed: Binding failed for port fc48dbbd-d086-4537-ab13-11a91a9ac0fc, please check neutron logs for more information. [ 781.415275] env[61923]: ERROR nova.compute.manager [instance: b497f3f3-001f-4e5e-b6b7-8e2006116856] [ 781.415275] env[61923]: INFO nova.compute.manager [None req-d3d04fc8-2bde-4c80-b2d6-dd8057edd757 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] [instance: b497f3f3-001f-4e5e-b6b7-8e2006116856] Terminating instance [ 781.417970] env[61923]: DEBUG oslo_concurrency.lockutils [None req-d3d04fc8-2bde-4c80-b2d6-dd8057edd757 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] Acquiring lock "refresh_cache-b497f3f3-001f-4e5e-b6b7-8e2006116856" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 781.671393] env[61923]: DEBUG oslo_concurrency.lockutils [None req-1d78b71c-3449-4286-a17a-dc5396f589fe tempest-ServersAdminTestJSON-678434436 tempest-ServersAdminTestJSON-678434436-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.378s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 781.671925] env[61923]: DEBUG nova.compute.manager [None req-1d78b71c-3449-4286-a17a-dc5396f589fe tempest-ServersAdminTestJSON-678434436 tempest-ServersAdminTestJSON-678434436-project-member] [instance: faff83a1-3532-4396-ac3f-600f86653749] Start building networks asynchronously for instance. {{(pid=61923) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 781.674824] env[61923]: DEBUG oslo_concurrency.lockutils [None req-6e6344ac-0cd2-410e-863e-3289b0c1fe3f tempest-ServersAdminTestJSON-678434436 tempest-ServersAdminTestJSON-678434436-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 23.949s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 781.676268] env[61923]: INFO nova.compute.claims [None req-6e6344ac-0cd2-410e-863e-3289b0c1fe3f tempest-ServersAdminTestJSON-678434436 tempest-ServersAdminTestJSON-678434436-project-member] [instance: 3f298e29-69ff-46ac-8018-ceb3990b9848] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 781.773705] env[61923]: DEBUG nova.network.neutron [req-2896729d-ee43-4b3b-b56e-2c0bc12631fe req-cd6a7933-84a9-4792-8d70-600b9884ef83 service nova] [instance: b497f3f3-001f-4e5e-b6b7-8e2006116856] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 781.859598] env[61923]: DEBUG oslo_vmware.api [None req-33807679-7c2d-4470-9448-1b4dbb76ab45 tempest-ServerShowV254Test-921228709 tempest-ServerShowV254Test-921228709-project-member] Task: {'id': task-1377511, 'name': PowerOffVM_Task, 'duration_secs': 0.133173} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 781.859919] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-33807679-7c2d-4470-9448-1b4dbb76ab45 tempest-ServerShowV254Test-921228709 tempest-ServerShowV254Test-921228709-project-member] [instance: ecedb77f-4a2a-4047-964f-80796954d847] Powered off the VM {{(pid=61923) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 781.860018] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-33807679-7c2d-4470-9448-1b4dbb76ab45 tempest-ServerShowV254Test-921228709 tempest-ServerShowV254Test-921228709-project-member] [instance: ecedb77f-4a2a-4047-964f-80796954d847] Unregistering the VM {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 781.860271] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f2a8678b-9fbe-4f6f-b5a6-ec2084d65014 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.863056] env[61923]: DEBUG nova.network.neutron [req-2896729d-ee43-4b3b-b56e-2c0bc12631fe req-cd6a7933-84a9-4792-8d70-600b9884ef83 service nova] [instance: b497f3f3-001f-4e5e-b6b7-8e2006116856] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 781.886445] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-33807679-7c2d-4470-9448-1b4dbb76ab45 tempest-ServerShowV254Test-921228709 tempest-ServerShowV254Test-921228709-project-member] [instance: ecedb77f-4a2a-4047-964f-80796954d847] Unregistered the VM {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 781.886670] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-33807679-7c2d-4470-9448-1b4dbb76ab45 tempest-ServerShowV254Test-921228709 tempest-ServerShowV254Test-921228709-project-member] [instance: ecedb77f-4a2a-4047-964f-80796954d847] Deleting contents of the VM from datastore datastore1 {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 781.886878] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-33807679-7c2d-4470-9448-1b4dbb76ab45 tempest-ServerShowV254Test-921228709 tempest-ServerShowV254Test-921228709-project-member] Deleting the datastore file [datastore1] ecedb77f-4a2a-4047-964f-80796954d847 {{(pid=61923) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 781.887155] env[61923]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-aaa06bc6-9430-4d67-8b56-05900f0abbe8 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 781.893874] env[61923]: DEBUG oslo_vmware.api [None req-33807679-7c2d-4470-9448-1b4dbb76ab45 tempest-ServerShowV254Test-921228709 tempest-ServerShowV254Test-921228709-project-member] Waiting for the task: (returnval){ [ 781.893874] env[61923]: value = "task-1377513" [ 781.893874] env[61923]: _type = "Task" [ 781.893874] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 781.901770] env[61923]: DEBUG oslo_vmware.api [None req-33807679-7c2d-4470-9448-1b4dbb76ab45 tempest-ServerShowV254Test-921228709 tempest-ServerShowV254Test-921228709-project-member] Task: {'id': task-1377513, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 782.181711] env[61923]: DEBUG nova.compute.utils [None req-1d78b71c-3449-4286-a17a-dc5396f589fe tempest-ServersAdminTestJSON-678434436 tempest-ServersAdminTestJSON-678434436-project-member] Using /dev/sd instead of None {{(pid=61923) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 782.184974] env[61923]: DEBUG nova.compute.manager [None req-1d78b71c-3449-4286-a17a-dc5396f589fe tempest-ServersAdminTestJSON-678434436 tempest-ServersAdminTestJSON-678434436-project-member] [instance: faff83a1-3532-4396-ac3f-600f86653749] Allocating IP information in the background. {{(pid=61923) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 782.185163] env[61923]: DEBUG nova.network.neutron [None req-1d78b71c-3449-4286-a17a-dc5396f589fe tempest-ServersAdminTestJSON-678434436 tempest-ServersAdminTestJSON-678434436-project-member] [instance: faff83a1-3532-4396-ac3f-600f86653749] allocate_for_instance() {{(pid=61923) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 782.248780] env[61923]: DEBUG nova.policy [None req-1d78b71c-3449-4286-a17a-dc5396f589fe tempest-ServersAdminTestJSON-678434436 tempest-ServersAdminTestJSON-678434436-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '492cf67e8d934790b177501a16161a55', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '9d757357e4754e67853dca46d17d9eb7', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61923) authorize /opt/stack/nova/nova/policy.py:201}} [ 782.365676] env[61923]: DEBUG oslo_concurrency.lockutils [req-2896729d-ee43-4b3b-b56e-2c0bc12631fe req-cd6a7933-84a9-4792-8d70-600b9884ef83 service nova] Releasing lock "refresh_cache-b497f3f3-001f-4e5e-b6b7-8e2006116856" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 782.366427] env[61923]: DEBUG oslo_concurrency.lockutils [None req-d3d04fc8-2bde-4c80-b2d6-dd8057edd757 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] Acquired lock "refresh_cache-b497f3f3-001f-4e5e-b6b7-8e2006116856" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 782.366427] env[61923]: DEBUG nova.network.neutron [None req-d3d04fc8-2bde-4c80-b2d6-dd8057edd757 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] [instance: b497f3f3-001f-4e5e-b6b7-8e2006116856] Building network info cache for instance {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 782.403609] env[61923]: DEBUG oslo_vmware.api [None req-33807679-7c2d-4470-9448-1b4dbb76ab45 tempest-ServerShowV254Test-921228709 tempest-ServerShowV254Test-921228709-project-member] Task: {'id': task-1377513, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.090634} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 782.403847] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-33807679-7c2d-4470-9448-1b4dbb76ab45 tempest-ServerShowV254Test-921228709 tempest-ServerShowV254Test-921228709-project-member] Deleted the datastore file {{(pid=61923) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 782.404034] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-33807679-7c2d-4470-9448-1b4dbb76ab45 tempest-ServerShowV254Test-921228709 tempest-ServerShowV254Test-921228709-project-member] [instance: ecedb77f-4a2a-4047-964f-80796954d847] Deleted contents of the VM from datastore datastore1 {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 782.404211] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-33807679-7c2d-4470-9448-1b4dbb76ab45 tempest-ServerShowV254Test-921228709 tempest-ServerShowV254Test-921228709-project-member] [instance: ecedb77f-4a2a-4047-964f-80796954d847] Instance destroyed {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 782.404377] env[61923]: INFO nova.compute.manager [None req-33807679-7c2d-4470-9448-1b4dbb76ab45 tempest-ServerShowV254Test-921228709 tempest-ServerShowV254Test-921228709-project-member] [instance: ecedb77f-4a2a-4047-964f-80796954d847] Took 1.07 seconds to destroy the instance on the hypervisor. [ 782.404607] env[61923]: DEBUG oslo.service.loopingcall [None req-33807679-7c2d-4470-9448-1b4dbb76ab45 tempest-ServerShowV254Test-921228709 tempest-ServerShowV254Test-921228709-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61923) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 782.404787] env[61923]: DEBUG nova.compute.manager [-] [instance: ecedb77f-4a2a-4047-964f-80796954d847] Deallocating network for instance {{(pid=61923) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 782.404878] env[61923]: DEBUG nova.network.neutron [-] [instance: ecedb77f-4a2a-4047-964f-80796954d847] deallocate_for_instance() {{(pid=61923) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 782.424672] env[61923]: DEBUG nova.network.neutron [-] [instance: ecedb77f-4a2a-4047-964f-80796954d847] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 782.508067] env[61923]: DEBUG nova.network.neutron [None req-1d78b71c-3449-4286-a17a-dc5396f589fe tempest-ServersAdminTestJSON-678434436 tempest-ServersAdminTestJSON-678434436-project-member] [instance: faff83a1-3532-4396-ac3f-600f86653749] Successfully created port: 556dc345-8ad3-4e4f-8dd5-1d38e265adae {{(pid=61923) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 782.686086] env[61923]: DEBUG nova.compute.manager [None req-1d78b71c-3449-4286-a17a-dc5396f589fe tempest-ServersAdminTestJSON-678434436 tempest-ServersAdminTestJSON-678434436-project-member] [instance: faff83a1-3532-4396-ac3f-600f86653749] Start building block device mappings for instance. {{(pid=61923) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 782.915929] env[61923]: DEBUG nova.network.neutron [None req-d3d04fc8-2bde-4c80-b2d6-dd8057edd757 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] [instance: b497f3f3-001f-4e5e-b6b7-8e2006116856] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 782.928455] env[61923]: DEBUG nova.network.neutron [-] [instance: ecedb77f-4a2a-4047-964f-80796954d847] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 783.015022] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b81174de-d4a5-447a-8236-79e4357dfe5b {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.023869] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23c7a87a-1b55-4b7f-be74-329b370f9152 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.055668] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-facf4894-6a32-4db8-8ec6-b031d4fe1765 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.063646] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75c68f7a-68db-443c-9ac0-57422de94b2e {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.078766] env[61923]: DEBUG nova.compute.provider_tree [None req-6e6344ac-0cd2-410e-863e-3289b0c1fe3f tempest-ServersAdminTestJSON-678434436 tempest-ServersAdminTestJSON-678434436-project-member] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 783.113554] env[61923]: DEBUG nova.network.neutron [None req-d3d04fc8-2bde-4c80-b2d6-dd8057edd757 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] [instance: b497f3f3-001f-4e5e-b6b7-8e2006116856] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 783.325613] env[61923]: DEBUG nova.compute.manager [req-b5507b69-2fc5-4373-aa68-bdf2e74d3c46 req-fc962a52-3878-4222-8d84-ef45700a6132 service nova] [instance: b497f3f3-001f-4e5e-b6b7-8e2006116856] Received event network-vif-deleted-fc48dbbd-d086-4537-ab13-11a91a9ac0fc {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 783.429966] env[61923]: INFO nova.compute.manager [-] [instance: ecedb77f-4a2a-4047-964f-80796954d847] Took 1.02 seconds to deallocate network for instance. [ 783.582771] env[61923]: DEBUG nova.scheduler.client.report [None req-6e6344ac-0cd2-410e-863e-3289b0c1fe3f tempest-ServersAdminTestJSON-678434436 tempest-ServersAdminTestJSON-678434436-project-member] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 783.615308] env[61923]: DEBUG oslo_concurrency.lockutils [None req-d3d04fc8-2bde-4c80-b2d6-dd8057edd757 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] Releasing lock "refresh_cache-b497f3f3-001f-4e5e-b6b7-8e2006116856" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 783.615762] env[61923]: DEBUG nova.compute.manager [None req-d3d04fc8-2bde-4c80-b2d6-dd8057edd757 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] [instance: b497f3f3-001f-4e5e-b6b7-8e2006116856] Start destroying the instance on the hypervisor. {{(pid=61923) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 783.615953] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-d3d04fc8-2bde-4c80-b2d6-dd8057edd757 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] [instance: b497f3f3-001f-4e5e-b6b7-8e2006116856] Destroying instance {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 783.616248] env[61923]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-23f3af5a-a5c8-4a40-858c-db8877121f98 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 783.626967] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5c52887-c2a9-4506-8cc5-87175f0c30dc {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.358591] env[61923]: ERROR nova.compute.manager [None req-1d78b71c-3449-4286-a17a-dc5396f589fe tempest-ServersAdminTestJSON-678434436 tempest-ServersAdminTestJSON-678434436-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 556dc345-8ad3-4e4f-8dd5-1d38e265adae, please check neutron logs for more information. [ 784.358591] env[61923]: ERROR nova.compute.manager Traceback (most recent call last): [ 784.358591] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 784.358591] env[61923]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 784.358591] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 784.358591] env[61923]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 784.358591] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 784.358591] env[61923]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 784.358591] env[61923]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 784.358591] env[61923]: ERROR nova.compute.manager self.force_reraise() [ 784.358591] env[61923]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 784.358591] env[61923]: ERROR nova.compute.manager raise self.value [ 784.358591] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 784.358591] env[61923]: ERROR nova.compute.manager updated_port = self._update_port( [ 784.358591] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 784.358591] env[61923]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 784.359495] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 784.359495] env[61923]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 784.359495] env[61923]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 556dc345-8ad3-4e4f-8dd5-1d38e265adae, please check neutron logs for more information. [ 784.359495] env[61923]: ERROR nova.compute.manager [ 784.359495] env[61923]: Traceback (most recent call last): [ 784.359495] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 784.359495] env[61923]: listener.cb(fileno) [ 784.359495] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 784.359495] env[61923]: result = function(*args, **kwargs) [ 784.359495] env[61923]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 784.359495] env[61923]: return func(*args, **kwargs) [ 784.359495] env[61923]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 784.359495] env[61923]: raise e [ 784.359495] env[61923]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 784.359495] env[61923]: nwinfo = self.network_api.allocate_for_instance( [ 784.359495] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 784.359495] env[61923]: created_port_ids = self._update_ports_for_instance( [ 784.359495] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 784.359495] env[61923]: with excutils.save_and_reraise_exception(): [ 784.359495] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 784.359495] env[61923]: self.force_reraise() [ 784.359495] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 784.359495] env[61923]: raise self.value [ 784.359495] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 784.359495] env[61923]: updated_port = self._update_port( [ 784.359495] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 784.359495] env[61923]: _ensure_no_port_binding_failure(port) [ 784.359495] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 784.359495] env[61923]: raise exception.PortBindingFailed(port_id=port['id']) [ 784.360146] env[61923]: nova.exception.PortBindingFailed: Binding failed for port 556dc345-8ad3-4e4f-8dd5-1d38e265adae, please check neutron logs for more information. [ 784.360146] env[61923]: Removing descriptor: 14 [ 784.360146] env[61923]: DEBUG nova.compute.manager [None req-1d78b71c-3449-4286-a17a-dc5396f589fe tempest-ServersAdminTestJSON-678434436 tempest-ServersAdminTestJSON-678434436-project-member] [instance: faff83a1-3532-4396-ac3f-600f86653749] Start spawning the instance on the hypervisor. {{(pid=61923) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 784.362672] env[61923]: DEBUG oslo_concurrency.lockutils [None req-33807679-7c2d-4470-9448-1b4dbb76ab45 tempest-ServerShowV254Test-921228709 tempest-ServerShowV254Test-921228709-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 784.363252] env[61923]: DEBUG oslo_concurrency.lockutils [None req-6e6344ac-0cd2-410e-863e-3289b0c1fe3f tempest-ServersAdminTestJSON-678434436 tempest-ServersAdminTestJSON-678434436-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.688s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 784.363674] env[61923]: DEBUG nova.compute.manager [None req-6e6344ac-0cd2-410e-863e-3289b0c1fe3f tempest-ServersAdminTestJSON-678434436 tempest-ServersAdminTestJSON-678434436-project-member] [instance: 3f298e29-69ff-46ac-8018-ceb3990b9848] Start building networks asynchronously for instance. {{(pid=61923) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 784.366119] env[61923]: DEBUG oslo_concurrency.lockutils [None req-37409b66-7d98-4871-81c4-a83dd0296328 tempest-ServerAddressesNegativeTestJSON-675891034 tempest-ServerAddressesNegativeTestJSON-675891034-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 25.710s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 784.367406] env[61923]: INFO nova.compute.claims [None req-37409b66-7d98-4871-81c4-a83dd0296328 tempest-ServerAddressesNegativeTestJSON-675891034 tempest-ServerAddressesNegativeTestJSON-675891034-project-member] [instance: 6fe76341-d578-4cb3-9dce-c025f0d798b8] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 784.381397] env[61923]: WARNING nova.virt.vmwareapi.vmops [None req-d3d04fc8-2bde-4c80-b2d6-dd8057edd757 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] [instance: b497f3f3-001f-4e5e-b6b7-8e2006116856] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance b497f3f3-001f-4e5e-b6b7-8e2006116856 could not be found. [ 784.381606] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-d3d04fc8-2bde-4c80-b2d6-dd8057edd757 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] [instance: b497f3f3-001f-4e5e-b6b7-8e2006116856] Instance destroyed {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 784.381786] env[61923]: INFO nova.compute.manager [None req-d3d04fc8-2bde-4c80-b2d6-dd8057edd757 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] [instance: b497f3f3-001f-4e5e-b6b7-8e2006116856] Took 0.77 seconds to destroy the instance on the hypervisor. [ 784.382033] env[61923]: DEBUG oslo.service.loopingcall [None req-d3d04fc8-2bde-4c80-b2d6-dd8057edd757 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61923) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 784.384819] env[61923]: DEBUG nova.compute.manager [-] [instance: b497f3f3-001f-4e5e-b6b7-8e2006116856] Deallocating network for instance {{(pid=61923) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 784.384819] env[61923]: DEBUG nova.network.neutron [-] [instance: b497f3f3-001f-4e5e-b6b7-8e2006116856] deallocate_for_instance() {{(pid=61923) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 784.391432] env[61923]: DEBUG nova.virt.hardware [None req-1d78b71c-3449-4286-a17a-dc5396f589fe tempest-ServersAdminTestJSON-678434436 tempest-ServersAdminTestJSON-678434436-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-29T20:07:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-29T20:07:35Z,direct_url=,disk_format='vmdk',id=0f153f63-ae0a-45b1-b7f5-7f9b673c947f,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='4e7b5e3b3c3443c8bd5c70f8a15739f5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-29T20:07:36Z,virtual_size=,visibility=), allow threads: False {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 784.392264] env[61923]: DEBUG nova.virt.hardware [None req-1d78b71c-3449-4286-a17a-dc5396f589fe tempest-ServersAdminTestJSON-678434436 tempest-ServersAdminTestJSON-678434436-project-member] Flavor limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 784.392264] env[61923]: DEBUG nova.virt.hardware [None req-1d78b71c-3449-4286-a17a-dc5396f589fe tempest-ServersAdminTestJSON-678434436 tempest-ServersAdminTestJSON-678434436-project-member] Image limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 784.392264] env[61923]: DEBUG nova.virt.hardware [None req-1d78b71c-3449-4286-a17a-dc5396f589fe tempest-ServersAdminTestJSON-678434436 tempest-ServersAdminTestJSON-678434436-project-member] Flavor pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 784.392264] env[61923]: DEBUG nova.virt.hardware [None req-1d78b71c-3449-4286-a17a-dc5396f589fe tempest-ServersAdminTestJSON-678434436 tempest-ServersAdminTestJSON-678434436-project-member] Image pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 784.392264] env[61923]: DEBUG nova.virt.hardware [None req-1d78b71c-3449-4286-a17a-dc5396f589fe tempest-ServersAdminTestJSON-678434436 tempest-ServersAdminTestJSON-678434436-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 784.392470] env[61923]: DEBUG nova.virt.hardware [None req-1d78b71c-3449-4286-a17a-dc5396f589fe tempest-ServersAdminTestJSON-678434436 tempest-ServersAdminTestJSON-678434436-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 784.392606] env[61923]: DEBUG nova.virt.hardware [None req-1d78b71c-3449-4286-a17a-dc5396f589fe tempest-ServersAdminTestJSON-678434436 tempest-ServersAdminTestJSON-678434436-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 784.392774] env[61923]: DEBUG nova.virt.hardware [None req-1d78b71c-3449-4286-a17a-dc5396f589fe tempest-ServersAdminTestJSON-678434436 tempest-ServersAdminTestJSON-678434436-project-member] Got 1 possible topologies {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 784.392935] env[61923]: DEBUG nova.virt.hardware [None req-1d78b71c-3449-4286-a17a-dc5396f589fe tempest-ServersAdminTestJSON-678434436 tempest-ServersAdminTestJSON-678434436-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 784.393118] env[61923]: DEBUG nova.virt.hardware [None req-1d78b71c-3449-4286-a17a-dc5396f589fe tempest-ServersAdminTestJSON-678434436 tempest-ServersAdminTestJSON-678434436-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 784.394358] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24db042e-d698-4b9b-8619-ca0a8ea58ab5 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.400937] env[61923]: DEBUG nova.network.neutron [-] [instance: b497f3f3-001f-4e5e-b6b7-8e2006116856] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 784.403841] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38ae75ad-d74b-4583-97d0-749faf354541 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.417466] env[61923]: ERROR nova.compute.manager [None req-1d78b71c-3449-4286-a17a-dc5396f589fe tempest-ServersAdminTestJSON-678434436 tempest-ServersAdminTestJSON-678434436-project-member] [instance: faff83a1-3532-4396-ac3f-600f86653749] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 556dc345-8ad3-4e4f-8dd5-1d38e265adae, please check neutron logs for more information. [ 784.417466] env[61923]: ERROR nova.compute.manager [instance: faff83a1-3532-4396-ac3f-600f86653749] Traceback (most recent call last): [ 784.417466] env[61923]: ERROR nova.compute.manager [instance: faff83a1-3532-4396-ac3f-600f86653749] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 784.417466] env[61923]: ERROR nova.compute.manager [instance: faff83a1-3532-4396-ac3f-600f86653749] yield resources [ 784.417466] env[61923]: ERROR nova.compute.manager [instance: faff83a1-3532-4396-ac3f-600f86653749] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 784.417466] env[61923]: ERROR nova.compute.manager [instance: faff83a1-3532-4396-ac3f-600f86653749] self.driver.spawn(context, instance, image_meta, [ 784.417466] env[61923]: ERROR nova.compute.manager [instance: faff83a1-3532-4396-ac3f-600f86653749] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 784.417466] env[61923]: ERROR nova.compute.manager [instance: faff83a1-3532-4396-ac3f-600f86653749] self._vmops.spawn(context, instance, image_meta, injected_files, [ 784.417466] env[61923]: ERROR nova.compute.manager [instance: faff83a1-3532-4396-ac3f-600f86653749] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 784.417466] env[61923]: ERROR nova.compute.manager [instance: faff83a1-3532-4396-ac3f-600f86653749] vm_ref = self.build_virtual_machine(instance, [ 784.417466] env[61923]: ERROR nova.compute.manager [instance: faff83a1-3532-4396-ac3f-600f86653749] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 784.417981] env[61923]: ERROR nova.compute.manager [instance: faff83a1-3532-4396-ac3f-600f86653749] vif_infos = vmwarevif.get_vif_info(self._session, [ 784.417981] env[61923]: ERROR nova.compute.manager [instance: faff83a1-3532-4396-ac3f-600f86653749] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 784.417981] env[61923]: ERROR nova.compute.manager [instance: faff83a1-3532-4396-ac3f-600f86653749] for vif in network_info: [ 784.417981] env[61923]: ERROR nova.compute.manager [instance: faff83a1-3532-4396-ac3f-600f86653749] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 784.417981] env[61923]: ERROR nova.compute.manager [instance: faff83a1-3532-4396-ac3f-600f86653749] return self._sync_wrapper(fn, *args, **kwargs) [ 784.417981] env[61923]: ERROR nova.compute.manager [instance: faff83a1-3532-4396-ac3f-600f86653749] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 784.417981] env[61923]: ERROR nova.compute.manager [instance: faff83a1-3532-4396-ac3f-600f86653749] self.wait() [ 784.417981] env[61923]: ERROR nova.compute.manager [instance: faff83a1-3532-4396-ac3f-600f86653749] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 784.417981] env[61923]: ERROR nova.compute.manager [instance: faff83a1-3532-4396-ac3f-600f86653749] self[:] = self._gt.wait() [ 784.417981] env[61923]: ERROR nova.compute.manager [instance: faff83a1-3532-4396-ac3f-600f86653749] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 784.417981] env[61923]: ERROR nova.compute.manager [instance: faff83a1-3532-4396-ac3f-600f86653749] return self._exit_event.wait() [ 784.417981] env[61923]: ERROR nova.compute.manager [instance: faff83a1-3532-4396-ac3f-600f86653749] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 784.417981] env[61923]: ERROR nova.compute.manager [instance: faff83a1-3532-4396-ac3f-600f86653749] current.throw(*self._exc) [ 784.418545] env[61923]: ERROR nova.compute.manager [instance: faff83a1-3532-4396-ac3f-600f86653749] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 784.418545] env[61923]: ERROR nova.compute.manager [instance: faff83a1-3532-4396-ac3f-600f86653749] result = function(*args, **kwargs) [ 784.418545] env[61923]: ERROR nova.compute.manager [instance: faff83a1-3532-4396-ac3f-600f86653749] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 784.418545] env[61923]: ERROR nova.compute.manager [instance: faff83a1-3532-4396-ac3f-600f86653749] return func(*args, **kwargs) [ 784.418545] env[61923]: ERROR nova.compute.manager [instance: faff83a1-3532-4396-ac3f-600f86653749] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 784.418545] env[61923]: ERROR nova.compute.manager [instance: faff83a1-3532-4396-ac3f-600f86653749] raise e [ 784.418545] env[61923]: ERROR nova.compute.manager [instance: faff83a1-3532-4396-ac3f-600f86653749] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 784.418545] env[61923]: ERROR nova.compute.manager [instance: faff83a1-3532-4396-ac3f-600f86653749] nwinfo = self.network_api.allocate_for_instance( [ 784.418545] env[61923]: ERROR nova.compute.manager [instance: faff83a1-3532-4396-ac3f-600f86653749] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 784.418545] env[61923]: ERROR nova.compute.manager [instance: faff83a1-3532-4396-ac3f-600f86653749] created_port_ids = self._update_ports_for_instance( [ 784.418545] env[61923]: ERROR nova.compute.manager [instance: faff83a1-3532-4396-ac3f-600f86653749] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 784.418545] env[61923]: ERROR nova.compute.manager [instance: faff83a1-3532-4396-ac3f-600f86653749] with excutils.save_and_reraise_exception(): [ 784.418545] env[61923]: ERROR nova.compute.manager [instance: faff83a1-3532-4396-ac3f-600f86653749] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 784.419079] env[61923]: ERROR nova.compute.manager [instance: faff83a1-3532-4396-ac3f-600f86653749] self.force_reraise() [ 784.419079] env[61923]: ERROR nova.compute.manager [instance: faff83a1-3532-4396-ac3f-600f86653749] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 784.419079] env[61923]: ERROR nova.compute.manager [instance: faff83a1-3532-4396-ac3f-600f86653749] raise self.value [ 784.419079] env[61923]: ERROR nova.compute.manager [instance: faff83a1-3532-4396-ac3f-600f86653749] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 784.419079] env[61923]: ERROR nova.compute.manager [instance: faff83a1-3532-4396-ac3f-600f86653749] updated_port = self._update_port( [ 784.419079] env[61923]: ERROR nova.compute.manager [instance: faff83a1-3532-4396-ac3f-600f86653749] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 784.419079] env[61923]: ERROR nova.compute.manager [instance: faff83a1-3532-4396-ac3f-600f86653749] _ensure_no_port_binding_failure(port) [ 784.419079] env[61923]: ERROR nova.compute.manager [instance: faff83a1-3532-4396-ac3f-600f86653749] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 784.419079] env[61923]: ERROR nova.compute.manager [instance: faff83a1-3532-4396-ac3f-600f86653749] raise exception.PortBindingFailed(port_id=port['id']) [ 784.419079] env[61923]: ERROR nova.compute.manager [instance: faff83a1-3532-4396-ac3f-600f86653749] nova.exception.PortBindingFailed: Binding failed for port 556dc345-8ad3-4e4f-8dd5-1d38e265adae, please check neutron logs for more information. [ 784.419079] env[61923]: ERROR nova.compute.manager [instance: faff83a1-3532-4396-ac3f-600f86653749] [ 784.419079] env[61923]: INFO nova.compute.manager [None req-1d78b71c-3449-4286-a17a-dc5396f589fe tempest-ServersAdminTestJSON-678434436 tempest-ServersAdminTestJSON-678434436-project-member] [instance: faff83a1-3532-4396-ac3f-600f86653749] Terminating instance [ 784.419896] env[61923]: DEBUG oslo_concurrency.lockutils [None req-1d78b71c-3449-4286-a17a-dc5396f589fe tempest-ServersAdminTestJSON-678434436 tempest-ServersAdminTestJSON-678434436-project-member] Acquiring lock "refresh_cache-faff83a1-3532-4396-ac3f-600f86653749" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 784.419896] env[61923]: DEBUG oslo_concurrency.lockutils [None req-1d78b71c-3449-4286-a17a-dc5396f589fe tempest-ServersAdminTestJSON-678434436 tempest-ServersAdminTestJSON-678434436-project-member] Acquired lock "refresh_cache-faff83a1-3532-4396-ac3f-600f86653749" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 784.419896] env[61923]: DEBUG nova.network.neutron [None req-1d78b71c-3449-4286-a17a-dc5396f589fe tempest-ServersAdminTestJSON-678434436 tempest-ServersAdminTestJSON-678434436-project-member] [instance: faff83a1-3532-4396-ac3f-600f86653749] Building network info cache for instance {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 784.871687] env[61923]: DEBUG nova.compute.utils [None req-6e6344ac-0cd2-410e-863e-3289b0c1fe3f tempest-ServersAdminTestJSON-678434436 tempest-ServersAdminTestJSON-678434436-project-member] Using /dev/sd instead of None {{(pid=61923) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 784.874997] env[61923]: DEBUG nova.compute.manager [None req-6e6344ac-0cd2-410e-863e-3289b0c1fe3f tempest-ServersAdminTestJSON-678434436 tempest-ServersAdminTestJSON-678434436-project-member] [instance: 3f298e29-69ff-46ac-8018-ceb3990b9848] Allocating IP information in the background. {{(pid=61923) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 784.875180] env[61923]: DEBUG nova.network.neutron [None req-6e6344ac-0cd2-410e-863e-3289b0c1fe3f tempest-ServersAdminTestJSON-678434436 tempest-ServersAdminTestJSON-678434436-project-member] [instance: 3f298e29-69ff-46ac-8018-ceb3990b9848] allocate_for_instance() {{(pid=61923) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 784.903418] env[61923]: DEBUG nova.network.neutron [-] [instance: b497f3f3-001f-4e5e-b6b7-8e2006116856] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 784.925311] env[61923]: DEBUG nova.policy [None req-6e6344ac-0cd2-410e-863e-3289b0c1fe3f tempest-ServersAdminTestJSON-678434436 tempest-ServersAdminTestJSON-678434436-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '492cf67e8d934790b177501a16161a55', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '9d757357e4754e67853dca46d17d9eb7', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61923) authorize /opt/stack/nova/nova/policy.py:201}} [ 784.941667] env[61923]: DEBUG nova.network.neutron [None req-1d78b71c-3449-4286-a17a-dc5396f589fe tempest-ServersAdminTestJSON-678434436 tempest-ServersAdminTestJSON-678434436-project-member] [instance: faff83a1-3532-4396-ac3f-600f86653749] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 785.090101] env[61923]: DEBUG nova.network.neutron [None req-1d78b71c-3449-4286-a17a-dc5396f589fe tempest-ServersAdminTestJSON-678434436 tempest-ServersAdminTestJSON-678434436-project-member] [instance: faff83a1-3532-4396-ac3f-600f86653749] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 785.314438] env[61923]: DEBUG nova.network.neutron [None req-6e6344ac-0cd2-410e-863e-3289b0c1fe3f tempest-ServersAdminTestJSON-678434436 tempest-ServersAdminTestJSON-678434436-project-member] [instance: 3f298e29-69ff-46ac-8018-ceb3990b9848] Successfully created port: bfaa52bf-dae4-491e-bd1e-681526f95731 {{(pid=61923) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 785.351476] env[61923]: DEBUG nova.compute.manager [req-e38b9897-4e1b-4e8f-88a8-a4dc2721effd req-4f8a4a3d-aaae-4d35-abf6-64e300cd8c97 service nova] [instance: faff83a1-3532-4396-ac3f-600f86653749] Received event network-changed-556dc345-8ad3-4e4f-8dd5-1d38e265adae {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 785.351679] env[61923]: DEBUG nova.compute.manager [req-e38b9897-4e1b-4e8f-88a8-a4dc2721effd req-4f8a4a3d-aaae-4d35-abf6-64e300cd8c97 service nova] [instance: faff83a1-3532-4396-ac3f-600f86653749] Refreshing instance network info cache due to event network-changed-556dc345-8ad3-4e4f-8dd5-1d38e265adae. {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 785.351863] env[61923]: DEBUG oslo_concurrency.lockutils [req-e38b9897-4e1b-4e8f-88a8-a4dc2721effd req-4f8a4a3d-aaae-4d35-abf6-64e300cd8c97 service nova] Acquiring lock "refresh_cache-faff83a1-3532-4396-ac3f-600f86653749" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 785.375949] env[61923]: DEBUG nova.compute.manager [None req-6e6344ac-0cd2-410e-863e-3289b0c1fe3f tempest-ServersAdminTestJSON-678434436 tempest-ServersAdminTestJSON-678434436-project-member] [instance: 3f298e29-69ff-46ac-8018-ceb3990b9848] Start building block device mappings for instance. {{(pid=61923) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 785.405697] env[61923]: INFO nova.compute.manager [-] [instance: b497f3f3-001f-4e5e-b6b7-8e2006116856] Took 1.02 seconds to deallocate network for instance. [ 785.407890] env[61923]: DEBUG nova.compute.claims [None req-d3d04fc8-2bde-4c80-b2d6-dd8057edd757 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] [instance: b497f3f3-001f-4e5e-b6b7-8e2006116856] Aborting claim: {{(pid=61923) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 785.408123] env[61923]: DEBUG oslo_concurrency.lockutils [None req-d3d04fc8-2bde-4c80-b2d6-dd8057edd757 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 785.593624] env[61923]: DEBUG oslo_concurrency.lockutils [None req-1d78b71c-3449-4286-a17a-dc5396f589fe tempest-ServersAdminTestJSON-678434436 tempest-ServersAdminTestJSON-678434436-project-member] Releasing lock "refresh_cache-faff83a1-3532-4396-ac3f-600f86653749" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 785.594041] env[61923]: DEBUG nova.compute.manager [None req-1d78b71c-3449-4286-a17a-dc5396f589fe tempest-ServersAdminTestJSON-678434436 tempest-ServersAdminTestJSON-678434436-project-member] [instance: faff83a1-3532-4396-ac3f-600f86653749] Start destroying the instance on the hypervisor. {{(pid=61923) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 785.594235] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-1d78b71c-3449-4286-a17a-dc5396f589fe tempest-ServersAdminTestJSON-678434436 tempest-ServersAdminTestJSON-678434436-project-member] [instance: faff83a1-3532-4396-ac3f-600f86653749] Destroying instance {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 785.594787] env[61923]: DEBUG oslo_concurrency.lockutils [req-e38b9897-4e1b-4e8f-88a8-a4dc2721effd req-4f8a4a3d-aaae-4d35-abf6-64e300cd8c97 service nova] Acquired lock "refresh_cache-faff83a1-3532-4396-ac3f-600f86653749" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 785.594960] env[61923]: DEBUG nova.network.neutron [req-e38b9897-4e1b-4e8f-88a8-a4dc2721effd req-4f8a4a3d-aaae-4d35-abf6-64e300cd8c97 service nova] [instance: faff83a1-3532-4396-ac3f-600f86653749] Refreshing network info cache for port 556dc345-8ad3-4e4f-8dd5-1d38e265adae {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 785.595968] env[61923]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-f5a6188e-4d82-43ca-ae36-e00dee6ec2c2 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.607160] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3383be72-69e3-481b-ace5-66885e21c1b4 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.632596] env[61923]: WARNING nova.virt.vmwareapi.vmops [None req-1d78b71c-3449-4286-a17a-dc5396f589fe tempest-ServersAdminTestJSON-678434436 tempest-ServersAdminTestJSON-678434436-project-member] [instance: faff83a1-3532-4396-ac3f-600f86653749] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance faff83a1-3532-4396-ac3f-600f86653749 could not be found. [ 785.632884] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-1d78b71c-3449-4286-a17a-dc5396f589fe tempest-ServersAdminTestJSON-678434436 tempest-ServersAdminTestJSON-678434436-project-member] [instance: faff83a1-3532-4396-ac3f-600f86653749] Instance destroyed {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 785.633113] env[61923]: INFO nova.compute.manager [None req-1d78b71c-3449-4286-a17a-dc5396f589fe tempest-ServersAdminTestJSON-678434436 tempest-ServersAdminTestJSON-678434436-project-member] [instance: faff83a1-3532-4396-ac3f-600f86653749] Took 0.04 seconds to destroy the instance on the hypervisor. [ 785.633351] env[61923]: DEBUG oslo.service.loopingcall [None req-1d78b71c-3449-4286-a17a-dc5396f589fe tempest-ServersAdminTestJSON-678434436 tempest-ServersAdminTestJSON-678434436-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61923) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 785.635762] env[61923]: DEBUG nova.compute.manager [-] [instance: faff83a1-3532-4396-ac3f-600f86653749] Deallocating network for instance {{(pid=61923) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 785.635903] env[61923]: DEBUG nova.network.neutron [-] [instance: faff83a1-3532-4396-ac3f-600f86653749] deallocate_for_instance() {{(pid=61923) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 785.648145] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5f34479-5446-4f22-a088-3d5534677ef2 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.655505] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-421224df-950e-46ce-b030-0bad0519a9a5 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.660198] env[61923]: DEBUG nova.network.neutron [-] [instance: faff83a1-3532-4396-ac3f-600f86653749] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 785.689656] env[61923]: DEBUG nova.network.neutron [-] [instance: faff83a1-3532-4396-ac3f-600f86653749] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 785.691192] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be4f5709-5e18-498a-83e7-ea0d00eb3acb {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.699268] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-617bb7f0-4469-4765-a72f-b73dd21c9637 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.714383] env[61923]: DEBUG nova.compute.provider_tree [None req-37409b66-7d98-4871-81c4-a83dd0296328 tempest-ServerAddressesNegativeTestJSON-675891034 tempest-ServerAddressesNegativeTestJSON-675891034-project-member] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 786.124766] env[61923]: DEBUG nova.network.neutron [req-e38b9897-4e1b-4e8f-88a8-a4dc2721effd req-4f8a4a3d-aaae-4d35-abf6-64e300cd8c97 service nova] [instance: faff83a1-3532-4396-ac3f-600f86653749] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 786.196938] env[61923]: INFO nova.compute.manager [-] [instance: faff83a1-3532-4396-ac3f-600f86653749] Took 0.56 seconds to deallocate network for instance. [ 786.201157] env[61923]: DEBUG nova.compute.claims [None req-1d78b71c-3449-4286-a17a-dc5396f589fe tempest-ServersAdminTestJSON-678434436 tempest-ServersAdminTestJSON-678434436-project-member] [instance: faff83a1-3532-4396-ac3f-600f86653749] Aborting claim: {{(pid=61923) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 786.201157] env[61923]: DEBUG oslo_concurrency.lockutils [None req-1d78b71c-3449-4286-a17a-dc5396f589fe tempest-ServersAdminTestJSON-678434436 tempest-ServersAdminTestJSON-678434436-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 786.219020] env[61923]: DEBUG nova.scheduler.client.report [None req-37409b66-7d98-4871-81c4-a83dd0296328 tempest-ServerAddressesNegativeTestJSON-675891034 tempest-ServerAddressesNegativeTestJSON-675891034-project-member] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 786.294704] env[61923]: DEBUG nova.network.neutron [req-e38b9897-4e1b-4e8f-88a8-a4dc2721effd req-4f8a4a3d-aaae-4d35-abf6-64e300cd8c97 service nova] [instance: faff83a1-3532-4396-ac3f-600f86653749] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 786.380797] env[61923]: ERROR nova.compute.manager [None req-6e6344ac-0cd2-410e-863e-3289b0c1fe3f tempest-ServersAdminTestJSON-678434436 tempest-ServersAdminTestJSON-678434436-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port bfaa52bf-dae4-491e-bd1e-681526f95731, please check neutron logs for more information. [ 786.380797] env[61923]: ERROR nova.compute.manager Traceback (most recent call last): [ 786.380797] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 786.380797] env[61923]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 786.380797] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 786.380797] env[61923]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 786.380797] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 786.380797] env[61923]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 786.380797] env[61923]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 786.380797] env[61923]: ERROR nova.compute.manager self.force_reraise() [ 786.380797] env[61923]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 786.380797] env[61923]: ERROR nova.compute.manager raise self.value [ 786.380797] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 786.380797] env[61923]: ERROR nova.compute.manager updated_port = self._update_port( [ 786.380797] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 786.380797] env[61923]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 786.381488] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 786.381488] env[61923]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 786.381488] env[61923]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port bfaa52bf-dae4-491e-bd1e-681526f95731, please check neutron logs for more information. [ 786.381488] env[61923]: ERROR nova.compute.manager [ 786.381488] env[61923]: Traceback (most recent call last): [ 786.381488] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 786.381488] env[61923]: listener.cb(fileno) [ 786.381488] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 786.381488] env[61923]: result = function(*args, **kwargs) [ 786.381488] env[61923]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 786.381488] env[61923]: return func(*args, **kwargs) [ 786.381488] env[61923]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 786.381488] env[61923]: raise e [ 786.381488] env[61923]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 786.381488] env[61923]: nwinfo = self.network_api.allocate_for_instance( [ 786.381488] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 786.381488] env[61923]: created_port_ids = self._update_ports_for_instance( [ 786.381488] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 786.381488] env[61923]: with excutils.save_and_reraise_exception(): [ 786.381488] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 786.381488] env[61923]: self.force_reraise() [ 786.381488] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 786.381488] env[61923]: raise self.value [ 786.381488] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 786.381488] env[61923]: updated_port = self._update_port( [ 786.381488] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 786.381488] env[61923]: _ensure_no_port_binding_failure(port) [ 786.381488] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 786.381488] env[61923]: raise exception.PortBindingFailed(port_id=port['id']) [ 786.382217] env[61923]: nova.exception.PortBindingFailed: Binding failed for port bfaa52bf-dae4-491e-bd1e-681526f95731, please check neutron logs for more information. [ 786.382217] env[61923]: Removing descriptor: 14 [ 786.386983] env[61923]: DEBUG nova.compute.manager [None req-6e6344ac-0cd2-410e-863e-3289b0c1fe3f tempest-ServersAdminTestJSON-678434436 tempest-ServersAdminTestJSON-678434436-project-member] [instance: 3f298e29-69ff-46ac-8018-ceb3990b9848] Start spawning the instance on the hypervisor. {{(pid=61923) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 786.412836] env[61923]: DEBUG nova.virt.hardware [None req-6e6344ac-0cd2-410e-863e-3289b0c1fe3f tempest-ServersAdminTestJSON-678434436 tempest-ServersAdminTestJSON-678434436-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-29T20:07:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-29T20:07:35Z,direct_url=,disk_format='vmdk',id=0f153f63-ae0a-45b1-b7f5-7f9b673c947f,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='4e7b5e3b3c3443c8bd5c70f8a15739f5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-29T20:07:36Z,virtual_size=,visibility=), allow threads: False {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 786.413086] env[61923]: DEBUG nova.virt.hardware [None req-6e6344ac-0cd2-410e-863e-3289b0c1fe3f tempest-ServersAdminTestJSON-678434436 tempest-ServersAdminTestJSON-678434436-project-member] Flavor limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 786.413256] env[61923]: DEBUG nova.virt.hardware [None req-6e6344ac-0cd2-410e-863e-3289b0c1fe3f tempest-ServersAdminTestJSON-678434436 tempest-ServersAdminTestJSON-678434436-project-member] Image limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 786.413438] env[61923]: DEBUG nova.virt.hardware [None req-6e6344ac-0cd2-410e-863e-3289b0c1fe3f tempest-ServersAdminTestJSON-678434436 tempest-ServersAdminTestJSON-678434436-project-member] Flavor pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 786.413582] env[61923]: DEBUG nova.virt.hardware [None req-6e6344ac-0cd2-410e-863e-3289b0c1fe3f tempest-ServersAdminTestJSON-678434436 tempest-ServersAdminTestJSON-678434436-project-member] Image pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 786.413725] env[61923]: DEBUG nova.virt.hardware [None req-6e6344ac-0cd2-410e-863e-3289b0c1fe3f tempest-ServersAdminTestJSON-678434436 tempest-ServersAdminTestJSON-678434436-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 786.413928] env[61923]: DEBUG nova.virt.hardware [None req-6e6344ac-0cd2-410e-863e-3289b0c1fe3f tempest-ServersAdminTestJSON-678434436 tempest-ServersAdminTestJSON-678434436-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 786.414721] env[61923]: DEBUG nova.virt.hardware [None req-6e6344ac-0cd2-410e-863e-3289b0c1fe3f tempest-ServersAdminTestJSON-678434436 tempest-ServersAdminTestJSON-678434436-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 786.414959] env[61923]: DEBUG nova.virt.hardware [None req-6e6344ac-0cd2-410e-863e-3289b0c1fe3f tempest-ServersAdminTestJSON-678434436 tempest-ServersAdminTestJSON-678434436-project-member] Got 1 possible topologies {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 786.415149] env[61923]: DEBUG nova.virt.hardware [None req-6e6344ac-0cd2-410e-863e-3289b0c1fe3f tempest-ServersAdminTestJSON-678434436 tempest-ServersAdminTestJSON-678434436-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 786.415323] env[61923]: DEBUG nova.virt.hardware [None req-6e6344ac-0cd2-410e-863e-3289b0c1fe3f tempest-ServersAdminTestJSON-678434436 tempest-ServersAdminTestJSON-678434436-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 786.416262] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71a6a5d0-c23e-4e4c-acaf-ecf12ead178d {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.424386] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d25338bf-1f7a-4cb0-ae1d-de943dd556d4 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.439368] env[61923]: ERROR nova.compute.manager [None req-6e6344ac-0cd2-410e-863e-3289b0c1fe3f tempest-ServersAdminTestJSON-678434436 tempest-ServersAdminTestJSON-678434436-project-member] [instance: 3f298e29-69ff-46ac-8018-ceb3990b9848] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port bfaa52bf-dae4-491e-bd1e-681526f95731, please check neutron logs for more information. [ 786.439368] env[61923]: ERROR nova.compute.manager [instance: 3f298e29-69ff-46ac-8018-ceb3990b9848] Traceback (most recent call last): [ 786.439368] env[61923]: ERROR nova.compute.manager [instance: 3f298e29-69ff-46ac-8018-ceb3990b9848] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 786.439368] env[61923]: ERROR nova.compute.manager [instance: 3f298e29-69ff-46ac-8018-ceb3990b9848] yield resources [ 786.439368] env[61923]: ERROR nova.compute.manager [instance: 3f298e29-69ff-46ac-8018-ceb3990b9848] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 786.439368] env[61923]: ERROR nova.compute.manager [instance: 3f298e29-69ff-46ac-8018-ceb3990b9848] self.driver.spawn(context, instance, image_meta, [ 786.439368] env[61923]: ERROR nova.compute.manager [instance: 3f298e29-69ff-46ac-8018-ceb3990b9848] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 786.439368] env[61923]: ERROR nova.compute.manager [instance: 3f298e29-69ff-46ac-8018-ceb3990b9848] self._vmops.spawn(context, instance, image_meta, injected_files, [ 786.439368] env[61923]: ERROR nova.compute.manager [instance: 3f298e29-69ff-46ac-8018-ceb3990b9848] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 786.439368] env[61923]: ERROR nova.compute.manager [instance: 3f298e29-69ff-46ac-8018-ceb3990b9848] vm_ref = self.build_virtual_machine(instance, [ 786.439368] env[61923]: ERROR nova.compute.manager [instance: 3f298e29-69ff-46ac-8018-ceb3990b9848] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 786.439641] env[61923]: ERROR nova.compute.manager [instance: 3f298e29-69ff-46ac-8018-ceb3990b9848] vif_infos = vmwarevif.get_vif_info(self._session, [ 786.439641] env[61923]: ERROR nova.compute.manager [instance: 3f298e29-69ff-46ac-8018-ceb3990b9848] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 786.439641] env[61923]: ERROR nova.compute.manager [instance: 3f298e29-69ff-46ac-8018-ceb3990b9848] for vif in network_info: [ 786.439641] env[61923]: ERROR nova.compute.manager [instance: 3f298e29-69ff-46ac-8018-ceb3990b9848] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 786.439641] env[61923]: ERROR nova.compute.manager [instance: 3f298e29-69ff-46ac-8018-ceb3990b9848] return self._sync_wrapper(fn, *args, **kwargs) [ 786.439641] env[61923]: ERROR nova.compute.manager [instance: 3f298e29-69ff-46ac-8018-ceb3990b9848] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 786.439641] env[61923]: ERROR nova.compute.manager [instance: 3f298e29-69ff-46ac-8018-ceb3990b9848] self.wait() [ 786.439641] env[61923]: ERROR nova.compute.manager [instance: 3f298e29-69ff-46ac-8018-ceb3990b9848] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 786.439641] env[61923]: ERROR nova.compute.manager [instance: 3f298e29-69ff-46ac-8018-ceb3990b9848] self[:] = self._gt.wait() [ 786.439641] env[61923]: ERROR nova.compute.manager [instance: 3f298e29-69ff-46ac-8018-ceb3990b9848] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 786.439641] env[61923]: ERROR nova.compute.manager [instance: 3f298e29-69ff-46ac-8018-ceb3990b9848] return self._exit_event.wait() [ 786.439641] env[61923]: ERROR nova.compute.manager [instance: 3f298e29-69ff-46ac-8018-ceb3990b9848] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 786.439641] env[61923]: ERROR nova.compute.manager [instance: 3f298e29-69ff-46ac-8018-ceb3990b9848] current.throw(*self._exc) [ 786.439910] env[61923]: ERROR nova.compute.manager [instance: 3f298e29-69ff-46ac-8018-ceb3990b9848] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 786.439910] env[61923]: ERROR nova.compute.manager [instance: 3f298e29-69ff-46ac-8018-ceb3990b9848] result = function(*args, **kwargs) [ 786.439910] env[61923]: ERROR nova.compute.manager [instance: 3f298e29-69ff-46ac-8018-ceb3990b9848] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 786.439910] env[61923]: ERROR nova.compute.manager [instance: 3f298e29-69ff-46ac-8018-ceb3990b9848] return func(*args, **kwargs) [ 786.439910] env[61923]: ERROR nova.compute.manager [instance: 3f298e29-69ff-46ac-8018-ceb3990b9848] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 786.439910] env[61923]: ERROR nova.compute.manager [instance: 3f298e29-69ff-46ac-8018-ceb3990b9848] raise e [ 786.439910] env[61923]: ERROR nova.compute.manager [instance: 3f298e29-69ff-46ac-8018-ceb3990b9848] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 786.439910] env[61923]: ERROR nova.compute.manager [instance: 3f298e29-69ff-46ac-8018-ceb3990b9848] nwinfo = self.network_api.allocate_for_instance( [ 786.439910] env[61923]: ERROR nova.compute.manager [instance: 3f298e29-69ff-46ac-8018-ceb3990b9848] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 786.439910] env[61923]: ERROR nova.compute.manager [instance: 3f298e29-69ff-46ac-8018-ceb3990b9848] created_port_ids = self._update_ports_for_instance( [ 786.439910] env[61923]: ERROR nova.compute.manager [instance: 3f298e29-69ff-46ac-8018-ceb3990b9848] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 786.439910] env[61923]: ERROR nova.compute.manager [instance: 3f298e29-69ff-46ac-8018-ceb3990b9848] with excutils.save_and_reraise_exception(): [ 786.439910] env[61923]: ERROR nova.compute.manager [instance: 3f298e29-69ff-46ac-8018-ceb3990b9848] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 786.440584] env[61923]: ERROR nova.compute.manager [instance: 3f298e29-69ff-46ac-8018-ceb3990b9848] self.force_reraise() [ 786.440584] env[61923]: ERROR nova.compute.manager [instance: 3f298e29-69ff-46ac-8018-ceb3990b9848] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 786.440584] env[61923]: ERROR nova.compute.manager [instance: 3f298e29-69ff-46ac-8018-ceb3990b9848] raise self.value [ 786.440584] env[61923]: ERROR nova.compute.manager [instance: 3f298e29-69ff-46ac-8018-ceb3990b9848] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 786.440584] env[61923]: ERROR nova.compute.manager [instance: 3f298e29-69ff-46ac-8018-ceb3990b9848] updated_port = self._update_port( [ 786.440584] env[61923]: ERROR nova.compute.manager [instance: 3f298e29-69ff-46ac-8018-ceb3990b9848] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 786.440584] env[61923]: ERROR nova.compute.manager [instance: 3f298e29-69ff-46ac-8018-ceb3990b9848] _ensure_no_port_binding_failure(port) [ 786.440584] env[61923]: ERROR nova.compute.manager [instance: 3f298e29-69ff-46ac-8018-ceb3990b9848] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 786.440584] env[61923]: ERROR nova.compute.manager [instance: 3f298e29-69ff-46ac-8018-ceb3990b9848] raise exception.PortBindingFailed(port_id=port['id']) [ 786.440584] env[61923]: ERROR nova.compute.manager [instance: 3f298e29-69ff-46ac-8018-ceb3990b9848] nova.exception.PortBindingFailed: Binding failed for port bfaa52bf-dae4-491e-bd1e-681526f95731, please check neutron logs for more information. [ 786.440584] env[61923]: ERROR nova.compute.manager [instance: 3f298e29-69ff-46ac-8018-ceb3990b9848] [ 786.440584] env[61923]: INFO nova.compute.manager [None req-6e6344ac-0cd2-410e-863e-3289b0c1fe3f tempest-ServersAdminTestJSON-678434436 tempest-ServersAdminTestJSON-678434436-project-member] [instance: 3f298e29-69ff-46ac-8018-ceb3990b9848] Terminating instance [ 786.442306] env[61923]: DEBUG oslo_concurrency.lockutils [None req-6e6344ac-0cd2-410e-863e-3289b0c1fe3f tempest-ServersAdminTestJSON-678434436 tempest-ServersAdminTestJSON-678434436-project-member] Acquiring lock "refresh_cache-3f298e29-69ff-46ac-8018-ceb3990b9848" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 786.442467] env[61923]: DEBUG oslo_concurrency.lockutils [None req-6e6344ac-0cd2-410e-863e-3289b0c1fe3f tempest-ServersAdminTestJSON-678434436 tempest-ServersAdminTestJSON-678434436-project-member] Acquired lock "refresh_cache-3f298e29-69ff-46ac-8018-ceb3990b9848" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 786.442632] env[61923]: DEBUG nova.network.neutron [None req-6e6344ac-0cd2-410e-863e-3289b0c1fe3f tempest-ServersAdminTestJSON-678434436 tempest-ServersAdminTestJSON-678434436-project-member] [instance: 3f298e29-69ff-46ac-8018-ceb3990b9848] Building network info cache for instance {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 786.722309] env[61923]: DEBUG oslo_concurrency.lockutils [None req-37409b66-7d98-4871-81c4-a83dd0296328 tempest-ServerAddressesNegativeTestJSON-675891034 tempest-ServerAddressesNegativeTestJSON-675891034-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.356s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 786.722879] env[61923]: DEBUG nova.compute.manager [None req-37409b66-7d98-4871-81c4-a83dd0296328 tempest-ServerAddressesNegativeTestJSON-675891034 tempest-ServerAddressesNegativeTestJSON-675891034-project-member] [instance: 6fe76341-d578-4cb3-9dce-c025f0d798b8] Start building networks asynchronously for instance. {{(pid=61923) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 786.725886] env[61923]: DEBUG oslo_concurrency.lockutils [None req-aed8d24a-43d8-4fbe-ac67-0dd6c37614c4 tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 27.003s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 786.727385] env[61923]: INFO nova.compute.claims [None req-aed8d24a-43d8-4fbe-ac67-0dd6c37614c4 tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] [instance: 4d709a3a-65af-4adf-9ad0-4d830a9518c2] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 786.796201] env[61923]: DEBUG oslo_concurrency.lockutils [req-e38b9897-4e1b-4e8f-88a8-a4dc2721effd req-4f8a4a3d-aaae-4d35-abf6-64e300cd8c97 service nova] Releasing lock "refresh_cache-faff83a1-3532-4396-ac3f-600f86653749" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 786.797046] env[61923]: DEBUG nova.compute.manager [req-e38b9897-4e1b-4e8f-88a8-a4dc2721effd req-4f8a4a3d-aaae-4d35-abf6-64e300cd8c97 service nova] [instance: faff83a1-3532-4396-ac3f-600f86653749] Received event network-vif-deleted-556dc345-8ad3-4e4f-8dd5-1d38e265adae {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 786.970233] env[61923]: DEBUG nova.network.neutron [None req-6e6344ac-0cd2-410e-863e-3289b0c1fe3f tempest-ServersAdminTestJSON-678434436 tempest-ServersAdminTestJSON-678434436-project-member] [instance: 3f298e29-69ff-46ac-8018-ceb3990b9848] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 787.046394] env[61923]: DEBUG nova.network.neutron [None req-6e6344ac-0cd2-410e-863e-3289b0c1fe3f tempest-ServersAdminTestJSON-678434436 tempest-ServersAdminTestJSON-678434436-project-member] [instance: 3f298e29-69ff-46ac-8018-ceb3990b9848] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 787.231427] env[61923]: DEBUG nova.compute.utils [None req-37409b66-7d98-4871-81c4-a83dd0296328 tempest-ServerAddressesNegativeTestJSON-675891034 tempest-ServerAddressesNegativeTestJSON-675891034-project-member] Using /dev/sd instead of None {{(pid=61923) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 787.235317] env[61923]: DEBUG nova.compute.manager [None req-37409b66-7d98-4871-81c4-a83dd0296328 tempest-ServerAddressesNegativeTestJSON-675891034 tempest-ServerAddressesNegativeTestJSON-675891034-project-member] [instance: 6fe76341-d578-4cb3-9dce-c025f0d798b8] Allocating IP information in the background. {{(pid=61923) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 787.235497] env[61923]: DEBUG nova.network.neutron [None req-37409b66-7d98-4871-81c4-a83dd0296328 tempest-ServerAddressesNegativeTestJSON-675891034 tempest-ServerAddressesNegativeTestJSON-675891034-project-member] [instance: 6fe76341-d578-4cb3-9dce-c025f0d798b8] allocate_for_instance() {{(pid=61923) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 787.273373] env[61923]: DEBUG nova.policy [None req-37409b66-7d98-4871-81c4-a83dd0296328 tempest-ServerAddressesNegativeTestJSON-675891034 tempest-ServerAddressesNegativeTestJSON-675891034-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'bd96c14aa22444dc8fa4b3a096ac4383', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4cbcf746907b438fb014d656b6c7d568', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61923) authorize /opt/stack/nova/nova/policy.py:201}} [ 787.384718] env[61923]: DEBUG nova.compute.manager [req-5618193d-296d-4726-9244-8531a5b149ed req-b90fa4ac-02fc-4a3a-8370-0bc9d876ce77 service nova] [instance: 3f298e29-69ff-46ac-8018-ceb3990b9848] Received event network-changed-bfaa52bf-dae4-491e-bd1e-681526f95731 {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 787.384989] env[61923]: DEBUG nova.compute.manager [req-5618193d-296d-4726-9244-8531a5b149ed req-b90fa4ac-02fc-4a3a-8370-0bc9d876ce77 service nova] [instance: 3f298e29-69ff-46ac-8018-ceb3990b9848] Refreshing instance network info cache due to event network-changed-bfaa52bf-dae4-491e-bd1e-681526f95731. {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 787.385360] env[61923]: DEBUG oslo_concurrency.lockutils [req-5618193d-296d-4726-9244-8531a5b149ed req-b90fa4ac-02fc-4a3a-8370-0bc9d876ce77 service nova] Acquiring lock "refresh_cache-3f298e29-69ff-46ac-8018-ceb3990b9848" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 787.548681] env[61923]: DEBUG oslo_concurrency.lockutils [None req-6e6344ac-0cd2-410e-863e-3289b0c1fe3f tempest-ServersAdminTestJSON-678434436 tempest-ServersAdminTestJSON-678434436-project-member] Releasing lock "refresh_cache-3f298e29-69ff-46ac-8018-ceb3990b9848" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 787.549601] env[61923]: DEBUG nova.compute.manager [None req-6e6344ac-0cd2-410e-863e-3289b0c1fe3f tempest-ServersAdminTestJSON-678434436 tempest-ServersAdminTestJSON-678434436-project-member] [instance: 3f298e29-69ff-46ac-8018-ceb3990b9848] Start destroying the instance on the hypervisor. {{(pid=61923) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 787.549601] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-6e6344ac-0cd2-410e-863e-3289b0c1fe3f tempest-ServersAdminTestJSON-678434436 tempest-ServersAdminTestJSON-678434436-project-member] [instance: 3f298e29-69ff-46ac-8018-ceb3990b9848] Destroying instance {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 787.549703] env[61923]: DEBUG oslo_concurrency.lockutils [req-5618193d-296d-4726-9244-8531a5b149ed req-b90fa4ac-02fc-4a3a-8370-0bc9d876ce77 service nova] Acquired lock "refresh_cache-3f298e29-69ff-46ac-8018-ceb3990b9848" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 787.549844] env[61923]: DEBUG nova.network.neutron [req-5618193d-296d-4726-9244-8531a5b149ed req-b90fa4ac-02fc-4a3a-8370-0bc9d876ce77 service nova] [instance: 3f298e29-69ff-46ac-8018-ceb3990b9848] Refreshing network info cache for port bfaa52bf-dae4-491e-bd1e-681526f95731 {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 787.550930] env[61923]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-9ac1bf7e-0538-4fe7-b9e9-ff04f9da3347 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.560547] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ae75a70-5ac2-44bc-9aef-ed44ff303e64 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.584134] env[61923]: WARNING nova.virt.vmwareapi.vmops [None req-6e6344ac-0cd2-410e-863e-3289b0c1fe3f tempest-ServersAdminTestJSON-678434436 tempest-ServersAdminTestJSON-678434436-project-member] [instance: 3f298e29-69ff-46ac-8018-ceb3990b9848] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 3f298e29-69ff-46ac-8018-ceb3990b9848 could not be found. [ 787.584384] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-6e6344ac-0cd2-410e-863e-3289b0c1fe3f tempest-ServersAdminTestJSON-678434436 tempest-ServersAdminTestJSON-678434436-project-member] [instance: 3f298e29-69ff-46ac-8018-ceb3990b9848] Instance destroyed {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 787.584561] env[61923]: INFO nova.compute.manager [None req-6e6344ac-0cd2-410e-863e-3289b0c1fe3f tempest-ServersAdminTestJSON-678434436 tempest-ServersAdminTestJSON-678434436-project-member] [instance: 3f298e29-69ff-46ac-8018-ceb3990b9848] Took 0.04 seconds to destroy the instance on the hypervisor. [ 787.584794] env[61923]: DEBUG oslo.service.loopingcall [None req-6e6344ac-0cd2-410e-863e-3289b0c1fe3f tempest-ServersAdminTestJSON-678434436 tempest-ServersAdminTestJSON-678434436-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61923) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 787.585012] env[61923]: DEBUG nova.compute.manager [-] [instance: 3f298e29-69ff-46ac-8018-ceb3990b9848] Deallocating network for instance {{(pid=61923) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 787.585180] env[61923]: DEBUG nova.network.neutron [-] [instance: 3f298e29-69ff-46ac-8018-ceb3990b9848] deallocate_for_instance() {{(pid=61923) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 787.602258] env[61923]: DEBUG nova.network.neutron [-] [instance: 3f298e29-69ff-46ac-8018-ceb3990b9848] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 787.632931] env[61923]: DEBUG nova.network.neutron [None req-37409b66-7d98-4871-81c4-a83dd0296328 tempest-ServerAddressesNegativeTestJSON-675891034 tempest-ServerAddressesNegativeTestJSON-675891034-project-member] [instance: 6fe76341-d578-4cb3-9dce-c025f0d798b8] Successfully created port: 4f871414-2810-4b6d-aaf8-2f56db898ba0 {{(pid=61923) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 787.738706] env[61923]: DEBUG nova.compute.manager [None req-37409b66-7d98-4871-81c4-a83dd0296328 tempest-ServerAddressesNegativeTestJSON-675891034 tempest-ServerAddressesNegativeTestJSON-675891034-project-member] [instance: 6fe76341-d578-4cb3-9dce-c025f0d798b8] Start building block device mappings for instance. {{(pid=61923) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 788.006089] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33106ee8-df2c-4e99-a1ce-470b08c3e851 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.012412] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df874f11-0766-4a02-af15-7a3151d9f18c {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.045079] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75fe7f10-245b-4f52-a614-6d178358af53 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.052806] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-112a3164-91bf-4871-a822-4b5d5dc60034 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.068167] env[61923]: DEBUG nova.compute.provider_tree [None req-aed8d24a-43d8-4fbe-ac67-0dd6c37614c4 tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 788.076697] env[61923]: DEBUG nova.network.neutron [req-5618193d-296d-4726-9244-8531a5b149ed req-b90fa4ac-02fc-4a3a-8370-0bc9d876ce77 service nova] [instance: 3f298e29-69ff-46ac-8018-ceb3990b9848] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 788.104870] env[61923]: DEBUG nova.network.neutron [-] [instance: 3f298e29-69ff-46ac-8018-ceb3990b9848] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 788.192395] env[61923]: DEBUG nova.network.neutron [req-5618193d-296d-4726-9244-8531a5b149ed req-b90fa4ac-02fc-4a3a-8370-0bc9d876ce77 service nova] [instance: 3f298e29-69ff-46ac-8018-ceb3990b9848] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 788.571819] env[61923]: DEBUG nova.scheduler.client.report [None req-aed8d24a-43d8-4fbe-ac67-0dd6c37614c4 tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 788.607669] env[61923]: INFO nova.compute.manager [-] [instance: 3f298e29-69ff-46ac-8018-ceb3990b9848] Took 1.02 seconds to deallocate network for instance. [ 788.611852] env[61923]: DEBUG nova.compute.claims [None req-6e6344ac-0cd2-410e-863e-3289b0c1fe3f tempest-ServersAdminTestJSON-678434436 tempest-ServersAdminTestJSON-678434436-project-member] [instance: 3f298e29-69ff-46ac-8018-ceb3990b9848] Aborting claim: {{(pid=61923) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 788.612042] env[61923]: DEBUG oslo_concurrency.lockutils [None req-6e6344ac-0cd2-410e-863e-3289b0c1fe3f tempest-ServersAdminTestJSON-678434436 tempest-ServersAdminTestJSON-678434436-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 788.694588] env[61923]: DEBUG oslo_concurrency.lockutils [req-5618193d-296d-4726-9244-8531a5b149ed req-b90fa4ac-02fc-4a3a-8370-0bc9d876ce77 service nova] Releasing lock "refresh_cache-3f298e29-69ff-46ac-8018-ceb3990b9848" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 788.694853] env[61923]: DEBUG nova.compute.manager [req-5618193d-296d-4726-9244-8531a5b149ed req-b90fa4ac-02fc-4a3a-8370-0bc9d876ce77 service nova] [instance: 3f298e29-69ff-46ac-8018-ceb3990b9848] Received event network-vif-deleted-bfaa52bf-dae4-491e-bd1e-681526f95731 {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 788.755122] env[61923]: DEBUG nova.compute.manager [None req-37409b66-7d98-4871-81c4-a83dd0296328 tempest-ServerAddressesNegativeTestJSON-675891034 tempest-ServerAddressesNegativeTestJSON-675891034-project-member] [instance: 6fe76341-d578-4cb3-9dce-c025f0d798b8] Start spawning the instance on the hypervisor. {{(pid=61923) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 788.777497] env[61923]: DEBUG nova.virt.hardware [None req-37409b66-7d98-4871-81c4-a83dd0296328 tempest-ServerAddressesNegativeTestJSON-675891034 tempest-ServerAddressesNegativeTestJSON-675891034-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-29T20:07:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-29T20:07:35Z,direct_url=,disk_format='vmdk',id=0f153f63-ae0a-45b1-b7f5-7f9b673c947f,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='4e7b5e3b3c3443c8bd5c70f8a15739f5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-29T20:07:36Z,virtual_size=,visibility=), allow threads: False {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 788.777738] env[61923]: DEBUG nova.virt.hardware [None req-37409b66-7d98-4871-81c4-a83dd0296328 tempest-ServerAddressesNegativeTestJSON-675891034 tempest-ServerAddressesNegativeTestJSON-675891034-project-member] Flavor limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 788.777888] env[61923]: DEBUG nova.virt.hardware [None req-37409b66-7d98-4871-81c4-a83dd0296328 tempest-ServerAddressesNegativeTestJSON-675891034 tempest-ServerAddressesNegativeTestJSON-675891034-project-member] Image limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 788.778078] env[61923]: DEBUG nova.virt.hardware [None req-37409b66-7d98-4871-81c4-a83dd0296328 tempest-ServerAddressesNegativeTestJSON-675891034 tempest-ServerAddressesNegativeTestJSON-675891034-project-member] Flavor pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 788.778228] env[61923]: DEBUG nova.virt.hardware [None req-37409b66-7d98-4871-81c4-a83dd0296328 tempest-ServerAddressesNegativeTestJSON-675891034 tempest-ServerAddressesNegativeTestJSON-675891034-project-member] Image pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 788.778372] env[61923]: DEBUG nova.virt.hardware [None req-37409b66-7d98-4871-81c4-a83dd0296328 tempest-ServerAddressesNegativeTestJSON-675891034 tempest-ServerAddressesNegativeTestJSON-675891034-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 788.778659] env[61923]: DEBUG nova.virt.hardware [None req-37409b66-7d98-4871-81c4-a83dd0296328 tempest-ServerAddressesNegativeTestJSON-675891034 tempest-ServerAddressesNegativeTestJSON-675891034-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 788.778821] env[61923]: DEBUG nova.virt.hardware [None req-37409b66-7d98-4871-81c4-a83dd0296328 tempest-ServerAddressesNegativeTestJSON-675891034 tempest-ServerAddressesNegativeTestJSON-675891034-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 788.778986] env[61923]: DEBUG nova.virt.hardware [None req-37409b66-7d98-4871-81c4-a83dd0296328 tempest-ServerAddressesNegativeTestJSON-675891034 tempest-ServerAddressesNegativeTestJSON-675891034-project-member] Got 1 possible topologies {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 788.779279] env[61923]: DEBUG nova.virt.hardware [None req-37409b66-7d98-4871-81c4-a83dd0296328 tempest-ServerAddressesNegativeTestJSON-675891034 tempest-ServerAddressesNegativeTestJSON-675891034-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 788.779494] env[61923]: DEBUG nova.virt.hardware [None req-37409b66-7d98-4871-81c4-a83dd0296328 tempest-ServerAddressesNegativeTestJSON-675891034 tempest-ServerAddressesNegativeTestJSON-675891034-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 788.780359] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce56a0ad-9c86-4d40-adbd-9d6880cc00c5 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.789121] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-917d2af8-6cdf-4285-84b8-3b939921c6ae {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.867385] env[61923]: ERROR nova.compute.manager [None req-37409b66-7d98-4871-81c4-a83dd0296328 tempest-ServerAddressesNegativeTestJSON-675891034 tempest-ServerAddressesNegativeTestJSON-675891034-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 4f871414-2810-4b6d-aaf8-2f56db898ba0, please check neutron logs for more information. [ 788.867385] env[61923]: ERROR nova.compute.manager Traceback (most recent call last): [ 788.867385] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 788.867385] env[61923]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 788.867385] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 788.867385] env[61923]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 788.867385] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 788.867385] env[61923]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 788.867385] env[61923]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 788.867385] env[61923]: ERROR nova.compute.manager self.force_reraise() [ 788.867385] env[61923]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 788.867385] env[61923]: ERROR nova.compute.manager raise self.value [ 788.867385] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 788.867385] env[61923]: ERROR nova.compute.manager updated_port = self._update_port( [ 788.867385] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 788.867385] env[61923]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 788.867748] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 788.867748] env[61923]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 788.867748] env[61923]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 4f871414-2810-4b6d-aaf8-2f56db898ba0, please check neutron logs for more information. [ 788.867748] env[61923]: ERROR nova.compute.manager [ 788.867748] env[61923]: Traceback (most recent call last): [ 788.867748] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 788.867748] env[61923]: listener.cb(fileno) [ 788.867748] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 788.867748] env[61923]: result = function(*args, **kwargs) [ 788.867748] env[61923]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 788.867748] env[61923]: return func(*args, **kwargs) [ 788.867748] env[61923]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 788.867748] env[61923]: raise e [ 788.867748] env[61923]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 788.867748] env[61923]: nwinfo = self.network_api.allocate_for_instance( [ 788.867748] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 788.867748] env[61923]: created_port_ids = self._update_ports_for_instance( [ 788.867748] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 788.867748] env[61923]: with excutils.save_and_reraise_exception(): [ 788.867748] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 788.867748] env[61923]: self.force_reraise() [ 788.867748] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 788.867748] env[61923]: raise self.value [ 788.867748] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 788.867748] env[61923]: updated_port = self._update_port( [ 788.867748] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 788.867748] env[61923]: _ensure_no_port_binding_failure(port) [ 788.867748] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 788.867748] env[61923]: raise exception.PortBindingFailed(port_id=port['id']) [ 788.868337] env[61923]: nova.exception.PortBindingFailed: Binding failed for port 4f871414-2810-4b6d-aaf8-2f56db898ba0, please check neutron logs for more information. [ 788.868337] env[61923]: Removing descriptor: 14 [ 788.868337] env[61923]: ERROR nova.compute.manager [None req-37409b66-7d98-4871-81c4-a83dd0296328 tempest-ServerAddressesNegativeTestJSON-675891034 tempest-ServerAddressesNegativeTestJSON-675891034-project-member] [instance: 6fe76341-d578-4cb3-9dce-c025f0d798b8] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 4f871414-2810-4b6d-aaf8-2f56db898ba0, please check neutron logs for more information. [ 788.868337] env[61923]: ERROR nova.compute.manager [instance: 6fe76341-d578-4cb3-9dce-c025f0d798b8] Traceback (most recent call last): [ 788.868337] env[61923]: ERROR nova.compute.manager [instance: 6fe76341-d578-4cb3-9dce-c025f0d798b8] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 788.868337] env[61923]: ERROR nova.compute.manager [instance: 6fe76341-d578-4cb3-9dce-c025f0d798b8] yield resources [ 788.868337] env[61923]: ERROR nova.compute.manager [instance: 6fe76341-d578-4cb3-9dce-c025f0d798b8] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 788.868337] env[61923]: ERROR nova.compute.manager [instance: 6fe76341-d578-4cb3-9dce-c025f0d798b8] self.driver.spawn(context, instance, image_meta, [ 788.868337] env[61923]: ERROR nova.compute.manager [instance: 6fe76341-d578-4cb3-9dce-c025f0d798b8] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 788.868337] env[61923]: ERROR nova.compute.manager [instance: 6fe76341-d578-4cb3-9dce-c025f0d798b8] self._vmops.spawn(context, instance, image_meta, injected_files, [ 788.868337] env[61923]: ERROR nova.compute.manager [instance: 6fe76341-d578-4cb3-9dce-c025f0d798b8] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 788.868337] env[61923]: ERROR nova.compute.manager [instance: 6fe76341-d578-4cb3-9dce-c025f0d798b8] vm_ref = self.build_virtual_machine(instance, [ 788.868609] env[61923]: ERROR nova.compute.manager [instance: 6fe76341-d578-4cb3-9dce-c025f0d798b8] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 788.868609] env[61923]: ERROR nova.compute.manager [instance: 6fe76341-d578-4cb3-9dce-c025f0d798b8] vif_infos = vmwarevif.get_vif_info(self._session, [ 788.868609] env[61923]: ERROR nova.compute.manager [instance: 6fe76341-d578-4cb3-9dce-c025f0d798b8] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 788.868609] env[61923]: ERROR nova.compute.manager [instance: 6fe76341-d578-4cb3-9dce-c025f0d798b8] for vif in network_info: [ 788.868609] env[61923]: ERROR nova.compute.manager [instance: 6fe76341-d578-4cb3-9dce-c025f0d798b8] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 788.868609] env[61923]: ERROR nova.compute.manager [instance: 6fe76341-d578-4cb3-9dce-c025f0d798b8] return self._sync_wrapper(fn, *args, **kwargs) [ 788.868609] env[61923]: ERROR nova.compute.manager [instance: 6fe76341-d578-4cb3-9dce-c025f0d798b8] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 788.868609] env[61923]: ERROR nova.compute.manager [instance: 6fe76341-d578-4cb3-9dce-c025f0d798b8] self.wait() [ 788.868609] env[61923]: ERROR nova.compute.manager [instance: 6fe76341-d578-4cb3-9dce-c025f0d798b8] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 788.868609] env[61923]: ERROR nova.compute.manager [instance: 6fe76341-d578-4cb3-9dce-c025f0d798b8] self[:] = self._gt.wait() [ 788.868609] env[61923]: ERROR nova.compute.manager [instance: 6fe76341-d578-4cb3-9dce-c025f0d798b8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 788.868609] env[61923]: ERROR nova.compute.manager [instance: 6fe76341-d578-4cb3-9dce-c025f0d798b8] return self._exit_event.wait() [ 788.868609] env[61923]: ERROR nova.compute.manager [instance: 6fe76341-d578-4cb3-9dce-c025f0d798b8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 788.868870] env[61923]: ERROR nova.compute.manager [instance: 6fe76341-d578-4cb3-9dce-c025f0d798b8] result = hub.switch() [ 788.868870] env[61923]: ERROR nova.compute.manager [instance: 6fe76341-d578-4cb3-9dce-c025f0d798b8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 788.868870] env[61923]: ERROR nova.compute.manager [instance: 6fe76341-d578-4cb3-9dce-c025f0d798b8] return self.greenlet.switch() [ 788.868870] env[61923]: ERROR nova.compute.manager [instance: 6fe76341-d578-4cb3-9dce-c025f0d798b8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 788.868870] env[61923]: ERROR nova.compute.manager [instance: 6fe76341-d578-4cb3-9dce-c025f0d798b8] result = function(*args, **kwargs) [ 788.868870] env[61923]: ERROR nova.compute.manager [instance: 6fe76341-d578-4cb3-9dce-c025f0d798b8] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 788.868870] env[61923]: ERROR nova.compute.manager [instance: 6fe76341-d578-4cb3-9dce-c025f0d798b8] return func(*args, **kwargs) [ 788.868870] env[61923]: ERROR nova.compute.manager [instance: 6fe76341-d578-4cb3-9dce-c025f0d798b8] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 788.868870] env[61923]: ERROR nova.compute.manager [instance: 6fe76341-d578-4cb3-9dce-c025f0d798b8] raise e [ 788.868870] env[61923]: ERROR nova.compute.manager [instance: 6fe76341-d578-4cb3-9dce-c025f0d798b8] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 788.868870] env[61923]: ERROR nova.compute.manager [instance: 6fe76341-d578-4cb3-9dce-c025f0d798b8] nwinfo = self.network_api.allocate_for_instance( [ 788.868870] env[61923]: ERROR nova.compute.manager [instance: 6fe76341-d578-4cb3-9dce-c025f0d798b8] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 788.868870] env[61923]: ERROR nova.compute.manager [instance: 6fe76341-d578-4cb3-9dce-c025f0d798b8] created_port_ids = self._update_ports_for_instance( [ 788.869149] env[61923]: ERROR nova.compute.manager [instance: 6fe76341-d578-4cb3-9dce-c025f0d798b8] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 788.869149] env[61923]: ERROR nova.compute.manager [instance: 6fe76341-d578-4cb3-9dce-c025f0d798b8] with excutils.save_and_reraise_exception(): [ 788.869149] env[61923]: ERROR nova.compute.manager [instance: 6fe76341-d578-4cb3-9dce-c025f0d798b8] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 788.869149] env[61923]: ERROR nova.compute.manager [instance: 6fe76341-d578-4cb3-9dce-c025f0d798b8] self.force_reraise() [ 788.869149] env[61923]: ERROR nova.compute.manager [instance: 6fe76341-d578-4cb3-9dce-c025f0d798b8] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 788.869149] env[61923]: ERROR nova.compute.manager [instance: 6fe76341-d578-4cb3-9dce-c025f0d798b8] raise self.value [ 788.869149] env[61923]: ERROR nova.compute.manager [instance: 6fe76341-d578-4cb3-9dce-c025f0d798b8] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 788.869149] env[61923]: ERROR nova.compute.manager [instance: 6fe76341-d578-4cb3-9dce-c025f0d798b8] updated_port = self._update_port( [ 788.869149] env[61923]: ERROR nova.compute.manager [instance: 6fe76341-d578-4cb3-9dce-c025f0d798b8] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 788.869149] env[61923]: ERROR nova.compute.manager [instance: 6fe76341-d578-4cb3-9dce-c025f0d798b8] _ensure_no_port_binding_failure(port) [ 788.869149] env[61923]: ERROR nova.compute.manager [instance: 6fe76341-d578-4cb3-9dce-c025f0d798b8] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 788.869149] env[61923]: ERROR nova.compute.manager [instance: 6fe76341-d578-4cb3-9dce-c025f0d798b8] raise exception.PortBindingFailed(port_id=port['id']) [ 788.869386] env[61923]: ERROR nova.compute.manager [instance: 6fe76341-d578-4cb3-9dce-c025f0d798b8] nova.exception.PortBindingFailed: Binding failed for port 4f871414-2810-4b6d-aaf8-2f56db898ba0, please check neutron logs for more information. [ 788.869386] env[61923]: ERROR nova.compute.manager [instance: 6fe76341-d578-4cb3-9dce-c025f0d798b8] [ 788.869386] env[61923]: INFO nova.compute.manager [None req-37409b66-7d98-4871-81c4-a83dd0296328 tempest-ServerAddressesNegativeTestJSON-675891034 tempest-ServerAddressesNegativeTestJSON-675891034-project-member] [instance: 6fe76341-d578-4cb3-9dce-c025f0d798b8] Terminating instance [ 788.871369] env[61923]: DEBUG oslo_concurrency.lockutils [None req-37409b66-7d98-4871-81c4-a83dd0296328 tempest-ServerAddressesNegativeTestJSON-675891034 tempest-ServerAddressesNegativeTestJSON-675891034-project-member] Acquiring lock "refresh_cache-6fe76341-d578-4cb3-9dce-c025f0d798b8" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 788.871557] env[61923]: DEBUG oslo_concurrency.lockutils [None req-37409b66-7d98-4871-81c4-a83dd0296328 tempest-ServerAddressesNegativeTestJSON-675891034 tempest-ServerAddressesNegativeTestJSON-675891034-project-member] Acquired lock "refresh_cache-6fe76341-d578-4cb3-9dce-c025f0d798b8" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 788.871740] env[61923]: DEBUG nova.network.neutron [None req-37409b66-7d98-4871-81c4-a83dd0296328 tempest-ServerAddressesNegativeTestJSON-675891034 tempest-ServerAddressesNegativeTestJSON-675891034-project-member] [instance: 6fe76341-d578-4cb3-9dce-c025f0d798b8] Building network info cache for instance {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 789.076901] env[61923]: DEBUG oslo_concurrency.lockutils [None req-aed8d24a-43d8-4fbe-ac67-0dd6c37614c4 tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.351s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 789.077465] env[61923]: DEBUG nova.compute.manager [None req-aed8d24a-43d8-4fbe-ac67-0dd6c37614c4 tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] [instance: 4d709a3a-65af-4adf-9ad0-4d830a9518c2] Start building networks asynchronously for instance. {{(pid=61923) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 789.081279] env[61923]: DEBUG oslo_concurrency.lockutils [None req-aed8d24a-43d8-4fbe-ac67-0dd6c37614c4 tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 27.479s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 789.082649] env[61923]: INFO nova.compute.claims [None req-aed8d24a-43d8-4fbe-ac67-0dd6c37614c4 tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] [instance: fe981195-1e2c-4c8d-8c53-327851f845c7] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 789.392092] env[61923]: DEBUG nova.network.neutron [None req-37409b66-7d98-4871-81c4-a83dd0296328 tempest-ServerAddressesNegativeTestJSON-675891034 tempest-ServerAddressesNegativeTestJSON-675891034-project-member] [instance: 6fe76341-d578-4cb3-9dce-c025f0d798b8] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 789.409800] env[61923]: DEBUG nova.compute.manager [req-6c848756-bccb-4647-836b-05fe788423f3 req-bb14318f-95ca-4cd0-9d64-08760f908c79 service nova] [instance: 6fe76341-d578-4cb3-9dce-c025f0d798b8] Received event network-changed-4f871414-2810-4b6d-aaf8-2f56db898ba0 {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 789.410007] env[61923]: DEBUG nova.compute.manager [req-6c848756-bccb-4647-836b-05fe788423f3 req-bb14318f-95ca-4cd0-9d64-08760f908c79 service nova] [instance: 6fe76341-d578-4cb3-9dce-c025f0d798b8] Refreshing instance network info cache due to event network-changed-4f871414-2810-4b6d-aaf8-2f56db898ba0. {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 789.410199] env[61923]: DEBUG oslo_concurrency.lockutils [req-6c848756-bccb-4647-836b-05fe788423f3 req-bb14318f-95ca-4cd0-9d64-08760f908c79 service nova] Acquiring lock "refresh_cache-6fe76341-d578-4cb3-9dce-c025f0d798b8" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 789.492172] env[61923]: DEBUG nova.network.neutron [None req-37409b66-7d98-4871-81c4-a83dd0296328 tempest-ServerAddressesNegativeTestJSON-675891034 tempest-ServerAddressesNegativeTestJSON-675891034-project-member] [instance: 6fe76341-d578-4cb3-9dce-c025f0d798b8] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 789.586534] env[61923]: DEBUG nova.compute.utils [None req-aed8d24a-43d8-4fbe-ac67-0dd6c37614c4 tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Using /dev/sd instead of None {{(pid=61923) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 789.591028] env[61923]: DEBUG nova.compute.manager [None req-aed8d24a-43d8-4fbe-ac67-0dd6c37614c4 tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] [instance: 4d709a3a-65af-4adf-9ad0-4d830a9518c2] Allocating IP information in the background. {{(pid=61923) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 789.591308] env[61923]: DEBUG nova.network.neutron [None req-aed8d24a-43d8-4fbe-ac67-0dd6c37614c4 tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] [instance: 4d709a3a-65af-4adf-9ad0-4d830a9518c2] allocate_for_instance() {{(pid=61923) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 789.640872] env[61923]: DEBUG nova.policy [None req-aed8d24a-43d8-4fbe-ac67-0dd6c37614c4 tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2631257787d44fcbac010b2bf1fd4b0e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '6e260342cd1b41ef8f0ceb15b6438daa', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61923) authorize /opt/stack/nova/nova/policy.py:201}} [ 789.899857] env[61923]: DEBUG nova.network.neutron [None req-aed8d24a-43d8-4fbe-ac67-0dd6c37614c4 tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] [instance: 4d709a3a-65af-4adf-9ad0-4d830a9518c2] Successfully created port: 89b347c2-d7a8-4e9a-bc5c-1468e948003b {{(pid=61923) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 789.996262] env[61923]: DEBUG oslo_concurrency.lockutils [None req-37409b66-7d98-4871-81c4-a83dd0296328 tempest-ServerAddressesNegativeTestJSON-675891034 tempest-ServerAddressesNegativeTestJSON-675891034-project-member] Releasing lock "refresh_cache-6fe76341-d578-4cb3-9dce-c025f0d798b8" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 789.996754] env[61923]: DEBUG nova.compute.manager [None req-37409b66-7d98-4871-81c4-a83dd0296328 tempest-ServerAddressesNegativeTestJSON-675891034 tempest-ServerAddressesNegativeTestJSON-675891034-project-member] [instance: 6fe76341-d578-4cb3-9dce-c025f0d798b8] Start destroying the instance on the hypervisor. {{(pid=61923) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 789.996955] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-37409b66-7d98-4871-81c4-a83dd0296328 tempest-ServerAddressesNegativeTestJSON-675891034 tempest-ServerAddressesNegativeTestJSON-675891034-project-member] [instance: 6fe76341-d578-4cb3-9dce-c025f0d798b8] Destroying instance {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 789.997290] env[61923]: DEBUG oslo_concurrency.lockutils [req-6c848756-bccb-4647-836b-05fe788423f3 req-bb14318f-95ca-4cd0-9d64-08760f908c79 service nova] Acquired lock "refresh_cache-6fe76341-d578-4cb3-9dce-c025f0d798b8" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 789.997784] env[61923]: DEBUG nova.network.neutron [req-6c848756-bccb-4647-836b-05fe788423f3 req-bb14318f-95ca-4cd0-9d64-08760f908c79 service nova] [instance: 6fe76341-d578-4cb3-9dce-c025f0d798b8] Refreshing network info cache for port 4f871414-2810-4b6d-aaf8-2f56db898ba0 {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 789.999180] env[61923]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-6b8b5642-7062-4f5f-b535-199aecc03ded {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.008522] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0793fa00-89a4-483e-a0a3-b23d5a7aefd4 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.032365] env[61923]: WARNING nova.virt.vmwareapi.vmops [None req-37409b66-7d98-4871-81c4-a83dd0296328 tempest-ServerAddressesNegativeTestJSON-675891034 tempest-ServerAddressesNegativeTestJSON-675891034-project-member] [instance: 6fe76341-d578-4cb3-9dce-c025f0d798b8] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 6fe76341-d578-4cb3-9dce-c025f0d798b8 could not be found. [ 790.032527] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-37409b66-7d98-4871-81c4-a83dd0296328 tempest-ServerAddressesNegativeTestJSON-675891034 tempest-ServerAddressesNegativeTestJSON-675891034-project-member] [instance: 6fe76341-d578-4cb3-9dce-c025f0d798b8] Instance destroyed {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 790.032707] env[61923]: INFO nova.compute.manager [None req-37409b66-7d98-4871-81c4-a83dd0296328 tempest-ServerAddressesNegativeTestJSON-675891034 tempest-ServerAddressesNegativeTestJSON-675891034-project-member] [instance: 6fe76341-d578-4cb3-9dce-c025f0d798b8] Took 0.04 seconds to destroy the instance on the hypervisor. [ 790.032940] env[61923]: DEBUG oslo.service.loopingcall [None req-37409b66-7d98-4871-81c4-a83dd0296328 tempest-ServerAddressesNegativeTestJSON-675891034 tempest-ServerAddressesNegativeTestJSON-675891034-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61923) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 790.033167] env[61923]: DEBUG nova.compute.manager [-] [instance: 6fe76341-d578-4cb3-9dce-c025f0d798b8] Deallocating network for instance {{(pid=61923) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 790.033259] env[61923]: DEBUG nova.network.neutron [-] [instance: 6fe76341-d578-4cb3-9dce-c025f0d798b8] deallocate_for_instance() {{(pid=61923) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 790.050393] env[61923]: DEBUG nova.network.neutron [-] [instance: 6fe76341-d578-4cb3-9dce-c025f0d798b8] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 790.094982] env[61923]: DEBUG nova.compute.manager [None req-aed8d24a-43d8-4fbe-ac67-0dd6c37614c4 tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] [instance: 4d709a3a-65af-4adf-9ad0-4d830a9518c2] Start building block device mappings for instance. {{(pid=61923) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 790.343071] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-afffaa67-8246-4a77-a844-0e2d7a024511 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.351709] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c835271-ee97-41f1-87e9-75dd605ffa04 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.383173] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-121435a2-6e37-4e49-b799-2b3724d91374 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.390493] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b53869cf-9fd9-4df8-b2d7-c46a67d79fbe {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.403953] env[61923]: DEBUG nova.compute.provider_tree [None req-aed8d24a-43d8-4fbe-ac67-0dd6c37614c4 tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 790.531551] env[61923]: DEBUG nova.network.neutron [req-6c848756-bccb-4647-836b-05fe788423f3 req-bb14318f-95ca-4cd0-9d64-08760f908c79 service nova] [instance: 6fe76341-d578-4cb3-9dce-c025f0d798b8] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 790.552589] env[61923]: DEBUG nova.network.neutron [-] [instance: 6fe76341-d578-4cb3-9dce-c025f0d798b8] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 790.720166] env[61923]: DEBUG nova.network.neutron [req-6c848756-bccb-4647-836b-05fe788423f3 req-bb14318f-95ca-4cd0-9d64-08760f908c79 service nova] [instance: 6fe76341-d578-4cb3-9dce-c025f0d798b8] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 790.910088] env[61923]: DEBUG nova.scheduler.client.report [None req-aed8d24a-43d8-4fbe-ac67-0dd6c37614c4 tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 791.006361] env[61923]: ERROR nova.compute.manager [None req-aed8d24a-43d8-4fbe-ac67-0dd6c37614c4 tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 89b347c2-d7a8-4e9a-bc5c-1468e948003b, please check neutron logs for more information. [ 791.006361] env[61923]: ERROR nova.compute.manager Traceback (most recent call last): [ 791.006361] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 791.006361] env[61923]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 791.006361] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 791.006361] env[61923]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 791.006361] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 791.006361] env[61923]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 791.006361] env[61923]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 791.006361] env[61923]: ERROR nova.compute.manager self.force_reraise() [ 791.006361] env[61923]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 791.006361] env[61923]: ERROR nova.compute.manager raise self.value [ 791.006361] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 791.006361] env[61923]: ERROR nova.compute.manager updated_port = self._update_port( [ 791.006361] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 791.006361] env[61923]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 791.006754] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 791.006754] env[61923]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 791.006754] env[61923]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 89b347c2-d7a8-4e9a-bc5c-1468e948003b, please check neutron logs for more information. [ 791.006754] env[61923]: ERROR nova.compute.manager [ 791.006754] env[61923]: Traceback (most recent call last): [ 791.006754] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 791.006754] env[61923]: listener.cb(fileno) [ 791.006754] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 791.006754] env[61923]: result = function(*args, **kwargs) [ 791.006754] env[61923]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 791.006754] env[61923]: return func(*args, **kwargs) [ 791.006754] env[61923]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 791.006754] env[61923]: raise e [ 791.006754] env[61923]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 791.006754] env[61923]: nwinfo = self.network_api.allocate_for_instance( [ 791.006754] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 791.006754] env[61923]: created_port_ids = self._update_ports_for_instance( [ 791.006754] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 791.006754] env[61923]: with excutils.save_and_reraise_exception(): [ 791.006754] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 791.006754] env[61923]: self.force_reraise() [ 791.006754] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 791.006754] env[61923]: raise self.value [ 791.006754] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 791.006754] env[61923]: updated_port = self._update_port( [ 791.006754] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 791.006754] env[61923]: _ensure_no_port_binding_failure(port) [ 791.006754] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 791.006754] env[61923]: raise exception.PortBindingFailed(port_id=port['id']) [ 791.007865] env[61923]: nova.exception.PortBindingFailed: Binding failed for port 89b347c2-d7a8-4e9a-bc5c-1468e948003b, please check neutron logs for more information. [ 791.007865] env[61923]: Removing descriptor: 14 [ 791.055269] env[61923]: INFO nova.compute.manager [-] [instance: 6fe76341-d578-4cb3-9dce-c025f0d798b8] Took 1.02 seconds to deallocate network for instance. [ 791.057672] env[61923]: DEBUG nova.compute.claims [None req-37409b66-7d98-4871-81c4-a83dd0296328 tempest-ServerAddressesNegativeTestJSON-675891034 tempest-ServerAddressesNegativeTestJSON-675891034-project-member] [instance: 6fe76341-d578-4cb3-9dce-c025f0d798b8] Aborting claim: {{(pid=61923) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 791.057853] env[61923]: DEBUG oslo_concurrency.lockutils [None req-37409b66-7d98-4871-81c4-a83dd0296328 tempest-ServerAddressesNegativeTestJSON-675891034 tempest-ServerAddressesNegativeTestJSON-675891034-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 791.106196] env[61923]: DEBUG nova.compute.manager [None req-aed8d24a-43d8-4fbe-ac67-0dd6c37614c4 tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] [instance: 4d709a3a-65af-4adf-9ad0-4d830a9518c2] Start spawning the instance on the hypervisor. {{(pid=61923) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 791.129995] env[61923]: DEBUG nova.virt.hardware [None req-aed8d24a-43d8-4fbe-ac67-0dd6c37614c4 tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-29T20:07:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-29T20:07:35Z,direct_url=,disk_format='vmdk',id=0f153f63-ae0a-45b1-b7f5-7f9b673c947f,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='4e7b5e3b3c3443c8bd5c70f8a15739f5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-29T20:07:36Z,virtual_size=,visibility=), allow threads: False {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 791.130260] env[61923]: DEBUG nova.virt.hardware [None req-aed8d24a-43d8-4fbe-ac67-0dd6c37614c4 tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Flavor limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 791.130413] env[61923]: DEBUG nova.virt.hardware [None req-aed8d24a-43d8-4fbe-ac67-0dd6c37614c4 tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Image limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 791.130592] env[61923]: DEBUG nova.virt.hardware [None req-aed8d24a-43d8-4fbe-ac67-0dd6c37614c4 tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Flavor pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 791.130735] env[61923]: DEBUG nova.virt.hardware [None req-aed8d24a-43d8-4fbe-ac67-0dd6c37614c4 tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Image pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 791.130872] env[61923]: DEBUG nova.virt.hardware [None req-aed8d24a-43d8-4fbe-ac67-0dd6c37614c4 tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 791.131090] env[61923]: DEBUG nova.virt.hardware [None req-aed8d24a-43d8-4fbe-ac67-0dd6c37614c4 tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 791.131255] env[61923]: DEBUG nova.virt.hardware [None req-aed8d24a-43d8-4fbe-ac67-0dd6c37614c4 tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 791.131451] env[61923]: DEBUG nova.virt.hardware [None req-aed8d24a-43d8-4fbe-ac67-0dd6c37614c4 tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Got 1 possible topologies {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 791.131623] env[61923]: DEBUG nova.virt.hardware [None req-aed8d24a-43d8-4fbe-ac67-0dd6c37614c4 tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 791.131792] env[61923]: DEBUG nova.virt.hardware [None req-aed8d24a-43d8-4fbe-ac67-0dd6c37614c4 tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 791.132636] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fae7209f-9e03-479e-bd46-d7a3519b1a79 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.140903] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c85e802b-d556-4285-b6a6-510b866f41b8 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.154342] env[61923]: ERROR nova.compute.manager [None req-aed8d24a-43d8-4fbe-ac67-0dd6c37614c4 tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] [instance: 4d709a3a-65af-4adf-9ad0-4d830a9518c2] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 89b347c2-d7a8-4e9a-bc5c-1468e948003b, please check neutron logs for more information. [ 791.154342] env[61923]: ERROR nova.compute.manager [instance: 4d709a3a-65af-4adf-9ad0-4d830a9518c2] Traceback (most recent call last): [ 791.154342] env[61923]: ERROR nova.compute.manager [instance: 4d709a3a-65af-4adf-9ad0-4d830a9518c2] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 791.154342] env[61923]: ERROR nova.compute.manager [instance: 4d709a3a-65af-4adf-9ad0-4d830a9518c2] yield resources [ 791.154342] env[61923]: ERROR nova.compute.manager [instance: 4d709a3a-65af-4adf-9ad0-4d830a9518c2] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 791.154342] env[61923]: ERROR nova.compute.manager [instance: 4d709a3a-65af-4adf-9ad0-4d830a9518c2] self.driver.spawn(context, instance, image_meta, [ 791.154342] env[61923]: ERROR nova.compute.manager [instance: 4d709a3a-65af-4adf-9ad0-4d830a9518c2] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 791.154342] env[61923]: ERROR nova.compute.manager [instance: 4d709a3a-65af-4adf-9ad0-4d830a9518c2] self._vmops.spawn(context, instance, image_meta, injected_files, [ 791.154342] env[61923]: ERROR nova.compute.manager [instance: 4d709a3a-65af-4adf-9ad0-4d830a9518c2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 791.154342] env[61923]: ERROR nova.compute.manager [instance: 4d709a3a-65af-4adf-9ad0-4d830a9518c2] vm_ref = self.build_virtual_machine(instance, [ 791.154342] env[61923]: ERROR nova.compute.manager [instance: 4d709a3a-65af-4adf-9ad0-4d830a9518c2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 791.154745] env[61923]: ERROR nova.compute.manager [instance: 4d709a3a-65af-4adf-9ad0-4d830a9518c2] vif_infos = vmwarevif.get_vif_info(self._session, [ 791.154745] env[61923]: ERROR nova.compute.manager [instance: 4d709a3a-65af-4adf-9ad0-4d830a9518c2] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 791.154745] env[61923]: ERROR nova.compute.manager [instance: 4d709a3a-65af-4adf-9ad0-4d830a9518c2] for vif in network_info: [ 791.154745] env[61923]: ERROR nova.compute.manager [instance: 4d709a3a-65af-4adf-9ad0-4d830a9518c2] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 791.154745] env[61923]: ERROR nova.compute.manager [instance: 4d709a3a-65af-4adf-9ad0-4d830a9518c2] return self._sync_wrapper(fn, *args, **kwargs) [ 791.154745] env[61923]: ERROR nova.compute.manager [instance: 4d709a3a-65af-4adf-9ad0-4d830a9518c2] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 791.154745] env[61923]: ERROR nova.compute.manager [instance: 4d709a3a-65af-4adf-9ad0-4d830a9518c2] self.wait() [ 791.154745] env[61923]: ERROR nova.compute.manager [instance: 4d709a3a-65af-4adf-9ad0-4d830a9518c2] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 791.154745] env[61923]: ERROR nova.compute.manager [instance: 4d709a3a-65af-4adf-9ad0-4d830a9518c2] self[:] = self._gt.wait() [ 791.154745] env[61923]: ERROR nova.compute.manager [instance: 4d709a3a-65af-4adf-9ad0-4d830a9518c2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 791.154745] env[61923]: ERROR nova.compute.manager [instance: 4d709a3a-65af-4adf-9ad0-4d830a9518c2] return self._exit_event.wait() [ 791.154745] env[61923]: ERROR nova.compute.manager [instance: 4d709a3a-65af-4adf-9ad0-4d830a9518c2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 791.154745] env[61923]: ERROR nova.compute.manager [instance: 4d709a3a-65af-4adf-9ad0-4d830a9518c2] current.throw(*self._exc) [ 791.155069] env[61923]: ERROR nova.compute.manager [instance: 4d709a3a-65af-4adf-9ad0-4d830a9518c2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 791.155069] env[61923]: ERROR nova.compute.manager [instance: 4d709a3a-65af-4adf-9ad0-4d830a9518c2] result = function(*args, **kwargs) [ 791.155069] env[61923]: ERROR nova.compute.manager [instance: 4d709a3a-65af-4adf-9ad0-4d830a9518c2] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 791.155069] env[61923]: ERROR nova.compute.manager [instance: 4d709a3a-65af-4adf-9ad0-4d830a9518c2] return func(*args, **kwargs) [ 791.155069] env[61923]: ERROR nova.compute.manager [instance: 4d709a3a-65af-4adf-9ad0-4d830a9518c2] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 791.155069] env[61923]: ERROR nova.compute.manager [instance: 4d709a3a-65af-4adf-9ad0-4d830a9518c2] raise e [ 791.155069] env[61923]: ERROR nova.compute.manager [instance: 4d709a3a-65af-4adf-9ad0-4d830a9518c2] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 791.155069] env[61923]: ERROR nova.compute.manager [instance: 4d709a3a-65af-4adf-9ad0-4d830a9518c2] nwinfo = self.network_api.allocate_for_instance( [ 791.155069] env[61923]: ERROR nova.compute.manager [instance: 4d709a3a-65af-4adf-9ad0-4d830a9518c2] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 791.155069] env[61923]: ERROR nova.compute.manager [instance: 4d709a3a-65af-4adf-9ad0-4d830a9518c2] created_port_ids = self._update_ports_for_instance( [ 791.155069] env[61923]: ERROR nova.compute.manager [instance: 4d709a3a-65af-4adf-9ad0-4d830a9518c2] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 791.155069] env[61923]: ERROR nova.compute.manager [instance: 4d709a3a-65af-4adf-9ad0-4d830a9518c2] with excutils.save_and_reraise_exception(): [ 791.155069] env[61923]: ERROR nova.compute.manager [instance: 4d709a3a-65af-4adf-9ad0-4d830a9518c2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 791.155337] env[61923]: ERROR nova.compute.manager [instance: 4d709a3a-65af-4adf-9ad0-4d830a9518c2] self.force_reraise() [ 791.155337] env[61923]: ERROR nova.compute.manager [instance: 4d709a3a-65af-4adf-9ad0-4d830a9518c2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 791.155337] env[61923]: ERROR nova.compute.manager [instance: 4d709a3a-65af-4adf-9ad0-4d830a9518c2] raise self.value [ 791.155337] env[61923]: ERROR nova.compute.manager [instance: 4d709a3a-65af-4adf-9ad0-4d830a9518c2] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 791.155337] env[61923]: ERROR nova.compute.manager [instance: 4d709a3a-65af-4adf-9ad0-4d830a9518c2] updated_port = self._update_port( [ 791.155337] env[61923]: ERROR nova.compute.manager [instance: 4d709a3a-65af-4adf-9ad0-4d830a9518c2] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 791.155337] env[61923]: ERROR nova.compute.manager [instance: 4d709a3a-65af-4adf-9ad0-4d830a9518c2] _ensure_no_port_binding_failure(port) [ 791.155337] env[61923]: ERROR nova.compute.manager [instance: 4d709a3a-65af-4adf-9ad0-4d830a9518c2] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 791.155337] env[61923]: ERROR nova.compute.manager [instance: 4d709a3a-65af-4adf-9ad0-4d830a9518c2] raise exception.PortBindingFailed(port_id=port['id']) [ 791.155337] env[61923]: ERROR nova.compute.manager [instance: 4d709a3a-65af-4adf-9ad0-4d830a9518c2] nova.exception.PortBindingFailed: Binding failed for port 89b347c2-d7a8-4e9a-bc5c-1468e948003b, please check neutron logs for more information. [ 791.155337] env[61923]: ERROR nova.compute.manager [instance: 4d709a3a-65af-4adf-9ad0-4d830a9518c2] [ 791.155337] env[61923]: INFO nova.compute.manager [None req-aed8d24a-43d8-4fbe-ac67-0dd6c37614c4 tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] [instance: 4d709a3a-65af-4adf-9ad0-4d830a9518c2] Terminating instance [ 791.156609] env[61923]: DEBUG oslo_concurrency.lockutils [None req-aed8d24a-43d8-4fbe-ac67-0dd6c37614c4 tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Acquiring lock "refresh_cache-4d709a3a-65af-4adf-9ad0-4d830a9518c2" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 791.156765] env[61923]: DEBUG oslo_concurrency.lockutils [None req-aed8d24a-43d8-4fbe-ac67-0dd6c37614c4 tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Acquired lock "refresh_cache-4d709a3a-65af-4adf-9ad0-4d830a9518c2" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 791.156928] env[61923]: DEBUG nova.network.neutron [None req-aed8d24a-43d8-4fbe-ac67-0dd6c37614c4 tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] [instance: 4d709a3a-65af-4adf-9ad0-4d830a9518c2] Building network info cache for instance {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 791.223637] env[61923]: DEBUG oslo_concurrency.lockutils [req-6c848756-bccb-4647-836b-05fe788423f3 req-bb14318f-95ca-4cd0-9d64-08760f908c79 service nova] Releasing lock "refresh_cache-6fe76341-d578-4cb3-9dce-c025f0d798b8" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 791.223933] env[61923]: DEBUG nova.compute.manager [req-6c848756-bccb-4647-836b-05fe788423f3 req-bb14318f-95ca-4cd0-9d64-08760f908c79 service nova] [instance: 6fe76341-d578-4cb3-9dce-c025f0d798b8] Received event network-vif-deleted-4f871414-2810-4b6d-aaf8-2f56db898ba0 {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 791.413143] env[61923]: DEBUG oslo_concurrency.lockutils [None req-aed8d24a-43d8-4fbe-ac67-0dd6c37614c4 tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.332s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 791.413708] env[61923]: DEBUG nova.compute.manager [None req-aed8d24a-43d8-4fbe-ac67-0dd6c37614c4 tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] [instance: fe981195-1e2c-4c8d-8c53-327851f845c7] Start building networks asynchronously for instance. {{(pid=61923) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 791.416275] env[61923]: DEBUG oslo_concurrency.lockutils [None req-a9ea97fa-71e5-4463-a1f3-ab69743d14a4 tempest-ServerActionsTestJSON-1295690556 tempest-ServerActionsTestJSON-1295690556-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 25.542s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 791.417772] env[61923]: INFO nova.compute.claims [None req-a9ea97fa-71e5-4463-a1f3-ab69743d14a4 tempest-ServerActionsTestJSON-1295690556 tempest-ServerActionsTestJSON-1295690556-project-member] [instance: ad3b3933-5bd9-444a-af7b-c81e868e0622] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 791.458624] env[61923]: DEBUG nova.compute.manager [req-36d1eca9-6df8-47e8-b79c-47ad641e2093 req-4b76002b-7208-4e80-ad65-f35e1e91ab2c service nova] [instance: 4d709a3a-65af-4adf-9ad0-4d830a9518c2] Received event network-changed-89b347c2-d7a8-4e9a-bc5c-1468e948003b {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 791.458624] env[61923]: DEBUG nova.compute.manager [req-36d1eca9-6df8-47e8-b79c-47ad641e2093 req-4b76002b-7208-4e80-ad65-f35e1e91ab2c service nova] [instance: 4d709a3a-65af-4adf-9ad0-4d830a9518c2] Refreshing instance network info cache due to event network-changed-89b347c2-d7a8-4e9a-bc5c-1468e948003b. {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 791.458624] env[61923]: DEBUG oslo_concurrency.lockutils [req-36d1eca9-6df8-47e8-b79c-47ad641e2093 req-4b76002b-7208-4e80-ad65-f35e1e91ab2c service nova] Acquiring lock "refresh_cache-4d709a3a-65af-4adf-9ad0-4d830a9518c2" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 791.679186] env[61923]: DEBUG nova.network.neutron [None req-aed8d24a-43d8-4fbe-ac67-0dd6c37614c4 tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] [instance: 4d709a3a-65af-4adf-9ad0-4d830a9518c2] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 791.785437] env[61923]: DEBUG nova.network.neutron [None req-aed8d24a-43d8-4fbe-ac67-0dd6c37614c4 tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] [instance: 4d709a3a-65af-4adf-9ad0-4d830a9518c2] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 791.922913] env[61923]: DEBUG nova.compute.utils [None req-aed8d24a-43d8-4fbe-ac67-0dd6c37614c4 tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Using /dev/sd instead of None {{(pid=61923) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 791.926781] env[61923]: DEBUG nova.compute.manager [None req-aed8d24a-43d8-4fbe-ac67-0dd6c37614c4 tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] [instance: fe981195-1e2c-4c8d-8c53-327851f845c7] Allocating IP information in the background. {{(pid=61923) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 791.926970] env[61923]: DEBUG nova.network.neutron [None req-aed8d24a-43d8-4fbe-ac67-0dd6c37614c4 tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] [instance: fe981195-1e2c-4c8d-8c53-327851f845c7] allocate_for_instance() {{(pid=61923) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 791.967301] env[61923]: DEBUG nova.policy [None req-aed8d24a-43d8-4fbe-ac67-0dd6c37614c4 tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2631257787d44fcbac010b2bf1fd4b0e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '6e260342cd1b41ef8f0ceb15b6438daa', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61923) authorize /opt/stack/nova/nova/policy.py:201}} [ 792.267276] env[61923]: DEBUG nova.network.neutron [None req-aed8d24a-43d8-4fbe-ac67-0dd6c37614c4 tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] [instance: fe981195-1e2c-4c8d-8c53-327851f845c7] Successfully created port: 5b85f4c9-ccc5-45a2-a383-7d3b903c5c61 {{(pid=61923) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 792.289059] env[61923]: DEBUG oslo_concurrency.lockutils [None req-aed8d24a-43d8-4fbe-ac67-0dd6c37614c4 tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Releasing lock "refresh_cache-4d709a3a-65af-4adf-9ad0-4d830a9518c2" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 792.289504] env[61923]: DEBUG nova.compute.manager [None req-aed8d24a-43d8-4fbe-ac67-0dd6c37614c4 tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] [instance: 4d709a3a-65af-4adf-9ad0-4d830a9518c2] Start destroying the instance on the hypervisor. {{(pid=61923) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 792.289739] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-aed8d24a-43d8-4fbe-ac67-0dd6c37614c4 tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] [instance: 4d709a3a-65af-4adf-9ad0-4d830a9518c2] Destroying instance {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 792.290063] env[61923]: DEBUG oslo_concurrency.lockutils [req-36d1eca9-6df8-47e8-b79c-47ad641e2093 req-4b76002b-7208-4e80-ad65-f35e1e91ab2c service nova] Acquired lock "refresh_cache-4d709a3a-65af-4adf-9ad0-4d830a9518c2" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 792.290250] env[61923]: DEBUG nova.network.neutron [req-36d1eca9-6df8-47e8-b79c-47ad641e2093 req-4b76002b-7208-4e80-ad65-f35e1e91ab2c service nova] [instance: 4d709a3a-65af-4adf-9ad0-4d830a9518c2] Refreshing network info cache for port 89b347c2-d7a8-4e9a-bc5c-1468e948003b {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 792.291359] env[61923]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-0bcd1557-2e6f-402d-9dff-ac2b3f952d68 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.302429] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86e26a99-0a14-4f56-96cb-d452876feb63 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.325599] env[61923]: WARNING nova.virt.vmwareapi.vmops [None req-aed8d24a-43d8-4fbe-ac67-0dd6c37614c4 tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] [instance: 4d709a3a-65af-4adf-9ad0-4d830a9518c2] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 4d709a3a-65af-4adf-9ad0-4d830a9518c2 could not be found. [ 792.325808] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-aed8d24a-43d8-4fbe-ac67-0dd6c37614c4 tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] [instance: 4d709a3a-65af-4adf-9ad0-4d830a9518c2] Instance destroyed {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 792.325986] env[61923]: INFO nova.compute.manager [None req-aed8d24a-43d8-4fbe-ac67-0dd6c37614c4 tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] [instance: 4d709a3a-65af-4adf-9ad0-4d830a9518c2] Took 0.04 seconds to destroy the instance on the hypervisor. [ 792.326241] env[61923]: DEBUG oslo.service.loopingcall [None req-aed8d24a-43d8-4fbe-ac67-0dd6c37614c4 tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61923) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 792.326693] env[61923]: DEBUG nova.compute.manager [-] [instance: 4d709a3a-65af-4adf-9ad0-4d830a9518c2] Deallocating network for instance {{(pid=61923) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 792.326794] env[61923]: DEBUG nova.network.neutron [-] [instance: 4d709a3a-65af-4adf-9ad0-4d830a9518c2] deallocate_for_instance() {{(pid=61923) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 792.343141] env[61923]: DEBUG nova.network.neutron [-] [instance: 4d709a3a-65af-4adf-9ad0-4d830a9518c2] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 792.428019] env[61923]: DEBUG nova.compute.manager [None req-aed8d24a-43d8-4fbe-ac67-0dd6c37614c4 tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] [instance: fe981195-1e2c-4c8d-8c53-327851f845c7] Start building block device mappings for instance. {{(pid=61923) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 792.680664] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c619d2ce-5802-45f9-8207-114457db0730 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.688342] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52cab39e-b07a-4963-8e4f-9f9c22637034 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.725518] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c971d26-0bab-4905-997b-71a10f914f8f {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.736030] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3a12ca4-1866-475a-a4dd-47a221064cd1 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.751714] env[61923]: DEBUG nova.compute.provider_tree [None req-a9ea97fa-71e5-4463-a1f3-ab69743d14a4 tempest-ServerActionsTestJSON-1295690556 tempest-ServerActionsTestJSON-1295690556-project-member] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 792.825328] env[61923]: DEBUG nova.network.neutron [req-36d1eca9-6df8-47e8-b79c-47ad641e2093 req-4b76002b-7208-4e80-ad65-f35e1e91ab2c service nova] [instance: 4d709a3a-65af-4adf-9ad0-4d830a9518c2] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 792.845349] env[61923]: DEBUG nova.network.neutron [-] [instance: 4d709a3a-65af-4adf-9ad0-4d830a9518c2] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 792.959579] env[61923]: DEBUG nova.network.neutron [req-36d1eca9-6df8-47e8-b79c-47ad641e2093 req-4b76002b-7208-4e80-ad65-f35e1e91ab2c service nova] [instance: 4d709a3a-65af-4adf-9ad0-4d830a9518c2] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 793.256497] env[61923]: DEBUG nova.scheduler.client.report [None req-a9ea97fa-71e5-4463-a1f3-ab69743d14a4 tempest-ServerActionsTestJSON-1295690556 tempest-ServerActionsTestJSON-1295690556-project-member] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 793.341303] env[61923]: ERROR nova.compute.manager [None req-aed8d24a-43d8-4fbe-ac67-0dd6c37614c4 tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 5b85f4c9-ccc5-45a2-a383-7d3b903c5c61, please check neutron logs for more information. [ 793.341303] env[61923]: ERROR nova.compute.manager Traceback (most recent call last): [ 793.341303] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 793.341303] env[61923]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 793.341303] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 793.341303] env[61923]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 793.341303] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 793.341303] env[61923]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 793.341303] env[61923]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 793.341303] env[61923]: ERROR nova.compute.manager self.force_reraise() [ 793.341303] env[61923]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 793.341303] env[61923]: ERROR nova.compute.manager raise self.value [ 793.341303] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 793.341303] env[61923]: ERROR nova.compute.manager updated_port = self._update_port( [ 793.341303] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 793.341303] env[61923]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 793.341762] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 793.341762] env[61923]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 793.341762] env[61923]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 5b85f4c9-ccc5-45a2-a383-7d3b903c5c61, please check neutron logs for more information. [ 793.341762] env[61923]: ERROR nova.compute.manager [ 793.341762] env[61923]: Traceback (most recent call last): [ 793.341762] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 793.341762] env[61923]: listener.cb(fileno) [ 793.341762] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 793.341762] env[61923]: result = function(*args, **kwargs) [ 793.341762] env[61923]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 793.341762] env[61923]: return func(*args, **kwargs) [ 793.341762] env[61923]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 793.341762] env[61923]: raise e [ 793.341762] env[61923]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 793.341762] env[61923]: nwinfo = self.network_api.allocate_for_instance( [ 793.341762] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 793.341762] env[61923]: created_port_ids = self._update_ports_for_instance( [ 793.341762] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 793.341762] env[61923]: with excutils.save_and_reraise_exception(): [ 793.341762] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 793.341762] env[61923]: self.force_reraise() [ 793.341762] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 793.341762] env[61923]: raise self.value [ 793.341762] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 793.341762] env[61923]: updated_port = self._update_port( [ 793.341762] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 793.341762] env[61923]: _ensure_no_port_binding_failure(port) [ 793.341762] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 793.341762] env[61923]: raise exception.PortBindingFailed(port_id=port['id']) [ 793.342525] env[61923]: nova.exception.PortBindingFailed: Binding failed for port 5b85f4c9-ccc5-45a2-a383-7d3b903c5c61, please check neutron logs for more information. [ 793.342525] env[61923]: Removing descriptor: 14 [ 793.348566] env[61923]: INFO nova.compute.manager [-] [instance: 4d709a3a-65af-4adf-9ad0-4d830a9518c2] Took 1.02 seconds to deallocate network for instance. [ 793.350937] env[61923]: DEBUG nova.compute.claims [None req-aed8d24a-43d8-4fbe-ac67-0dd6c37614c4 tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] [instance: 4d709a3a-65af-4adf-9ad0-4d830a9518c2] Aborting claim: {{(pid=61923) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 793.351158] env[61923]: DEBUG oslo_concurrency.lockutils [None req-aed8d24a-43d8-4fbe-ac67-0dd6c37614c4 tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 793.440534] env[61923]: DEBUG nova.compute.manager [None req-aed8d24a-43d8-4fbe-ac67-0dd6c37614c4 tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] [instance: fe981195-1e2c-4c8d-8c53-327851f845c7] Start spawning the instance on the hypervisor. {{(pid=61923) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 793.462195] env[61923]: DEBUG oslo_concurrency.lockutils [req-36d1eca9-6df8-47e8-b79c-47ad641e2093 req-4b76002b-7208-4e80-ad65-f35e1e91ab2c service nova] Releasing lock "refresh_cache-4d709a3a-65af-4adf-9ad0-4d830a9518c2" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 793.462455] env[61923]: DEBUG nova.compute.manager [req-36d1eca9-6df8-47e8-b79c-47ad641e2093 req-4b76002b-7208-4e80-ad65-f35e1e91ab2c service nova] [instance: 4d709a3a-65af-4adf-9ad0-4d830a9518c2] Received event network-vif-deleted-89b347c2-d7a8-4e9a-bc5c-1468e948003b {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 793.466227] env[61923]: DEBUG nova.virt.hardware [None req-aed8d24a-43d8-4fbe-ac67-0dd6c37614c4 tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-29T20:07:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-29T20:07:35Z,direct_url=,disk_format='vmdk',id=0f153f63-ae0a-45b1-b7f5-7f9b673c947f,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='4e7b5e3b3c3443c8bd5c70f8a15739f5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-29T20:07:36Z,virtual_size=,visibility=), allow threads: False {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 793.466507] env[61923]: DEBUG nova.virt.hardware [None req-aed8d24a-43d8-4fbe-ac67-0dd6c37614c4 tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Flavor limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 793.466745] env[61923]: DEBUG nova.virt.hardware [None req-aed8d24a-43d8-4fbe-ac67-0dd6c37614c4 tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Image limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 793.467014] env[61923]: DEBUG nova.virt.hardware [None req-aed8d24a-43d8-4fbe-ac67-0dd6c37614c4 tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Flavor pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 793.467213] env[61923]: DEBUG nova.virt.hardware [None req-aed8d24a-43d8-4fbe-ac67-0dd6c37614c4 tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Image pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 793.467374] env[61923]: DEBUG nova.virt.hardware [None req-aed8d24a-43d8-4fbe-ac67-0dd6c37614c4 tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 793.467617] env[61923]: DEBUG nova.virt.hardware [None req-aed8d24a-43d8-4fbe-ac67-0dd6c37614c4 tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 793.467797] env[61923]: DEBUG nova.virt.hardware [None req-aed8d24a-43d8-4fbe-ac67-0dd6c37614c4 tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 793.467991] env[61923]: DEBUG nova.virt.hardware [None req-aed8d24a-43d8-4fbe-ac67-0dd6c37614c4 tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Got 1 possible topologies {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 793.468208] env[61923]: DEBUG nova.virt.hardware [None req-aed8d24a-43d8-4fbe-ac67-0dd6c37614c4 tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 793.468411] env[61923]: DEBUG nova.virt.hardware [None req-aed8d24a-43d8-4fbe-ac67-0dd6c37614c4 tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 793.469377] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b30dc4ec-d208-4d5b-bdc0-93d633bd08d5 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.477764] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90692c57-413e-4a3a-8c44-a6a2f928609b {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.493063] env[61923]: ERROR nova.compute.manager [None req-aed8d24a-43d8-4fbe-ac67-0dd6c37614c4 tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] [instance: fe981195-1e2c-4c8d-8c53-327851f845c7] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 5b85f4c9-ccc5-45a2-a383-7d3b903c5c61, please check neutron logs for more information. [ 793.493063] env[61923]: ERROR nova.compute.manager [instance: fe981195-1e2c-4c8d-8c53-327851f845c7] Traceback (most recent call last): [ 793.493063] env[61923]: ERROR nova.compute.manager [instance: fe981195-1e2c-4c8d-8c53-327851f845c7] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 793.493063] env[61923]: ERROR nova.compute.manager [instance: fe981195-1e2c-4c8d-8c53-327851f845c7] yield resources [ 793.493063] env[61923]: ERROR nova.compute.manager [instance: fe981195-1e2c-4c8d-8c53-327851f845c7] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 793.493063] env[61923]: ERROR nova.compute.manager [instance: fe981195-1e2c-4c8d-8c53-327851f845c7] self.driver.spawn(context, instance, image_meta, [ 793.493063] env[61923]: ERROR nova.compute.manager [instance: fe981195-1e2c-4c8d-8c53-327851f845c7] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 793.493063] env[61923]: ERROR nova.compute.manager [instance: fe981195-1e2c-4c8d-8c53-327851f845c7] self._vmops.spawn(context, instance, image_meta, injected_files, [ 793.493063] env[61923]: ERROR nova.compute.manager [instance: fe981195-1e2c-4c8d-8c53-327851f845c7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 793.493063] env[61923]: ERROR nova.compute.manager [instance: fe981195-1e2c-4c8d-8c53-327851f845c7] vm_ref = self.build_virtual_machine(instance, [ 793.493063] env[61923]: ERROR nova.compute.manager [instance: fe981195-1e2c-4c8d-8c53-327851f845c7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 793.493653] env[61923]: ERROR nova.compute.manager [instance: fe981195-1e2c-4c8d-8c53-327851f845c7] vif_infos = vmwarevif.get_vif_info(self._session, [ 793.493653] env[61923]: ERROR nova.compute.manager [instance: fe981195-1e2c-4c8d-8c53-327851f845c7] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 793.493653] env[61923]: ERROR nova.compute.manager [instance: fe981195-1e2c-4c8d-8c53-327851f845c7] for vif in network_info: [ 793.493653] env[61923]: ERROR nova.compute.manager [instance: fe981195-1e2c-4c8d-8c53-327851f845c7] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 793.493653] env[61923]: ERROR nova.compute.manager [instance: fe981195-1e2c-4c8d-8c53-327851f845c7] return self._sync_wrapper(fn, *args, **kwargs) [ 793.493653] env[61923]: ERROR nova.compute.manager [instance: fe981195-1e2c-4c8d-8c53-327851f845c7] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 793.493653] env[61923]: ERROR nova.compute.manager [instance: fe981195-1e2c-4c8d-8c53-327851f845c7] self.wait() [ 793.493653] env[61923]: ERROR nova.compute.manager [instance: fe981195-1e2c-4c8d-8c53-327851f845c7] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 793.493653] env[61923]: ERROR nova.compute.manager [instance: fe981195-1e2c-4c8d-8c53-327851f845c7] self[:] = self._gt.wait() [ 793.493653] env[61923]: ERROR nova.compute.manager [instance: fe981195-1e2c-4c8d-8c53-327851f845c7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 793.493653] env[61923]: ERROR nova.compute.manager [instance: fe981195-1e2c-4c8d-8c53-327851f845c7] return self._exit_event.wait() [ 793.493653] env[61923]: ERROR nova.compute.manager [instance: fe981195-1e2c-4c8d-8c53-327851f845c7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 793.493653] env[61923]: ERROR nova.compute.manager [instance: fe981195-1e2c-4c8d-8c53-327851f845c7] current.throw(*self._exc) [ 793.493960] env[61923]: ERROR nova.compute.manager [instance: fe981195-1e2c-4c8d-8c53-327851f845c7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 793.493960] env[61923]: ERROR nova.compute.manager [instance: fe981195-1e2c-4c8d-8c53-327851f845c7] result = function(*args, **kwargs) [ 793.493960] env[61923]: ERROR nova.compute.manager [instance: fe981195-1e2c-4c8d-8c53-327851f845c7] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 793.493960] env[61923]: ERROR nova.compute.manager [instance: fe981195-1e2c-4c8d-8c53-327851f845c7] return func(*args, **kwargs) [ 793.493960] env[61923]: ERROR nova.compute.manager [instance: fe981195-1e2c-4c8d-8c53-327851f845c7] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 793.493960] env[61923]: ERROR nova.compute.manager [instance: fe981195-1e2c-4c8d-8c53-327851f845c7] raise e [ 793.493960] env[61923]: ERROR nova.compute.manager [instance: fe981195-1e2c-4c8d-8c53-327851f845c7] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 793.493960] env[61923]: ERROR nova.compute.manager [instance: fe981195-1e2c-4c8d-8c53-327851f845c7] nwinfo = self.network_api.allocate_for_instance( [ 793.493960] env[61923]: ERROR nova.compute.manager [instance: fe981195-1e2c-4c8d-8c53-327851f845c7] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 793.493960] env[61923]: ERROR nova.compute.manager [instance: fe981195-1e2c-4c8d-8c53-327851f845c7] created_port_ids = self._update_ports_for_instance( [ 793.493960] env[61923]: ERROR nova.compute.manager [instance: fe981195-1e2c-4c8d-8c53-327851f845c7] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 793.493960] env[61923]: ERROR nova.compute.manager [instance: fe981195-1e2c-4c8d-8c53-327851f845c7] with excutils.save_and_reraise_exception(): [ 793.493960] env[61923]: ERROR nova.compute.manager [instance: fe981195-1e2c-4c8d-8c53-327851f845c7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 793.494298] env[61923]: ERROR nova.compute.manager [instance: fe981195-1e2c-4c8d-8c53-327851f845c7] self.force_reraise() [ 793.494298] env[61923]: ERROR nova.compute.manager [instance: fe981195-1e2c-4c8d-8c53-327851f845c7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 793.494298] env[61923]: ERROR nova.compute.manager [instance: fe981195-1e2c-4c8d-8c53-327851f845c7] raise self.value [ 793.494298] env[61923]: ERROR nova.compute.manager [instance: fe981195-1e2c-4c8d-8c53-327851f845c7] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 793.494298] env[61923]: ERROR nova.compute.manager [instance: fe981195-1e2c-4c8d-8c53-327851f845c7] updated_port = self._update_port( [ 793.494298] env[61923]: ERROR nova.compute.manager [instance: fe981195-1e2c-4c8d-8c53-327851f845c7] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 793.494298] env[61923]: ERROR nova.compute.manager [instance: fe981195-1e2c-4c8d-8c53-327851f845c7] _ensure_no_port_binding_failure(port) [ 793.494298] env[61923]: ERROR nova.compute.manager [instance: fe981195-1e2c-4c8d-8c53-327851f845c7] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 793.494298] env[61923]: ERROR nova.compute.manager [instance: fe981195-1e2c-4c8d-8c53-327851f845c7] raise exception.PortBindingFailed(port_id=port['id']) [ 793.494298] env[61923]: ERROR nova.compute.manager [instance: fe981195-1e2c-4c8d-8c53-327851f845c7] nova.exception.PortBindingFailed: Binding failed for port 5b85f4c9-ccc5-45a2-a383-7d3b903c5c61, please check neutron logs for more information. [ 793.494298] env[61923]: ERROR nova.compute.manager [instance: fe981195-1e2c-4c8d-8c53-327851f845c7] [ 793.494298] env[61923]: INFO nova.compute.manager [None req-aed8d24a-43d8-4fbe-ac67-0dd6c37614c4 tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] [instance: fe981195-1e2c-4c8d-8c53-327851f845c7] Terminating instance [ 793.495526] env[61923]: DEBUG oslo_concurrency.lockutils [None req-aed8d24a-43d8-4fbe-ac67-0dd6c37614c4 tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Acquiring lock "refresh_cache-fe981195-1e2c-4c8d-8c53-327851f845c7" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 793.495694] env[61923]: DEBUG oslo_concurrency.lockutils [None req-aed8d24a-43d8-4fbe-ac67-0dd6c37614c4 tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Acquired lock "refresh_cache-fe981195-1e2c-4c8d-8c53-327851f845c7" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 793.495862] env[61923]: DEBUG nova.network.neutron [None req-aed8d24a-43d8-4fbe-ac67-0dd6c37614c4 tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] [instance: fe981195-1e2c-4c8d-8c53-327851f845c7] Building network info cache for instance {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 793.498215] env[61923]: DEBUG nova.compute.manager [req-fc386e65-b4e7-434f-8fa4-79e15ec321f8 req-57ac4144-8360-40a8-9175-845695e96f86 service nova] [instance: fe981195-1e2c-4c8d-8c53-327851f845c7] Received event network-changed-5b85f4c9-ccc5-45a2-a383-7d3b903c5c61 {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 793.498396] env[61923]: DEBUG nova.compute.manager [req-fc386e65-b4e7-434f-8fa4-79e15ec321f8 req-57ac4144-8360-40a8-9175-845695e96f86 service nova] [instance: fe981195-1e2c-4c8d-8c53-327851f845c7] Refreshing instance network info cache due to event network-changed-5b85f4c9-ccc5-45a2-a383-7d3b903c5c61. {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 793.498597] env[61923]: DEBUG oslo_concurrency.lockutils [req-fc386e65-b4e7-434f-8fa4-79e15ec321f8 req-57ac4144-8360-40a8-9175-845695e96f86 service nova] Acquiring lock "refresh_cache-fe981195-1e2c-4c8d-8c53-327851f845c7" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 793.762833] env[61923]: DEBUG oslo_concurrency.lockutils [None req-a9ea97fa-71e5-4463-a1f3-ab69743d14a4 tempest-ServerActionsTestJSON-1295690556 tempest-ServerActionsTestJSON-1295690556-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.345s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 793.762833] env[61923]: DEBUG nova.compute.manager [None req-a9ea97fa-71e5-4463-a1f3-ab69743d14a4 tempest-ServerActionsTestJSON-1295690556 tempest-ServerActionsTestJSON-1295690556-project-member] [instance: ad3b3933-5bd9-444a-af7b-c81e868e0622] Start building networks asynchronously for instance. {{(pid=61923) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 793.764528] env[61923]: DEBUG oslo_concurrency.lockutils [None req-0adbd41a-bb08-4b5a-bc10-ab9778329dd8 tempest-VolumesAdminNegativeTest-1422120048 tempest-VolumesAdminNegativeTest-1422120048-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 26.100s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 793.769922] env[61923]: INFO nova.compute.claims [None req-0adbd41a-bb08-4b5a-bc10-ab9778329dd8 tempest-VolumesAdminNegativeTest-1422120048 tempest-VolumesAdminNegativeTest-1422120048-project-member] [instance: 7bd3204b-5754-4ff2-b0ff-f30556dd72a1] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 794.012821] env[61923]: DEBUG nova.network.neutron [None req-aed8d24a-43d8-4fbe-ac67-0dd6c37614c4 tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] [instance: fe981195-1e2c-4c8d-8c53-327851f845c7] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 794.097581] env[61923]: DEBUG nova.network.neutron [None req-aed8d24a-43d8-4fbe-ac67-0dd6c37614c4 tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] [instance: fe981195-1e2c-4c8d-8c53-327851f845c7] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 794.275054] env[61923]: DEBUG nova.compute.utils [None req-a9ea97fa-71e5-4463-a1f3-ab69743d14a4 tempest-ServerActionsTestJSON-1295690556 tempest-ServerActionsTestJSON-1295690556-project-member] Using /dev/sd instead of None {{(pid=61923) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 794.275697] env[61923]: DEBUG nova.compute.manager [None req-a9ea97fa-71e5-4463-a1f3-ab69743d14a4 tempest-ServerActionsTestJSON-1295690556 tempest-ServerActionsTestJSON-1295690556-project-member] [instance: ad3b3933-5bd9-444a-af7b-c81e868e0622] Allocating IP information in the background. {{(pid=61923) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 794.275864] env[61923]: DEBUG nova.network.neutron [None req-a9ea97fa-71e5-4463-a1f3-ab69743d14a4 tempest-ServerActionsTestJSON-1295690556 tempest-ServerActionsTestJSON-1295690556-project-member] [instance: ad3b3933-5bd9-444a-af7b-c81e868e0622] allocate_for_instance() {{(pid=61923) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 794.312775] env[61923]: DEBUG nova.policy [None req-a9ea97fa-71e5-4463-a1f3-ab69743d14a4 tempest-ServerActionsTestJSON-1295690556 tempest-ServerActionsTestJSON-1295690556-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '493a1c0bd4e94e68a8ca2ff6b5f087ff', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '074bdb17890f4ec8983356dadd8661a0', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61923) authorize /opt/stack/nova/nova/policy.py:201}} [ 794.600290] env[61923]: DEBUG oslo_concurrency.lockutils [None req-aed8d24a-43d8-4fbe-ac67-0dd6c37614c4 tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Releasing lock "refresh_cache-fe981195-1e2c-4c8d-8c53-327851f845c7" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 794.600510] env[61923]: DEBUG nova.compute.manager [None req-aed8d24a-43d8-4fbe-ac67-0dd6c37614c4 tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] [instance: fe981195-1e2c-4c8d-8c53-327851f845c7] Start destroying the instance on the hypervisor. {{(pid=61923) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 794.600783] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-aed8d24a-43d8-4fbe-ac67-0dd6c37614c4 tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] [instance: fe981195-1e2c-4c8d-8c53-327851f845c7] Destroying instance {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 794.601131] env[61923]: DEBUG oslo_concurrency.lockutils [req-fc386e65-b4e7-434f-8fa4-79e15ec321f8 req-57ac4144-8360-40a8-9175-845695e96f86 service nova] Acquired lock "refresh_cache-fe981195-1e2c-4c8d-8c53-327851f845c7" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 794.601301] env[61923]: DEBUG nova.network.neutron [req-fc386e65-b4e7-434f-8fa4-79e15ec321f8 req-57ac4144-8360-40a8-9175-845695e96f86 service nova] [instance: fe981195-1e2c-4c8d-8c53-327851f845c7] Refreshing network info cache for port 5b85f4c9-ccc5-45a2-a383-7d3b903c5c61 {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 794.602388] env[61923]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-aab36dcd-cd6c-40fd-b338-2834a0e18d30 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.616268] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8b1e219-e1ac-42cc-a109-e1c5f1411a39 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.627280] env[61923]: DEBUG nova.network.neutron [None req-a9ea97fa-71e5-4463-a1f3-ab69743d14a4 tempest-ServerActionsTestJSON-1295690556 tempest-ServerActionsTestJSON-1295690556-project-member] [instance: ad3b3933-5bd9-444a-af7b-c81e868e0622] Successfully created port: d7bef7b0-fae5-4d49-917d-36c20321c890 {{(pid=61923) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 794.641457] env[61923]: WARNING nova.virt.vmwareapi.vmops [None req-aed8d24a-43d8-4fbe-ac67-0dd6c37614c4 tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] [instance: fe981195-1e2c-4c8d-8c53-327851f845c7] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance fe981195-1e2c-4c8d-8c53-327851f845c7 could not be found. [ 794.641703] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-aed8d24a-43d8-4fbe-ac67-0dd6c37614c4 tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] [instance: fe981195-1e2c-4c8d-8c53-327851f845c7] Instance destroyed {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 794.641886] env[61923]: INFO nova.compute.manager [None req-aed8d24a-43d8-4fbe-ac67-0dd6c37614c4 tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] [instance: fe981195-1e2c-4c8d-8c53-327851f845c7] Took 0.04 seconds to destroy the instance on the hypervisor. [ 794.642143] env[61923]: DEBUG oslo.service.loopingcall [None req-aed8d24a-43d8-4fbe-ac67-0dd6c37614c4 tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61923) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 794.642412] env[61923]: DEBUG nova.compute.manager [-] [instance: fe981195-1e2c-4c8d-8c53-327851f845c7] Deallocating network for instance {{(pid=61923) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 794.642521] env[61923]: DEBUG nova.network.neutron [-] [instance: fe981195-1e2c-4c8d-8c53-327851f845c7] deallocate_for_instance() {{(pid=61923) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 794.668077] env[61923]: DEBUG nova.network.neutron [-] [instance: fe981195-1e2c-4c8d-8c53-327851f845c7] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 794.779893] env[61923]: DEBUG nova.compute.manager [None req-a9ea97fa-71e5-4463-a1f3-ab69743d14a4 tempest-ServerActionsTestJSON-1295690556 tempest-ServerActionsTestJSON-1295690556-project-member] [instance: ad3b3933-5bd9-444a-af7b-c81e868e0622] Start building block device mappings for instance. {{(pid=61923) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 795.052981] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c865882d-b0bb-4c1c-a807-fb4e64259963 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.060524] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a63db32-6d4f-427f-be6a-ad9a763ca55d {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.092330] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f37cdd36-abac-4132-bf9c-bf9c6da29f63 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.099424] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-540eb2f7-f48a-4e85-85a0-fbb4fa5d291f {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.114556] env[61923]: DEBUG nova.compute.provider_tree [None req-0adbd41a-bb08-4b5a-bc10-ab9778329dd8 tempest-VolumesAdminNegativeTest-1422120048 tempest-VolumesAdminNegativeTest-1422120048-project-member] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 795.133815] env[61923]: DEBUG nova.network.neutron [req-fc386e65-b4e7-434f-8fa4-79e15ec321f8 req-57ac4144-8360-40a8-9175-845695e96f86 service nova] [instance: fe981195-1e2c-4c8d-8c53-327851f845c7] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 795.173849] env[61923]: DEBUG nova.network.neutron [-] [instance: fe981195-1e2c-4c8d-8c53-327851f845c7] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 795.439724] env[61923]: DEBUG nova.network.neutron [req-fc386e65-b4e7-434f-8fa4-79e15ec321f8 req-57ac4144-8360-40a8-9175-845695e96f86 service nova] [instance: fe981195-1e2c-4c8d-8c53-327851f845c7] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 795.617490] env[61923]: DEBUG nova.scheduler.client.report [None req-0adbd41a-bb08-4b5a-bc10-ab9778329dd8 tempest-VolumesAdminNegativeTest-1422120048 tempest-VolumesAdminNegativeTest-1422120048-project-member] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 795.676220] env[61923]: INFO nova.compute.manager [-] [instance: fe981195-1e2c-4c8d-8c53-327851f845c7] Took 1.03 seconds to deallocate network for instance. [ 795.678418] env[61923]: DEBUG nova.compute.claims [None req-aed8d24a-43d8-4fbe-ac67-0dd6c37614c4 tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] [instance: fe981195-1e2c-4c8d-8c53-327851f845c7] Aborting claim: {{(pid=61923) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 795.678819] env[61923]: DEBUG oslo_concurrency.lockutils [None req-aed8d24a-43d8-4fbe-ac67-0dd6c37614c4 tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 795.725965] env[61923]: DEBUG nova.compute.manager [req-7d2dcfbf-18f1-49b8-a812-ff3c25739c2d req-419a079c-80b4-4e13-bdf9-57b554e72d53 service nova] [instance: ad3b3933-5bd9-444a-af7b-c81e868e0622] Received event network-changed-d7bef7b0-fae5-4d49-917d-36c20321c890 {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 795.726261] env[61923]: DEBUG nova.compute.manager [req-7d2dcfbf-18f1-49b8-a812-ff3c25739c2d req-419a079c-80b4-4e13-bdf9-57b554e72d53 service nova] [instance: ad3b3933-5bd9-444a-af7b-c81e868e0622] Refreshing instance network info cache due to event network-changed-d7bef7b0-fae5-4d49-917d-36c20321c890. {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 795.726553] env[61923]: DEBUG oslo_concurrency.lockutils [req-7d2dcfbf-18f1-49b8-a812-ff3c25739c2d req-419a079c-80b4-4e13-bdf9-57b554e72d53 service nova] Acquiring lock "refresh_cache-ad3b3933-5bd9-444a-af7b-c81e868e0622" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 795.726766] env[61923]: DEBUG oslo_concurrency.lockutils [req-7d2dcfbf-18f1-49b8-a812-ff3c25739c2d req-419a079c-80b4-4e13-bdf9-57b554e72d53 service nova] Acquired lock "refresh_cache-ad3b3933-5bd9-444a-af7b-c81e868e0622" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 795.726995] env[61923]: DEBUG nova.network.neutron [req-7d2dcfbf-18f1-49b8-a812-ff3c25739c2d req-419a079c-80b4-4e13-bdf9-57b554e72d53 service nova] [instance: ad3b3933-5bd9-444a-af7b-c81e868e0622] Refreshing network info cache for port d7bef7b0-fae5-4d49-917d-36c20321c890 {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 795.789066] env[61923]: DEBUG nova.compute.manager [None req-a9ea97fa-71e5-4463-a1f3-ab69743d14a4 tempest-ServerActionsTestJSON-1295690556 tempest-ServerActionsTestJSON-1295690556-project-member] [instance: ad3b3933-5bd9-444a-af7b-c81e868e0622] Start spawning the instance on the hypervisor. {{(pid=61923) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 795.824038] env[61923]: DEBUG nova.virt.hardware [None req-a9ea97fa-71e5-4463-a1f3-ab69743d14a4 tempest-ServerActionsTestJSON-1295690556 tempest-ServerActionsTestJSON-1295690556-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-29T20:07:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-29T20:07:35Z,direct_url=,disk_format='vmdk',id=0f153f63-ae0a-45b1-b7f5-7f9b673c947f,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='4e7b5e3b3c3443c8bd5c70f8a15739f5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-29T20:07:36Z,virtual_size=,visibility=), allow threads: False {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 795.824038] env[61923]: DEBUG nova.virt.hardware [None req-a9ea97fa-71e5-4463-a1f3-ab69743d14a4 tempest-ServerActionsTestJSON-1295690556 tempest-ServerActionsTestJSON-1295690556-project-member] Flavor limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 795.824038] env[61923]: DEBUG nova.virt.hardware [None req-a9ea97fa-71e5-4463-a1f3-ab69743d14a4 tempest-ServerActionsTestJSON-1295690556 tempest-ServerActionsTestJSON-1295690556-project-member] Image limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 795.824319] env[61923]: DEBUG nova.virt.hardware [None req-a9ea97fa-71e5-4463-a1f3-ab69743d14a4 tempest-ServerActionsTestJSON-1295690556 tempest-ServerActionsTestJSON-1295690556-project-member] Flavor pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 795.824319] env[61923]: DEBUG nova.virt.hardware [None req-a9ea97fa-71e5-4463-a1f3-ab69743d14a4 tempest-ServerActionsTestJSON-1295690556 tempest-ServerActionsTestJSON-1295690556-project-member] Image pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 795.824319] env[61923]: DEBUG nova.virt.hardware [None req-a9ea97fa-71e5-4463-a1f3-ab69743d14a4 tempest-ServerActionsTestJSON-1295690556 tempest-ServerActionsTestJSON-1295690556-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 795.824319] env[61923]: DEBUG nova.virt.hardware [None req-a9ea97fa-71e5-4463-a1f3-ab69743d14a4 tempest-ServerActionsTestJSON-1295690556 tempest-ServerActionsTestJSON-1295690556-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 795.824504] env[61923]: DEBUG nova.virt.hardware [None req-a9ea97fa-71e5-4463-a1f3-ab69743d14a4 tempest-ServerActionsTestJSON-1295690556 tempest-ServerActionsTestJSON-1295690556-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 795.824538] env[61923]: DEBUG nova.virt.hardware [None req-a9ea97fa-71e5-4463-a1f3-ab69743d14a4 tempest-ServerActionsTestJSON-1295690556 tempest-ServerActionsTestJSON-1295690556-project-member] Got 1 possible topologies {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 795.825815] env[61923]: DEBUG nova.virt.hardware [None req-a9ea97fa-71e5-4463-a1f3-ab69743d14a4 tempest-ServerActionsTestJSON-1295690556 tempest-ServerActionsTestJSON-1295690556-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 795.826066] env[61923]: DEBUG nova.virt.hardware [None req-a9ea97fa-71e5-4463-a1f3-ab69743d14a4 tempest-ServerActionsTestJSON-1295690556 tempest-ServerActionsTestJSON-1295690556-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 795.826962] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d4b890b-9e2d-44da-a1f8-2a050c6c26bf {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.835213] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40104a16-eb96-43be-bff8-533b3f3c9e05 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 795.914968] env[61923]: ERROR nova.compute.manager [None req-a9ea97fa-71e5-4463-a1f3-ab69743d14a4 tempest-ServerActionsTestJSON-1295690556 tempest-ServerActionsTestJSON-1295690556-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port d7bef7b0-fae5-4d49-917d-36c20321c890, please check neutron logs for more information. [ 795.914968] env[61923]: ERROR nova.compute.manager Traceback (most recent call last): [ 795.914968] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 795.914968] env[61923]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 795.914968] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 795.914968] env[61923]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 795.914968] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 795.914968] env[61923]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 795.914968] env[61923]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 795.914968] env[61923]: ERROR nova.compute.manager self.force_reraise() [ 795.914968] env[61923]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 795.914968] env[61923]: ERROR nova.compute.manager raise self.value [ 795.914968] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 795.914968] env[61923]: ERROR nova.compute.manager updated_port = self._update_port( [ 795.914968] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 795.914968] env[61923]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 795.915526] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 795.915526] env[61923]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 795.915526] env[61923]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port d7bef7b0-fae5-4d49-917d-36c20321c890, please check neutron logs for more information. [ 795.915526] env[61923]: ERROR nova.compute.manager [ 795.915526] env[61923]: Traceback (most recent call last): [ 795.915526] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 795.915526] env[61923]: listener.cb(fileno) [ 795.915526] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 795.915526] env[61923]: result = function(*args, **kwargs) [ 795.915526] env[61923]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 795.915526] env[61923]: return func(*args, **kwargs) [ 795.915526] env[61923]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 795.915526] env[61923]: raise e [ 795.915526] env[61923]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 795.915526] env[61923]: nwinfo = self.network_api.allocate_for_instance( [ 795.915526] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 795.915526] env[61923]: created_port_ids = self._update_ports_for_instance( [ 795.915526] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 795.915526] env[61923]: with excutils.save_and_reraise_exception(): [ 795.915526] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 795.915526] env[61923]: self.force_reraise() [ 795.915526] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 795.915526] env[61923]: raise self.value [ 795.915526] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 795.915526] env[61923]: updated_port = self._update_port( [ 795.915526] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 795.915526] env[61923]: _ensure_no_port_binding_failure(port) [ 795.915526] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 795.915526] env[61923]: raise exception.PortBindingFailed(port_id=port['id']) [ 795.916685] env[61923]: nova.exception.PortBindingFailed: Binding failed for port d7bef7b0-fae5-4d49-917d-36c20321c890, please check neutron logs for more information. [ 795.916685] env[61923]: Removing descriptor: 14 [ 795.916685] env[61923]: ERROR nova.compute.manager [None req-a9ea97fa-71e5-4463-a1f3-ab69743d14a4 tempest-ServerActionsTestJSON-1295690556 tempest-ServerActionsTestJSON-1295690556-project-member] [instance: ad3b3933-5bd9-444a-af7b-c81e868e0622] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port d7bef7b0-fae5-4d49-917d-36c20321c890, please check neutron logs for more information. [ 795.916685] env[61923]: ERROR nova.compute.manager [instance: ad3b3933-5bd9-444a-af7b-c81e868e0622] Traceback (most recent call last): [ 795.916685] env[61923]: ERROR nova.compute.manager [instance: ad3b3933-5bd9-444a-af7b-c81e868e0622] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 795.916685] env[61923]: ERROR nova.compute.manager [instance: ad3b3933-5bd9-444a-af7b-c81e868e0622] yield resources [ 795.916685] env[61923]: ERROR nova.compute.manager [instance: ad3b3933-5bd9-444a-af7b-c81e868e0622] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 795.916685] env[61923]: ERROR nova.compute.manager [instance: ad3b3933-5bd9-444a-af7b-c81e868e0622] self.driver.spawn(context, instance, image_meta, [ 795.916685] env[61923]: ERROR nova.compute.manager [instance: ad3b3933-5bd9-444a-af7b-c81e868e0622] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 795.916685] env[61923]: ERROR nova.compute.manager [instance: ad3b3933-5bd9-444a-af7b-c81e868e0622] self._vmops.spawn(context, instance, image_meta, injected_files, [ 795.916685] env[61923]: ERROR nova.compute.manager [instance: ad3b3933-5bd9-444a-af7b-c81e868e0622] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 795.916685] env[61923]: ERROR nova.compute.manager [instance: ad3b3933-5bd9-444a-af7b-c81e868e0622] vm_ref = self.build_virtual_machine(instance, [ 795.916969] env[61923]: ERROR nova.compute.manager [instance: ad3b3933-5bd9-444a-af7b-c81e868e0622] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 795.916969] env[61923]: ERROR nova.compute.manager [instance: ad3b3933-5bd9-444a-af7b-c81e868e0622] vif_infos = vmwarevif.get_vif_info(self._session, [ 795.916969] env[61923]: ERROR nova.compute.manager [instance: ad3b3933-5bd9-444a-af7b-c81e868e0622] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 795.916969] env[61923]: ERROR nova.compute.manager [instance: ad3b3933-5bd9-444a-af7b-c81e868e0622] for vif in network_info: [ 795.916969] env[61923]: ERROR nova.compute.manager [instance: ad3b3933-5bd9-444a-af7b-c81e868e0622] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 795.916969] env[61923]: ERROR nova.compute.manager [instance: ad3b3933-5bd9-444a-af7b-c81e868e0622] return self._sync_wrapper(fn, *args, **kwargs) [ 795.916969] env[61923]: ERROR nova.compute.manager [instance: ad3b3933-5bd9-444a-af7b-c81e868e0622] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 795.916969] env[61923]: ERROR nova.compute.manager [instance: ad3b3933-5bd9-444a-af7b-c81e868e0622] self.wait() [ 795.916969] env[61923]: ERROR nova.compute.manager [instance: ad3b3933-5bd9-444a-af7b-c81e868e0622] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 795.916969] env[61923]: ERROR nova.compute.manager [instance: ad3b3933-5bd9-444a-af7b-c81e868e0622] self[:] = self._gt.wait() [ 795.916969] env[61923]: ERROR nova.compute.manager [instance: ad3b3933-5bd9-444a-af7b-c81e868e0622] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 795.916969] env[61923]: ERROR nova.compute.manager [instance: ad3b3933-5bd9-444a-af7b-c81e868e0622] return self._exit_event.wait() [ 795.916969] env[61923]: ERROR nova.compute.manager [instance: ad3b3933-5bd9-444a-af7b-c81e868e0622] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 795.917360] env[61923]: ERROR nova.compute.manager [instance: ad3b3933-5bd9-444a-af7b-c81e868e0622] result = hub.switch() [ 795.917360] env[61923]: ERROR nova.compute.manager [instance: ad3b3933-5bd9-444a-af7b-c81e868e0622] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 795.917360] env[61923]: ERROR nova.compute.manager [instance: ad3b3933-5bd9-444a-af7b-c81e868e0622] return self.greenlet.switch() [ 795.917360] env[61923]: ERROR nova.compute.manager [instance: ad3b3933-5bd9-444a-af7b-c81e868e0622] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 795.917360] env[61923]: ERROR nova.compute.manager [instance: ad3b3933-5bd9-444a-af7b-c81e868e0622] result = function(*args, **kwargs) [ 795.917360] env[61923]: ERROR nova.compute.manager [instance: ad3b3933-5bd9-444a-af7b-c81e868e0622] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 795.917360] env[61923]: ERROR nova.compute.manager [instance: ad3b3933-5bd9-444a-af7b-c81e868e0622] return func(*args, **kwargs) [ 795.917360] env[61923]: ERROR nova.compute.manager [instance: ad3b3933-5bd9-444a-af7b-c81e868e0622] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 795.917360] env[61923]: ERROR nova.compute.manager [instance: ad3b3933-5bd9-444a-af7b-c81e868e0622] raise e [ 795.917360] env[61923]: ERROR nova.compute.manager [instance: ad3b3933-5bd9-444a-af7b-c81e868e0622] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 795.917360] env[61923]: ERROR nova.compute.manager [instance: ad3b3933-5bd9-444a-af7b-c81e868e0622] nwinfo = self.network_api.allocate_for_instance( [ 795.917360] env[61923]: ERROR nova.compute.manager [instance: ad3b3933-5bd9-444a-af7b-c81e868e0622] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 795.917360] env[61923]: ERROR nova.compute.manager [instance: ad3b3933-5bd9-444a-af7b-c81e868e0622] created_port_ids = self._update_ports_for_instance( [ 795.917710] env[61923]: ERROR nova.compute.manager [instance: ad3b3933-5bd9-444a-af7b-c81e868e0622] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 795.917710] env[61923]: ERROR nova.compute.manager [instance: ad3b3933-5bd9-444a-af7b-c81e868e0622] with excutils.save_and_reraise_exception(): [ 795.917710] env[61923]: ERROR nova.compute.manager [instance: ad3b3933-5bd9-444a-af7b-c81e868e0622] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 795.917710] env[61923]: ERROR nova.compute.manager [instance: ad3b3933-5bd9-444a-af7b-c81e868e0622] self.force_reraise() [ 795.917710] env[61923]: ERROR nova.compute.manager [instance: ad3b3933-5bd9-444a-af7b-c81e868e0622] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 795.917710] env[61923]: ERROR nova.compute.manager [instance: ad3b3933-5bd9-444a-af7b-c81e868e0622] raise self.value [ 795.917710] env[61923]: ERROR nova.compute.manager [instance: ad3b3933-5bd9-444a-af7b-c81e868e0622] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 795.917710] env[61923]: ERROR nova.compute.manager [instance: ad3b3933-5bd9-444a-af7b-c81e868e0622] updated_port = self._update_port( [ 795.917710] env[61923]: ERROR nova.compute.manager [instance: ad3b3933-5bd9-444a-af7b-c81e868e0622] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 795.917710] env[61923]: ERROR nova.compute.manager [instance: ad3b3933-5bd9-444a-af7b-c81e868e0622] _ensure_no_port_binding_failure(port) [ 795.917710] env[61923]: ERROR nova.compute.manager [instance: ad3b3933-5bd9-444a-af7b-c81e868e0622] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 795.917710] env[61923]: ERROR nova.compute.manager [instance: ad3b3933-5bd9-444a-af7b-c81e868e0622] raise exception.PortBindingFailed(port_id=port['id']) [ 795.918043] env[61923]: ERROR nova.compute.manager [instance: ad3b3933-5bd9-444a-af7b-c81e868e0622] nova.exception.PortBindingFailed: Binding failed for port d7bef7b0-fae5-4d49-917d-36c20321c890, please check neutron logs for more information. [ 795.918043] env[61923]: ERROR nova.compute.manager [instance: ad3b3933-5bd9-444a-af7b-c81e868e0622] [ 795.918043] env[61923]: INFO nova.compute.manager [None req-a9ea97fa-71e5-4463-a1f3-ab69743d14a4 tempest-ServerActionsTestJSON-1295690556 tempest-ServerActionsTestJSON-1295690556-project-member] [instance: ad3b3933-5bd9-444a-af7b-c81e868e0622] Terminating instance [ 795.918867] env[61923]: DEBUG oslo_concurrency.lockutils [None req-a9ea97fa-71e5-4463-a1f3-ab69743d14a4 tempest-ServerActionsTestJSON-1295690556 tempest-ServerActionsTestJSON-1295690556-project-member] Acquiring lock "refresh_cache-ad3b3933-5bd9-444a-af7b-c81e868e0622" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 795.942596] env[61923]: DEBUG oslo_concurrency.lockutils [req-fc386e65-b4e7-434f-8fa4-79e15ec321f8 req-57ac4144-8360-40a8-9175-845695e96f86 service nova] Releasing lock "refresh_cache-fe981195-1e2c-4c8d-8c53-327851f845c7" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 795.942871] env[61923]: DEBUG nova.compute.manager [req-fc386e65-b4e7-434f-8fa4-79e15ec321f8 req-57ac4144-8360-40a8-9175-845695e96f86 service nova] [instance: fe981195-1e2c-4c8d-8c53-327851f845c7] Received event network-vif-deleted-5b85f4c9-ccc5-45a2-a383-7d3b903c5c61 {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 796.123396] env[61923]: DEBUG oslo_concurrency.lockutils [None req-0adbd41a-bb08-4b5a-bc10-ab9778329dd8 tempest-VolumesAdminNegativeTest-1422120048 tempest-VolumesAdminNegativeTest-1422120048-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.359s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 796.123927] env[61923]: DEBUG nova.compute.manager [None req-0adbd41a-bb08-4b5a-bc10-ab9778329dd8 tempest-VolumesAdminNegativeTest-1422120048 tempest-VolumesAdminNegativeTest-1422120048-project-member] [instance: 7bd3204b-5754-4ff2-b0ff-f30556dd72a1] Start building networks asynchronously for instance. {{(pid=61923) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 796.126474] env[61923]: DEBUG oslo_concurrency.lockutils [None req-28282894-8344-49f7-8696-a5bcafc57c69 tempest-ServerRescueTestJSONUnderV235-1784422495 tempest-ServerRescueTestJSONUnderV235-1784422495-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 26.553s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 796.127864] env[61923]: INFO nova.compute.claims [None req-28282894-8344-49f7-8696-a5bcafc57c69 tempest-ServerRescueTestJSONUnderV235-1784422495 tempest-ServerRescueTestJSONUnderV235-1784422495-project-member] [instance: b2d2b34d-827f-4f96-af8e-936135ba8700] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 796.245658] env[61923]: DEBUG nova.network.neutron [req-7d2dcfbf-18f1-49b8-a812-ff3c25739c2d req-419a079c-80b4-4e13-bdf9-57b554e72d53 service nova] [instance: ad3b3933-5bd9-444a-af7b-c81e868e0622] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 796.314387] env[61923]: DEBUG nova.network.neutron [req-7d2dcfbf-18f1-49b8-a812-ff3c25739c2d req-419a079c-80b4-4e13-bdf9-57b554e72d53 service nova] [instance: ad3b3933-5bd9-444a-af7b-c81e868e0622] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 796.632101] env[61923]: DEBUG nova.compute.utils [None req-0adbd41a-bb08-4b5a-bc10-ab9778329dd8 tempest-VolumesAdminNegativeTest-1422120048 tempest-VolumesAdminNegativeTest-1422120048-project-member] Using /dev/sd instead of None {{(pid=61923) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 796.635333] env[61923]: DEBUG nova.compute.manager [None req-0adbd41a-bb08-4b5a-bc10-ab9778329dd8 tempest-VolumesAdminNegativeTest-1422120048 tempest-VolumesAdminNegativeTest-1422120048-project-member] [instance: 7bd3204b-5754-4ff2-b0ff-f30556dd72a1] Allocating IP information in the background. {{(pid=61923) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 796.635497] env[61923]: DEBUG nova.network.neutron [None req-0adbd41a-bb08-4b5a-bc10-ab9778329dd8 tempest-VolumesAdminNegativeTest-1422120048 tempest-VolumesAdminNegativeTest-1422120048-project-member] [instance: 7bd3204b-5754-4ff2-b0ff-f30556dd72a1] allocate_for_instance() {{(pid=61923) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 796.671166] env[61923]: DEBUG nova.policy [None req-0adbd41a-bb08-4b5a-bc10-ab9778329dd8 tempest-VolumesAdminNegativeTest-1422120048 tempest-VolumesAdminNegativeTest-1422120048-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '98dcbc29b88741d5b51f311bb39f8a7a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '97570799e5f54fdeb7e8dfd2f50cc8a6', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61923) authorize /opt/stack/nova/nova/policy.py:201}} [ 796.817459] env[61923]: DEBUG oslo_concurrency.lockutils [req-7d2dcfbf-18f1-49b8-a812-ff3c25739c2d req-419a079c-80b4-4e13-bdf9-57b554e72d53 service nova] Releasing lock "refresh_cache-ad3b3933-5bd9-444a-af7b-c81e868e0622" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 796.818039] env[61923]: DEBUG oslo_concurrency.lockutils [None req-a9ea97fa-71e5-4463-a1f3-ab69743d14a4 tempest-ServerActionsTestJSON-1295690556 tempest-ServerActionsTestJSON-1295690556-project-member] Acquired lock "refresh_cache-ad3b3933-5bd9-444a-af7b-c81e868e0622" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 796.818352] env[61923]: DEBUG nova.network.neutron [None req-a9ea97fa-71e5-4463-a1f3-ab69743d14a4 tempest-ServerActionsTestJSON-1295690556 tempest-ServerActionsTestJSON-1295690556-project-member] [instance: ad3b3933-5bd9-444a-af7b-c81e868e0622] Building network info cache for instance {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 796.974907] env[61923]: DEBUG nova.network.neutron [None req-0adbd41a-bb08-4b5a-bc10-ab9778329dd8 tempest-VolumesAdminNegativeTest-1422120048 tempest-VolumesAdminNegativeTest-1422120048-project-member] [instance: 7bd3204b-5754-4ff2-b0ff-f30556dd72a1] Successfully created port: 9712b870-b8eb-4048-8ebf-0be8f9c5dfc2 {{(pid=61923) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 797.135968] env[61923]: DEBUG nova.compute.manager [None req-0adbd41a-bb08-4b5a-bc10-ab9778329dd8 tempest-VolumesAdminNegativeTest-1422120048 tempest-VolumesAdminNegativeTest-1422120048-project-member] [instance: 7bd3204b-5754-4ff2-b0ff-f30556dd72a1] Start building block device mappings for instance. {{(pid=61923) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 797.336425] env[61923]: DEBUG nova.network.neutron [None req-a9ea97fa-71e5-4463-a1f3-ab69743d14a4 tempest-ServerActionsTestJSON-1295690556 tempest-ServerActionsTestJSON-1295690556-project-member] [instance: ad3b3933-5bd9-444a-af7b-c81e868e0622] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 797.393702] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab23aabc-1611-4567-9c89-41ac5ae9de21 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.407700] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8c46b0c-3743-42d7-bf29-5357f8631fca {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.463525] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c8b0d8d-987a-411a-adf2-94e9fdd8fbeb {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.478749] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ad21bf3-08ec-44d7-b7fa-5666caa20314 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 797.486429] env[61923]: DEBUG nova.network.neutron [None req-a9ea97fa-71e5-4463-a1f3-ab69743d14a4 tempest-ServerActionsTestJSON-1295690556 tempest-ServerActionsTestJSON-1295690556-project-member] [instance: ad3b3933-5bd9-444a-af7b-c81e868e0622] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 797.501469] env[61923]: DEBUG nova.compute.provider_tree [None req-28282894-8344-49f7-8696-a5bcafc57c69 tempest-ServerRescueTestJSONUnderV235-1784422495 tempest-ServerRescueTestJSONUnderV235-1784422495-project-member] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 797.754317] env[61923]: DEBUG nova.compute.manager [req-a1fc531f-0ef9-4833-89d3-29149a045788 req-3793abb4-754e-4ab0-8012-0f8b44bec2c8 service nova] [instance: ad3b3933-5bd9-444a-af7b-c81e868e0622] Received event network-vif-deleted-d7bef7b0-fae5-4d49-917d-36c20321c890 {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 797.989845] env[61923]: DEBUG oslo_concurrency.lockutils [None req-a9ea97fa-71e5-4463-a1f3-ab69743d14a4 tempest-ServerActionsTestJSON-1295690556 tempest-ServerActionsTestJSON-1295690556-project-member] Releasing lock "refresh_cache-ad3b3933-5bd9-444a-af7b-c81e868e0622" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 797.990296] env[61923]: DEBUG nova.compute.manager [None req-a9ea97fa-71e5-4463-a1f3-ab69743d14a4 tempest-ServerActionsTestJSON-1295690556 tempest-ServerActionsTestJSON-1295690556-project-member] [instance: ad3b3933-5bd9-444a-af7b-c81e868e0622] Start destroying the instance on the hypervisor. {{(pid=61923) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 797.990497] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-a9ea97fa-71e5-4463-a1f3-ab69743d14a4 tempest-ServerActionsTestJSON-1295690556 tempest-ServerActionsTestJSON-1295690556-project-member] [instance: ad3b3933-5bd9-444a-af7b-c81e868e0622] Destroying instance {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 797.990804] env[61923]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-4c16b91c-f30d-4f4a-a16b-0c84c7a9eb10 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.000199] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84b135b0-8cb6-4b0d-bfdb-db7272229fea {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.011014] env[61923]: DEBUG nova.scheduler.client.report [None req-28282894-8344-49f7-8696-a5bcafc57c69 tempest-ServerRescueTestJSONUnderV235-1784422495 tempest-ServerRescueTestJSONUnderV235-1784422495-project-member] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 798.027114] env[61923]: WARNING nova.virt.vmwareapi.vmops [None req-a9ea97fa-71e5-4463-a1f3-ab69743d14a4 tempest-ServerActionsTestJSON-1295690556 tempest-ServerActionsTestJSON-1295690556-project-member] [instance: ad3b3933-5bd9-444a-af7b-c81e868e0622] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance ad3b3933-5bd9-444a-af7b-c81e868e0622 could not be found. [ 798.027489] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-a9ea97fa-71e5-4463-a1f3-ab69743d14a4 tempest-ServerActionsTestJSON-1295690556 tempest-ServerActionsTestJSON-1295690556-project-member] [instance: ad3b3933-5bd9-444a-af7b-c81e868e0622] Instance destroyed {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 798.027658] env[61923]: INFO nova.compute.manager [None req-a9ea97fa-71e5-4463-a1f3-ab69743d14a4 tempest-ServerActionsTestJSON-1295690556 tempest-ServerActionsTestJSON-1295690556-project-member] [instance: ad3b3933-5bd9-444a-af7b-c81e868e0622] Took 0.04 seconds to destroy the instance on the hypervisor. [ 798.027961] env[61923]: DEBUG oslo.service.loopingcall [None req-a9ea97fa-71e5-4463-a1f3-ab69743d14a4 tempest-ServerActionsTestJSON-1295690556 tempest-ServerActionsTestJSON-1295690556-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61923) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 798.028195] env[61923]: DEBUG nova.compute.manager [-] [instance: ad3b3933-5bd9-444a-af7b-c81e868e0622] Deallocating network for instance {{(pid=61923) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 798.028288] env[61923]: DEBUG nova.network.neutron [-] [instance: ad3b3933-5bd9-444a-af7b-c81e868e0622] deallocate_for_instance() {{(pid=61923) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 798.048443] env[61923]: DEBUG nova.network.neutron [-] [instance: ad3b3933-5bd9-444a-af7b-c81e868e0622] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 798.100898] env[61923]: ERROR nova.compute.manager [None req-0adbd41a-bb08-4b5a-bc10-ab9778329dd8 tempest-VolumesAdminNegativeTest-1422120048 tempest-VolumesAdminNegativeTest-1422120048-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 9712b870-b8eb-4048-8ebf-0be8f9c5dfc2, please check neutron logs for more information. [ 798.100898] env[61923]: ERROR nova.compute.manager Traceback (most recent call last): [ 798.100898] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 798.100898] env[61923]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 798.100898] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 798.100898] env[61923]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 798.100898] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 798.100898] env[61923]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 798.100898] env[61923]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 798.100898] env[61923]: ERROR nova.compute.manager self.force_reraise() [ 798.100898] env[61923]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 798.100898] env[61923]: ERROR nova.compute.manager raise self.value [ 798.100898] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 798.100898] env[61923]: ERROR nova.compute.manager updated_port = self._update_port( [ 798.100898] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 798.100898] env[61923]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 798.101366] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 798.101366] env[61923]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 798.101366] env[61923]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 9712b870-b8eb-4048-8ebf-0be8f9c5dfc2, please check neutron logs for more information. [ 798.101366] env[61923]: ERROR nova.compute.manager [ 798.101366] env[61923]: Traceback (most recent call last): [ 798.101366] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 798.101366] env[61923]: listener.cb(fileno) [ 798.101366] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 798.101366] env[61923]: result = function(*args, **kwargs) [ 798.101366] env[61923]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 798.101366] env[61923]: return func(*args, **kwargs) [ 798.101366] env[61923]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 798.101366] env[61923]: raise e [ 798.101366] env[61923]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 798.101366] env[61923]: nwinfo = self.network_api.allocate_for_instance( [ 798.101366] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 798.101366] env[61923]: created_port_ids = self._update_ports_for_instance( [ 798.101366] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 798.101366] env[61923]: with excutils.save_and_reraise_exception(): [ 798.101366] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 798.101366] env[61923]: self.force_reraise() [ 798.101366] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 798.101366] env[61923]: raise self.value [ 798.101366] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 798.101366] env[61923]: updated_port = self._update_port( [ 798.101366] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 798.101366] env[61923]: _ensure_no_port_binding_failure(port) [ 798.101366] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 798.101366] env[61923]: raise exception.PortBindingFailed(port_id=port['id']) [ 798.102129] env[61923]: nova.exception.PortBindingFailed: Binding failed for port 9712b870-b8eb-4048-8ebf-0be8f9c5dfc2, please check neutron logs for more information. [ 798.102129] env[61923]: Removing descriptor: 14 [ 798.150225] env[61923]: DEBUG nova.compute.manager [None req-0adbd41a-bb08-4b5a-bc10-ab9778329dd8 tempest-VolumesAdminNegativeTest-1422120048 tempest-VolumesAdminNegativeTest-1422120048-project-member] [instance: 7bd3204b-5754-4ff2-b0ff-f30556dd72a1] Start spawning the instance on the hypervisor. {{(pid=61923) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 798.177251] env[61923]: DEBUG nova.virt.hardware [None req-0adbd41a-bb08-4b5a-bc10-ab9778329dd8 tempest-VolumesAdminNegativeTest-1422120048 tempest-VolumesAdminNegativeTest-1422120048-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-29T20:07:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-29T20:07:35Z,direct_url=,disk_format='vmdk',id=0f153f63-ae0a-45b1-b7f5-7f9b673c947f,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='4e7b5e3b3c3443c8bd5c70f8a15739f5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-29T20:07:36Z,virtual_size=,visibility=), allow threads: False {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 798.177510] env[61923]: DEBUG nova.virt.hardware [None req-0adbd41a-bb08-4b5a-bc10-ab9778329dd8 tempest-VolumesAdminNegativeTest-1422120048 tempest-VolumesAdminNegativeTest-1422120048-project-member] Flavor limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 798.177666] env[61923]: DEBUG nova.virt.hardware [None req-0adbd41a-bb08-4b5a-bc10-ab9778329dd8 tempest-VolumesAdminNegativeTest-1422120048 tempest-VolumesAdminNegativeTest-1422120048-project-member] Image limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 798.177846] env[61923]: DEBUG nova.virt.hardware [None req-0adbd41a-bb08-4b5a-bc10-ab9778329dd8 tempest-VolumesAdminNegativeTest-1422120048 tempest-VolumesAdminNegativeTest-1422120048-project-member] Flavor pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 798.177992] env[61923]: DEBUG nova.virt.hardware [None req-0adbd41a-bb08-4b5a-bc10-ab9778329dd8 tempest-VolumesAdminNegativeTest-1422120048 tempest-VolumesAdminNegativeTest-1422120048-project-member] Image pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 798.178170] env[61923]: DEBUG nova.virt.hardware [None req-0adbd41a-bb08-4b5a-bc10-ab9778329dd8 tempest-VolumesAdminNegativeTest-1422120048 tempest-VolumesAdminNegativeTest-1422120048-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 798.178374] env[61923]: DEBUG nova.virt.hardware [None req-0adbd41a-bb08-4b5a-bc10-ab9778329dd8 tempest-VolumesAdminNegativeTest-1422120048 tempest-VolumesAdminNegativeTest-1422120048-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 798.178553] env[61923]: DEBUG nova.virt.hardware [None req-0adbd41a-bb08-4b5a-bc10-ab9778329dd8 tempest-VolumesAdminNegativeTest-1422120048 tempest-VolumesAdminNegativeTest-1422120048-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 798.178730] env[61923]: DEBUG nova.virt.hardware [None req-0adbd41a-bb08-4b5a-bc10-ab9778329dd8 tempest-VolumesAdminNegativeTest-1422120048 tempest-VolumesAdminNegativeTest-1422120048-project-member] Got 1 possible topologies {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 798.178889] env[61923]: DEBUG nova.virt.hardware [None req-0adbd41a-bb08-4b5a-bc10-ab9778329dd8 tempest-VolumesAdminNegativeTest-1422120048 tempest-VolumesAdminNegativeTest-1422120048-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 798.179073] env[61923]: DEBUG nova.virt.hardware [None req-0adbd41a-bb08-4b5a-bc10-ab9778329dd8 tempest-VolumesAdminNegativeTest-1422120048 tempest-VolumesAdminNegativeTest-1422120048-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 798.180035] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a85610c1-de19-4207-9176-f3d59734ac5d {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.188408] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-397433bb-983e-4711-b348-b3be23f728fa {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 798.204206] env[61923]: ERROR nova.compute.manager [None req-0adbd41a-bb08-4b5a-bc10-ab9778329dd8 tempest-VolumesAdminNegativeTest-1422120048 tempest-VolumesAdminNegativeTest-1422120048-project-member] [instance: 7bd3204b-5754-4ff2-b0ff-f30556dd72a1] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 9712b870-b8eb-4048-8ebf-0be8f9c5dfc2, please check neutron logs for more information. [ 798.204206] env[61923]: ERROR nova.compute.manager [instance: 7bd3204b-5754-4ff2-b0ff-f30556dd72a1] Traceback (most recent call last): [ 798.204206] env[61923]: ERROR nova.compute.manager [instance: 7bd3204b-5754-4ff2-b0ff-f30556dd72a1] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 798.204206] env[61923]: ERROR nova.compute.manager [instance: 7bd3204b-5754-4ff2-b0ff-f30556dd72a1] yield resources [ 798.204206] env[61923]: ERROR nova.compute.manager [instance: 7bd3204b-5754-4ff2-b0ff-f30556dd72a1] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 798.204206] env[61923]: ERROR nova.compute.manager [instance: 7bd3204b-5754-4ff2-b0ff-f30556dd72a1] self.driver.spawn(context, instance, image_meta, [ 798.204206] env[61923]: ERROR nova.compute.manager [instance: 7bd3204b-5754-4ff2-b0ff-f30556dd72a1] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 798.204206] env[61923]: ERROR nova.compute.manager [instance: 7bd3204b-5754-4ff2-b0ff-f30556dd72a1] self._vmops.spawn(context, instance, image_meta, injected_files, [ 798.204206] env[61923]: ERROR nova.compute.manager [instance: 7bd3204b-5754-4ff2-b0ff-f30556dd72a1] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 798.204206] env[61923]: ERROR nova.compute.manager [instance: 7bd3204b-5754-4ff2-b0ff-f30556dd72a1] vm_ref = self.build_virtual_machine(instance, [ 798.204206] env[61923]: ERROR nova.compute.manager [instance: 7bd3204b-5754-4ff2-b0ff-f30556dd72a1] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 798.204626] env[61923]: ERROR nova.compute.manager [instance: 7bd3204b-5754-4ff2-b0ff-f30556dd72a1] vif_infos = vmwarevif.get_vif_info(self._session, [ 798.204626] env[61923]: ERROR nova.compute.manager [instance: 7bd3204b-5754-4ff2-b0ff-f30556dd72a1] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 798.204626] env[61923]: ERROR nova.compute.manager [instance: 7bd3204b-5754-4ff2-b0ff-f30556dd72a1] for vif in network_info: [ 798.204626] env[61923]: ERROR nova.compute.manager [instance: 7bd3204b-5754-4ff2-b0ff-f30556dd72a1] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 798.204626] env[61923]: ERROR nova.compute.manager [instance: 7bd3204b-5754-4ff2-b0ff-f30556dd72a1] return self._sync_wrapper(fn, *args, **kwargs) [ 798.204626] env[61923]: ERROR nova.compute.manager [instance: 7bd3204b-5754-4ff2-b0ff-f30556dd72a1] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 798.204626] env[61923]: ERROR nova.compute.manager [instance: 7bd3204b-5754-4ff2-b0ff-f30556dd72a1] self.wait() [ 798.204626] env[61923]: ERROR nova.compute.manager [instance: 7bd3204b-5754-4ff2-b0ff-f30556dd72a1] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 798.204626] env[61923]: ERROR nova.compute.manager [instance: 7bd3204b-5754-4ff2-b0ff-f30556dd72a1] self[:] = self._gt.wait() [ 798.204626] env[61923]: ERROR nova.compute.manager [instance: 7bd3204b-5754-4ff2-b0ff-f30556dd72a1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 798.204626] env[61923]: ERROR nova.compute.manager [instance: 7bd3204b-5754-4ff2-b0ff-f30556dd72a1] return self._exit_event.wait() [ 798.204626] env[61923]: ERROR nova.compute.manager [instance: 7bd3204b-5754-4ff2-b0ff-f30556dd72a1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 798.204626] env[61923]: ERROR nova.compute.manager [instance: 7bd3204b-5754-4ff2-b0ff-f30556dd72a1] current.throw(*self._exc) [ 798.204915] env[61923]: ERROR nova.compute.manager [instance: 7bd3204b-5754-4ff2-b0ff-f30556dd72a1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 798.204915] env[61923]: ERROR nova.compute.manager [instance: 7bd3204b-5754-4ff2-b0ff-f30556dd72a1] result = function(*args, **kwargs) [ 798.204915] env[61923]: ERROR nova.compute.manager [instance: 7bd3204b-5754-4ff2-b0ff-f30556dd72a1] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 798.204915] env[61923]: ERROR nova.compute.manager [instance: 7bd3204b-5754-4ff2-b0ff-f30556dd72a1] return func(*args, **kwargs) [ 798.204915] env[61923]: ERROR nova.compute.manager [instance: 7bd3204b-5754-4ff2-b0ff-f30556dd72a1] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 798.204915] env[61923]: ERROR nova.compute.manager [instance: 7bd3204b-5754-4ff2-b0ff-f30556dd72a1] raise e [ 798.204915] env[61923]: ERROR nova.compute.manager [instance: 7bd3204b-5754-4ff2-b0ff-f30556dd72a1] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 798.204915] env[61923]: ERROR nova.compute.manager [instance: 7bd3204b-5754-4ff2-b0ff-f30556dd72a1] nwinfo = self.network_api.allocate_for_instance( [ 798.204915] env[61923]: ERROR nova.compute.manager [instance: 7bd3204b-5754-4ff2-b0ff-f30556dd72a1] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 798.204915] env[61923]: ERROR nova.compute.manager [instance: 7bd3204b-5754-4ff2-b0ff-f30556dd72a1] created_port_ids = self._update_ports_for_instance( [ 798.204915] env[61923]: ERROR nova.compute.manager [instance: 7bd3204b-5754-4ff2-b0ff-f30556dd72a1] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 798.204915] env[61923]: ERROR nova.compute.manager [instance: 7bd3204b-5754-4ff2-b0ff-f30556dd72a1] with excutils.save_and_reraise_exception(): [ 798.204915] env[61923]: ERROR nova.compute.manager [instance: 7bd3204b-5754-4ff2-b0ff-f30556dd72a1] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 798.205266] env[61923]: ERROR nova.compute.manager [instance: 7bd3204b-5754-4ff2-b0ff-f30556dd72a1] self.force_reraise() [ 798.205266] env[61923]: ERROR nova.compute.manager [instance: 7bd3204b-5754-4ff2-b0ff-f30556dd72a1] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 798.205266] env[61923]: ERROR nova.compute.manager [instance: 7bd3204b-5754-4ff2-b0ff-f30556dd72a1] raise self.value [ 798.205266] env[61923]: ERROR nova.compute.manager [instance: 7bd3204b-5754-4ff2-b0ff-f30556dd72a1] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 798.205266] env[61923]: ERROR nova.compute.manager [instance: 7bd3204b-5754-4ff2-b0ff-f30556dd72a1] updated_port = self._update_port( [ 798.205266] env[61923]: ERROR nova.compute.manager [instance: 7bd3204b-5754-4ff2-b0ff-f30556dd72a1] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 798.205266] env[61923]: ERROR nova.compute.manager [instance: 7bd3204b-5754-4ff2-b0ff-f30556dd72a1] _ensure_no_port_binding_failure(port) [ 798.205266] env[61923]: ERROR nova.compute.manager [instance: 7bd3204b-5754-4ff2-b0ff-f30556dd72a1] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 798.205266] env[61923]: ERROR nova.compute.manager [instance: 7bd3204b-5754-4ff2-b0ff-f30556dd72a1] raise exception.PortBindingFailed(port_id=port['id']) [ 798.205266] env[61923]: ERROR nova.compute.manager [instance: 7bd3204b-5754-4ff2-b0ff-f30556dd72a1] nova.exception.PortBindingFailed: Binding failed for port 9712b870-b8eb-4048-8ebf-0be8f9c5dfc2, please check neutron logs for more information. [ 798.205266] env[61923]: ERROR nova.compute.manager [instance: 7bd3204b-5754-4ff2-b0ff-f30556dd72a1] [ 798.205266] env[61923]: INFO nova.compute.manager [None req-0adbd41a-bb08-4b5a-bc10-ab9778329dd8 tempest-VolumesAdminNegativeTest-1422120048 tempest-VolumesAdminNegativeTest-1422120048-project-member] [instance: 7bd3204b-5754-4ff2-b0ff-f30556dd72a1] Terminating instance [ 798.206553] env[61923]: DEBUG oslo_concurrency.lockutils [None req-0adbd41a-bb08-4b5a-bc10-ab9778329dd8 tempest-VolumesAdminNegativeTest-1422120048 tempest-VolumesAdminNegativeTest-1422120048-project-member] Acquiring lock "refresh_cache-7bd3204b-5754-4ff2-b0ff-f30556dd72a1" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 798.206724] env[61923]: DEBUG oslo_concurrency.lockutils [None req-0adbd41a-bb08-4b5a-bc10-ab9778329dd8 tempest-VolumesAdminNegativeTest-1422120048 tempest-VolumesAdminNegativeTest-1422120048-project-member] Acquired lock "refresh_cache-7bd3204b-5754-4ff2-b0ff-f30556dd72a1" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 798.206891] env[61923]: DEBUG nova.network.neutron [None req-0adbd41a-bb08-4b5a-bc10-ab9778329dd8 tempest-VolumesAdminNegativeTest-1422120048 tempest-VolumesAdminNegativeTest-1422120048-project-member] [instance: 7bd3204b-5754-4ff2-b0ff-f30556dd72a1] Building network info cache for instance {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 798.517347] env[61923]: DEBUG oslo_concurrency.lockutils [None req-28282894-8344-49f7-8696-a5bcafc57c69 tempest-ServerRescueTestJSONUnderV235-1784422495 tempest-ServerRescueTestJSONUnderV235-1784422495-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.391s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 798.517944] env[61923]: DEBUG nova.compute.manager [None req-28282894-8344-49f7-8696-a5bcafc57c69 tempest-ServerRescueTestJSONUnderV235-1784422495 tempest-ServerRescueTestJSONUnderV235-1784422495-project-member] [instance: b2d2b34d-827f-4f96-af8e-936135ba8700] Start building networks asynchronously for instance. {{(pid=61923) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 798.521036] env[61923]: DEBUG oslo_concurrency.lockutils [None req-998fcb75-b821-4586-ae48-2fe4d02a9864 tempest-AttachVolumeShelveTestJSON-1252849694 tempest-AttachVolumeShelveTestJSON-1252849694-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 28.916s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 798.522463] env[61923]: INFO nova.compute.claims [None req-998fcb75-b821-4586-ae48-2fe4d02a9864 tempest-AttachVolumeShelveTestJSON-1252849694 tempest-AttachVolumeShelveTestJSON-1252849694-project-member] [instance: 870602dd-1b85-4c97-9506-f08e9f1bade6] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 798.551969] env[61923]: DEBUG nova.network.neutron [-] [instance: ad3b3933-5bd9-444a-af7b-c81e868e0622] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 798.733746] env[61923]: DEBUG nova.network.neutron [None req-0adbd41a-bb08-4b5a-bc10-ab9778329dd8 tempest-VolumesAdminNegativeTest-1422120048 tempest-VolumesAdminNegativeTest-1422120048-project-member] [instance: 7bd3204b-5754-4ff2-b0ff-f30556dd72a1] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 798.810352] env[61923]: DEBUG nova.network.neutron [None req-0adbd41a-bb08-4b5a-bc10-ab9778329dd8 tempest-VolumesAdminNegativeTest-1422120048 tempest-VolumesAdminNegativeTest-1422120048-project-member] [instance: 7bd3204b-5754-4ff2-b0ff-f30556dd72a1] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 799.027023] env[61923]: DEBUG nova.compute.utils [None req-28282894-8344-49f7-8696-a5bcafc57c69 tempest-ServerRescueTestJSONUnderV235-1784422495 tempest-ServerRescueTestJSONUnderV235-1784422495-project-member] Using /dev/sd instead of None {{(pid=61923) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 799.032019] env[61923]: DEBUG nova.compute.manager [None req-28282894-8344-49f7-8696-a5bcafc57c69 tempest-ServerRescueTestJSONUnderV235-1784422495 tempest-ServerRescueTestJSONUnderV235-1784422495-project-member] [instance: b2d2b34d-827f-4f96-af8e-936135ba8700] Allocating IP information in the background. {{(pid=61923) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 799.032019] env[61923]: DEBUG nova.network.neutron [None req-28282894-8344-49f7-8696-a5bcafc57c69 tempest-ServerRescueTestJSONUnderV235-1784422495 tempest-ServerRescueTestJSONUnderV235-1784422495-project-member] [instance: b2d2b34d-827f-4f96-af8e-936135ba8700] allocate_for_instance() {{(pid=61923) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 799.054699] env[61923]: INFO nova.compute.manager [-] [instance: ad3b3933-5bd9-444a-af7b-c81e868e0622] Took 1.03 seconds to deallocate network for instance. [ 799.056974] env[61923]: DEBUG nova.compute.claims [None req-a9ea97fa-71e5-4463-a1f3-ab69743d14a4 tempest-ServerActionsTestJSON-1295690556 tempest-ServerActionsTestJSON-1295690556-project-member] [instance: ad3b3933-5bd9-444a-af7b-c81e868e0622] Aborting claim: {{(pid=61923) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 799.057167] env[61923]: DEBUG oslo_concurrency.lockutils [None req-a9ea97fa-71e5-4463-a1f3-ab69743d14a4 tempest-ServerActionsTestJSON-1295690556 tempest-ServerActionsTestJSON-1295690556-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 799.070291] env[61923]: DEBUG nova.policy [None req-28282894-8344-49f7-8696-a5bcafc57c69 tempest-ServerRescueTestJSONUnderV235-1784422495 tempest-ServerRescueTestJSONUnderV235-1784422495-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'fc605decd45a4b448e2e4aed1f2f76e7', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7c3aebbbdf3e40ce9d1379a9f5a6378f', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61923) authorize /opt/stack/nova/nova/policy.py:201}} [ 799.312768] env[61923]: DEBUG oslo_concurrency.lockutils [None req-0adbd41a-bb08-4b5a-bc10-ab9778329dd8 tempest-VolumesAdminNegativeTest-1422120048 tempest-VolumesAdminNegativeTest-1422120048-project-member] Releasing lock "refresh_cache-7bd3204b-5754-4ff2-b0ff-f30556dd72a1" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 799.313253] env[61923]: DEBUG nova.compute.manager [None req-0adbd41a-bb08-4b5a-bc10-ab9778329dd8 tempest-VolumesAdminNegativeTest-1422120048 tempest-VolumesAdminNegativeTest-1422120048-project-member] [instance: 7bd3204b-5754-4ff2-b0ff-f30556dd72a1] Start destroying the instance on the hypervisor. {{(pid=61923) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 799.313468] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-0adbd41a-bb08-4b5a-bc10-ab9778329dd8 tempest-VolumesAdminNegativeTest-1422120048 tempest-VolumesAdminNegativeTest-1422120048-project-member] [instance: 7bd3204b-5754-4ff2-b0ff-f30556dd72a1] Destroying instance {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 799.313764] env[61923]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-2a926d37-1d89-4815-9d56-736b6e775437 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.322441] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2076579-03bc-4556-8d2b-23e94d15912a {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.344165] env[61923]: WARNING nova.virt.vmwareapi.vmops [None req-0adbd41a-bb08-4b5a-bc10-ab9778329dd8 tempest-VolumesAdminNegativeTest-1422120048 tempest-VolumesAdminNegativeTest-1422120048-project-member] [instance: 7bd3204b-5754-4ff2-b0ff-f30556dd72a1] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 7bd3204b-5754-4ff2-b0ff-f30556dd72a1 could not be found. [ 799.344386] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-0adbd41a-bb08-4b5a-bc10-ab9778329dd8 tempest-VolumesAdminNegativeTest-1422120048 tempest-VolumesAdminNegativeTest-1422120048-project-member] [instance: 7bd3204b-5754-4ff2-b0ff-f30556dd72a1] Instance destroyed {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 799.344565] env[61923]: INFO nova.compute.manager [None req-0adbd41a-bb08-4b5a-bc10-ab9778329dd8 tempest-VolumesAdminNegativeTest-1422120048 tempest-VolumesAdminNegativeTest-1422120048-project-member] [instance: 7bd3204b-5754-4ff2-b0ff-f30556dd72a1] Took 0.03 seconds to destroy the instance on the hypervisor. [ 799.344812] env[61923]: DEBUG oslo.service.loopingcall [None req-0adbd41a-bb08-4b5a-bc10-ab9778329dd8 tempest-VolumesAdminNegativeTest-1422120048 tempest-VolumesAdminNegativeTest-1422120048-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61923) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 799.345042] env[61923]: DEBUG nova.compute.manager [-] [instance: 7bd3204b-5754-4ff2-b0ff-f30556dd72a1] Deallocating network for instance {{(pid=61923) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 799.345138] env[61923]: DEBUG nova.network.neutron [-] [instance: 7bd3204b-5754-4ff2-b0ff-f30556dd72a1] deallocate_for_instance() {{(pid=61923) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 799.360237] env[61923]: DEBUG nova.network.neutron [-] [instance: 7bd3204b-5754-4ff2-b0ff-f30556dd72a1] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 799.422728] env[61923]: DEBUG nova.network.neutron [None req-28282894-8344-49f7-8696-a5bcafc57c69 tempest-ServerRescueTestJSONUnderV235-1784422495 tempest-ServerRescueTestJSONUnderV235-1784422495-project-member] [instance: b2d2b34d-827f-4f96-af8e-936135ba8700] Successfully created port: 85462a2f-eba7-481a-8f13-c572e92d6634 {{(pid=61923) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 799.536543] env[61923]: DEBUG nova.compute.manager [None req-28282894-8344-49f7-8696-a5bcafc57c69 tempest-ServerRescueTestJSONUnderV235-1784422495 tempest-ServerRescueTestJSONUnderV235-1784422495-project-member] [instance: b2d2b34d-827f-4f96-af8e-936135ba8700] Start building block device mappings for instance. {{(pid=61923) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 799.779388] env[61923]: DEBUG nova.compute.manager [req-e64f4413-6a17-4fab-acdb-907b880fb93a req-3baa98ae-2a4e-4f9c-ad06-2055e953ca1c service nova] [instance: 7bd3204b-5754-4ff2-b0ff-f30556dd72a1] Received event network-changed-9712b870-b8eb-4048-8ebf-0be8f9c5dfc2 {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 799.779598] env[61923]: DEBUG nova.compute.manager [req-e64f4413-6a17-4fab-acdb-907b880fb93a req-3baa98ae-2a4e-4f9c-ad06-2055e953ca1c service nova] [instance: 7bd3204b-5754-4ff2-b0ff-f30556dd72a1] Refreshing instance network info cache due to event network-changed-9712b870-b8eb-4048-8ebf-0be8f9c5dfc2. {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 799.779876] env[61923]: DEBUG oslo_concurrency.lockutils [req-e64f4413-6a17-4fab-acdb-907b880fb93a req-3baa98ae-2a4e-4f9c-ad06-2055e953ca1c service nova] Acquiring lock "refresh_cache-7bd3204b-5754-4ff2-b0ff-f30556dd72a1" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 799.779971] env[61923]: DEBUG oslo_concurrency.lockutils [req-e64f4413-6a17-4fab-acdb-907b880fb93a req-3baa98ae-2a4e-4f9c-ad06-2055e953ca1c service nova] Acquired lock "refresh_cache-7bd3204b-5754-4ff2-b0ff-f30556dd72a1" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 799.780189] env[61923]: DEBUG nova.network.neutron [req-e64f4413-6a17-4fab-acdb-907b880fb93a req-3baa98ae-2a4e-4f9c-ad06-2055e953ca1c service nova] [instance: 7bd3204b-5754-4ff2-b0ff-f30556dd72a1] Refreshing network info cache for port 9712b870-b8eb-4048-8ebf-0be8f9c5dfc2 {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 799.795439] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7fc1591-9b26-4bbb-99e3-7e0ec769e00f {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.803477] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a51b1375-267a-4797-a05e-105ddfad2175 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.834608] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f33d75ac-e62d-4c38-a957-c857d9a20b8b {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.841862] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74dc21f5-40b7-4537-9d26-667d7c1989f8 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.854736] env[61923]: DEBUG nova.compute.provider_tree [None req-998fcb75-b821-4586-ae48-2fe4d02a9864 tempest-AttachVolumeShelveTestJSON-1252849694 tempest-AttachVolumeShelveTestJSON-1252849694-project-member] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 799.865774] env[61923]: DEBUG nova.network.neutron [-] [instance: 7bd3204b-5754-4ff2-b0ff-f30556dd72a1] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 800.321921] env[61923]: DEBUG nova.network.neutron [req-e64f4413-6a17-4fab-acdb-907b880fb93a req-3baa98ae-2a4e-4f9c-ad06-2055e953ca1c service nova] [instance: 7bd3204b-5754-4ff2-b0ff-f30556dd72a1] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 800.358168] env[61923]: DEBUG nova.scheduler.client.report [None req-998fcb75-b821-4586-ae48-2fe4d02a9864 tempest-AttachVolumeShelveTestJSON-1252849694 tempest-AttachVolumeShelveTestJSON-1252849694-project-member] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 800.374625] env[61923]: INFO nova.compute.manager [-] [instance: 7bd3204b-5754-4ff2-b0ff-f30556dd72a1] Took 1.02 seconds to deallocate network for instance. [ 800.374625] env[61923]: DEBUG nova.compute.claims [None req-0adbd41a-bb08-4b5a-bc10-ab9778329dd8 tempest-VolumesAdminNegativeTest-1422120048 tempest-VolumesAdminNegativeTest-1422120048-project-member] [instance: 7bd3204b-5754-4ff2-b0ff-f30556dd72a1] Aborting claim: {{(pid=61923) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 800.374625] env[61923]: DEBUG oslo_concurrency.lockutils [None req-0adbd41a-bb08-4b5a-bc10-ab9778329dd8 tempest-VolumesAdminNegativeTest-1422120048 tempest-VolumesAdminNegativeTest-1422120048-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 800.465019] env[61923]: ERROR nova.compute.manager [None req-28282894-8344-49f7-8696-a5bcafc57c69 tempest-ServerRescueTestJSONUnderV235-1784422495 tempest-ServerRescueTestJSONUnderV235-1784422495-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 85462a2f-eba7-481a-8f13-c572e92d6634, please check neutron logs for more information. [ 800.465019] env[61923]: ERROR nova.compute.manager Traceback (most recent call last): [ 800.465019] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 800.465019] env[61923]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 800.465019] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 800.465019] env[61923]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 800.465019] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 800.465019] env[61923]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 800.465019] env[61923]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 800.465019] env[61923]: ERROR nova.compute.manager self.force_reraise() [ 800.465019] env[61923]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 800.465019] env[61923]: ERROR nova.compute.manager raise self.value [ 800.465019] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 800.465019] env[61923]: ERROR nova.compute.manager updated_port = self._update_port( [ 800.465019] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 800.465019] env[61923]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 800.465396] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 800.465396] env[61923]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 800.465396] env[61923]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 85462a2f-eba7-481a-8f13-c572e92d6634, please check neutron logs for more information. [ 800.465396] env[61923]: ERROR nova.compute.manager [ 800.465396] env[61923]: Traceback (most recent call last): [ 800.465396] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 800.465396] env[61923]: listener.cb(fileno) [ 800.465396] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 800.465396] env[61923]: result = function(*args, **kwargs) [ 800.465396] env[61923]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 800.465396] env[61923]: return func(*args, **kwargs) [ 800.465396] env[61923]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 800.465396] env[61923]: raise e [ 800.465396] env[61923]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 800.465396] env[61923]: nwinfo = self.network_api.allocate_for_instance( [ 800.465396] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 800.465396] env[61923]: created_port_ids = self._update_ports_for_instance( [ 800.465396] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 800.465396] env[61923]: with excutils.save_and_reraise_exception(): [ 800.465396] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 800.465396] env[61923]: self.force_reraise() [ 800.465396] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 800.465396] env[61923]: raise self.value [ 800.465396] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 800.465396] env[61923]: updated_port = self._update_port( [ 800.465396] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 800.465396] env[61923]: _ensure_no_port_binding_failure(port) [ 800.465396] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 800.465396] env[61923]: raise exception.PortBindingFailed(port_id=port['id']) [ 800.466021] env[61923]: nova.exception.PortBindingFailed: Binding failed for port 85462a2f-eba7-481a-8f13-c572e92d6634, please check neutron logs for more information. [ 800.466021] env[61923]: Removing descriptor: 14 [ 800.474698] env[61923]: DEBUG nova.network.neutron [req-e64f4413-6a17-4fab-acdb-907b880fb93a req-3baa98ae-2a4e-4f9c-ad06-2055e953ca1c service nova] [instance: 7bd3204b-5754-4ff2-b0ff-f30556dd72a1] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 800.546638] env[61923]: DEBUG nova.compute.manager [None req-28282894-8344-49f7-8696-a5bcafc57c69 tempest-ServerRescueTestJSONUnderV235-1784422495 tempest-ServerRescueTestJSONUnderV235-1784422495-project-member] [instance: b2d2b34d-827f-4f96-af8e-936135ba8700] Start spawning the instance on the hypervisor. {{(pid=61923) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 800.570137] env[61923]: DEBUG nova.virt.hardware [None req-28282894-8344-49f7-8696-a5bcafc57c69 tempest-ServerRescueTestJSONUnderV235-1784422495 tempest-ServerRescueTestJSONUnderV235-1784422495-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-29T20:07:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-29T20:07:35Z,direct_url=,disk_format='vmdk',id=0f153f63-ae0a-45b1-b7f5-7f9b673c947f,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='4e7b5e3b3c3443c8bd5c70f8a15739f5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-29T20:07:36Z,virtual_size=,visibility=), allow threads: False {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 800.570383] env[61923]: DEBUG nova.virt.hardware [None req-28282894-8344-49f7-8696-a5bcafc57c69 tempest-ServerRescueTestJSONUnderV235-1784422495 tempest-ServerRescueTestJSONUnderV235-1784422495-project-member] Flavor limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 800.570536] env[61923]: DEBUG nova.virt.hardware [None req-28282894-8344-49f7-8696-a5bcafc57c69 tempest-ServerRescueTestJSONUnderV235-1784422495 tempest-ServerRescueTestJSONUnderV235-1784422495-project-member] Image limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 800.570713] env[61923]: DEBUG nova.virt.hardware [None req-28282894-8344-49f7-8696-a5bcafc57c69 tempest-ServerRescueTestJSONUnderV235-1784422495 tempest-ServerRescueTestJSONUnderV235-1784422495-project-member] Flavor pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 800.570857] env[61923]: DEBUG nova.virt.hardware [None req-28282894-8344-49f7-8696-a5bcafc57c69 tempest-ServerRescueTestJSONUnderV235-1784422495 tempest-ServerRescueTestJSONUnderV235-1784422495-project-member] Image pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 800.571009] env[61923]: DEBUG nova.virt.hardware [None req-28282894-8344-49f7-8696-a5bcafc57c69 tempest-ServerRescueTestJSONUnderV235-1784422495 tempest-ServerRescueTestJSONUnderV235-1784422495-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 800.571222] env[61923]: DEBUG nova.virt.hardware [None req-28282894-8344-49f7-8696-a5bcafc57c69 tempest-ServerRescueTestJSONUnderV235-1784422495 tempest-ServerRescueTestJSONUnderV235-1784422495-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 800.571378] env[61923]: DEBUG nova.virt.hardware [None req-28282894-8344-49f7-8696-a5bcafc57c69 tempest-ServerRescueTestJSONUnderV235-1784422495 tempest-ServerRescueTestJSONUnderV235-1784422495-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 800.571577] env[61923]: DEBUG nova.virt.hardware [None req-28282894-8344-49f7-8696-a5bcafc57c69 tempest-ServerRescueTestJSONUnderV235-1784422495 tempest-ServerRescueTestJSONUnderV235-1784422495-project-member] Got 1 possible topologies {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 800.571699] env[61923]: DEBUG nova.virt.hardware [None req-28282894-8344-49f7-8696-a5bcafc57c69 tempest-ServerRescueTestJSONUnderV235-1784422495 tempest-ServerRescueTestJSONUnderV235-1784422495-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 800.571866] env[61923]: DEBUG nova.virt.hardware [None req-28282894-8344-49f7-8696-a5bcafc57c69 tempest-ServerRescueTestJSONUnderV235-1784422495 tempest-ServerRescueTestJSONUnderV235-1784422495-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 800.572718] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7441e59f-8872-4925-92a6-ac1c0e614f77 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.580489] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81484f6e-3019-4fa5-82a6-6283773c8b27 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.593686] env[61923]: ERROR nova.compute.manager [None req-28282894-8344-49f7-8696-a5bcafc57c69 tempest-ServerRescueTestJSONUnderV235-1784422495 tempest-ServerRescueTestJSONUnderV235-1784422495-project-member] [instance: b2d2b34d-827f-4f96-af8e-936135ba8700] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 85462a2f-eba7-481a-8f13-c572e92d6634, please check neutron logs for more information. [ 800.593686] env[61923]: ERROR nova.compute.manager [instance: b2d2b34d-827f-4f96-af8e-936135ba8700] Traceback (most recent call last): [ 800.593686] env[61923]: ERROR nova.compute.manager [instance: b2d2b34d-827f-4f96-af8e-936135ba8700] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 800.593686] env[61923]: ERROR nova.compute.manager [instance: b2d2b34d-827f-4f96-af8e-936135ba8700] yield resources [ 800.593686] env[61923]: ERROR nova.compute.manager [instance: b2d2b34d-827f-4f96-af8e-936135ba8700] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 800.593686] env[61923]: ERROR nova.compute.manager [instance: b2d2b34d-827f-4f96-af8e-936135ba8700] self.driver.spawn(context, instance, image_meta, [ 800.593686] env[61923]: ERROR nova.compute.manager [instance: b2d2b34d-827f-4f96-af8e-936135ba8700] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 800.593686] env[61923]: ERROR nova.compute.manager [instance: b2d2b34d-827f-4f96-af8e-936135ba8700] self._vmops.spawn(context, instance, image_meta, injected_files, [ 800.593686] env[61923]: ERROR nova.compute.manager [instance: b2d2b34d-827f-4f96-af8e-936135ba8700] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 800.593686] env[61923]: ERROR nova.compute.manager [instance: b2d2b34d-827f-4f96-af8e-936135ba8700] vm_ref = self.build_virtual_machine(instance, [ 800.593686] env[61923]: ERROR nova.compute.manager [instance: b2d2b34d-827f-4f96-af8e-936135ba8700] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 800.593978] env[61923]: ERROR nova.compute.manager [instance: b2d2b34d-827f-4f96-af8e-936135ba8700] vif_infos = vmwarevif.get_vif_info(self._session, [ 800.593978] env[61923]: ERROR nova.compute.manager [instance: b2d2b34d-827f-4f96-af8e-936135ba8700] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 800.593978] env[61923]: ERROR nova.compute.manager [instance: b2d2b34d-827f-4f96-af8e-936135ba8700] for vif in network_info: [ 800.593978] env[61923]: ERROR nova.compute.manager [instance: b2d2b34d-827f-4f96-af8e-936135ba8700] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 800.593978] env[61923]: ERROR nova.compute.manager [instance: b2d2b34d-827f-4f96-af8e-936135ba8700] return self._sync_wrapper(fn, *args, **kwargs) [ 800.593978] env[61923]: ERROR nova.compute.manager [instance: b2d2b34d-827f-4f96-af8e-936135ba8700] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 800.593978] env[61923]: ERROR nova.compute.manager [instance: b2d2b34d-827f-4f96-af8e-936135ba8700] self.wait() [ 800.593978] env[61923]: ERROR nova.compute.manager [instance: b2d2b34d-827f-4f96-af8e-936135ba8700] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 800.593978] env[61923]: ERROR nova.compute.manager [instance: b2d2b34d-827f-4f96-af8e-936135ba8700] self[:] = self._gt.wait() [ 800.593978] env[61923]: ERROR nova.compute.manager [instance: b2d2b34d-827f-4f96-af8e-936135ba8700] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 800.593978] env[61923]: ERROR nova.compute.manager [instance: b2d2b34d-827f-4f96-af8e-936135ba8700] return self._exit_event.wait() [ 800.593978] env[61923]: ERROR nova.compute.manager [instance: b2d2b34d-827f-4f96-af8e-936135ba8700] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 800.593978] env[61923]: ERROR nova.compute.manager [instance: b2d2b34d-827f-4f96-af8e-936135ba8700] current.throw(*self._exc) [ 800.594394] env[61923]: ERROR nova.compute.manager [instance: b2d2b34d-827f-4f96-af8e-936135ba8700] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 800.594394] env[61923]: ERROR nova.compute.manager [instance: b2d2b34d-827f-4f96-af8e-936135ba8700] result = function(*args, **kwargs) [ 800.594394] env[61923]: ERROR nova.compute.manager [instance: b2d2b34d-827f-4f96-af8e-936135ba8700] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 800.594394] env[61923]: ERROR nova.compute.manager [instance: b2d2b34d-827f-4f96-af8e-936135ba8700] return func(*args, **kwargs) [ 800.594394] env[61923]: ERROR nova.compute.manager [instance: b2d2b34d-827f-4f96-af8e-936135ba8700] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 800.594394] env[61923]: ERROR nova.compute.manager [instance: b2d2b34d-827f-4f96-af8e-936135ba8700] raise e [ 800.594394] env[61923]: ERROR nova.compute.manager [instance: b2d2b34d-827f-4f96-af8e-936135ba8700] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 800.594394] env[61923]: ERROR nova.compute.manager [instance: b2d2b34d-827f-4f96-af8e-936135ba8700] nwinfo = self.network_api.allocate_for_instance( [ 800.594394] env[61923]: ERROR nova.compute.manager [instance: b2d2b34d-827f-4f96-af8e-936135ba8700] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 800.594394] env[61923]: ERROR nova.compute.manager [instance: b2d2b34d-827f-4f96-af8e-936135ba8700] created_port_ids = self._update_ports_for_instance( [ 800.594394] env[61923]: ERROR nova.compute.manager [instance: b2d2b34d-827f-4f96-af8e-936135ba8700] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 800.594394] env[61923]: ERROR nova.compute.manager [instance: b2d2b34d-827f-4f96-af8e-936135ba8700] with excutils.save_and_reraise_exception(): [ 800.594394] env[61923]: ERROR nova.compute.manager [instance: b2d2b34d-827f-4f96-af8e-936135ba8700] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 800.594872] env[61923]: ERROR nova.compute.manager [instance: b2d2b34d-827f-4f96-af8e-936135ba8700] self.force_reraise() [ 800.594872] env[61923]: ERROR nova.compute.manager [instance: b2d2b34d-827f-4f96-af8e-936135ba8700] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 800.594872] env[61923]: ERROR nova.compute.manager [instance: b2d2b34d-827f-4f96-af8e-936135ba8700] raise self.value [ 800.594872] env[61923]: ERROR nova.compute.manager [instance: b2d2b34d-827f-4f96-af8e-936135ba8700] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 800.594872] env[61923]: ERROR nova.compute.manager [instance: b2d2b34d-827f-4f96-af8e-936135ba8700] updated_port = self._update_port( [ 800.594872] env[61923]: ERROR nova.compute.manager [instance: b2d2b34d-827f-4f96-af8e-936135ba8700] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 800.594872] env[61923]: ERROR nova.compute.manager [instance: b2d2b34d-827f-4f96-af8e-936135ba8700] _ensure_no_port_binding_failure(port) [ 800.594872] env[61923]: ERROR nova.compute.manager [instance: b2d2b34d-827f-4f96-af8e-936135ba8700] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 800.594872] env[61923]: ERROR nova.compute.manager [instance: b2d2b34d-827f-4f96-af8e-936135ba8700] raise exception.PortBindingFailed(port_id=port['id']) [ 800.594872] env[61923]: ERROR nova.compute.manager [instance: b2d2b34d-827f-4f96-af8e-936135ba8700] nova.exception.PortBindingFailed: Binding failed for port 85462a2f-eba7-481a-8f13-c572e92d6634, please check neutron logs for more information. [ 800.594872] env[61923]: ERROR nova.compute.manager [instance: b2d2b34d-827f-4f96-af8e-936135ba8700] [ 800.594872] env[61923]: INFO nova.compute.manager [None req-28282894-8344-49f7-8696-a5bcafc57c69 tempest-ServerRescueTestJSONUnderV235-1784422495 tempest-ServerRescueTestJSONUnderV235-1784422495-project-member] [instance: b2d2b34d-827f-4f96-af8e-936135ba8700] Terminating instance [ 800.595814] env[61923]: DEBUG oslo_concurrency.lockutils [None req-28282894-8344-49f7-8696-a5bcafc57c69 tempest-ServerRescueTestJSONUnderV235-1784422495 tempest-ServerRescueTestJSONUnderV235-1784422495-project-member] Acquiring lock "refresh_cache-b2d2b34d-827f-4f96-af8e-936135ba8700" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 800.595968] env[61923]: DEBUG oslo_concurrency.lockutils [None req-28282894-8344-49f7-8696-a5bcafc57c69 tempest-ServerRescueTestJSONUnderV235-1784422495 tempest-ServerRescueTestJSONUnderV235-1784422495-project-member] Acquired lock "refresh_cache-b2d2b34d-827f-4f96-af8e-936135ba8700" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 800.596152] env[61923]: DEBUG nova.network.neutron [None req-28282894-8344-49f7-8696-a5bcafc57c69 tempest-ServerRescueTestJSONUnderV235-1784422495 tempest-ServerRescueTestJSONUnderV235-1784422495-project-member] [instance: b2d2b34d-827f-4f96-af8e-936135ba8700] Building network info cache for instance {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 800.863241] env[61923]: DEBUG oslo_concurrency.lockutils [None req-998fcb75-b821-4586-ae48-2fe4d02a9864 tempest-AttachVolumeShelveTestJSON-1252849694 tempest-AttachVolumeShelveTestJSON-1252849694-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.342s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 800.863771] env[61923]: DEBUG nova.compute.manager [None req-998fcb75-b821-4586-ae48-2fe4d02a9864 tempest-AttachVolumeShelveTestJSON-1252849694 tempest-AttachVolumeShelveTestJSON-1252849694-project-member] [instance: 870602dd-1b85-4c97-9506-f08e9f1bade6] Start building networks asynchronously for instance. {{(pid=61923) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 800.866443] env[61923]: DEBUG oslo_concurrency.lockutils [None req-f8f53256-e468-4f43-98b4-c171269c77f6 tempest-ServerDiagnosticsV248Test-1822646244 tempest-ServerDiagnosticsV248Test-1822646244-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 28.407s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 800.866679] env[61923]: DEBUG nova.objects.instance [None req-f8f53256-e468-4f43-98b4-c171269c77f6 tempest-ServerDiagnosticsV248Test-1822646244 tempest-ServerDiagnosticsV248Test-1822646244-project-member] Lazy-loading 'resources' on Instance uuid 655b01ae-76a5-4c09-9245-e629fec8fd4b {{(pid=61923) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 800.977222] env[61923]: DEBUG oslo_concurrency.lockutils [req-e64f4413-6a17-4fab-acdb-907b880fb93a req-3baa98ae-2a4e-4f9c-ad06-2055e953ca1c service nova] Releasing lock "refresh_cache-7bd3204b-5754-4ff2-b0ff-f30556dd72a1" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 800.977474] env[61923]: DEBUG nova.compute.manager [req-e64f4413-6a17-4fab-acdb-907b880fb93a req-3baa98ae-2a4e-4f9c-ad06-2055e953ca1c service nova] [instance: 7bd3204b-5754-4ff2-b0ff-f30556dd72a1] Received event network-vif-deleted-9712b870-b8eb-4048-8ebf-0be8f9c5dfc2 {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 801.112572] env[61923]: DEBUG nova.network.neutron [None req-28282894-8344-49f7-8696-a5bcafc57c69 tempest-ServerRescueTestJSONUnderV235-1784422495 tempest-ServerRescueTestJSONUnderV235-1784422495-project-member] [instance: b2d2b34d-827f-4f96-af8e-936135ba8700] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 801.188018] env[61923]: DEBUG nova.network.neutron [None req-28282894-8344-49f7-8696-a5bcafc57c69 tempest-ServerRescueTestJSONUnderV235-1784422495 tempest-ServerRescueTestJSONUnderV235-1784422495-project-member] [instance: b2d2b34d-827f-4f96-af8e-936135ba8700] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 801.370392] env[61923]: DEBUG nova.compute.utils [None req-998fcb75-b821-4586-ae48-2fe4d02a9864 tempest-AttachVolumeShelveTestJSON-1252849694 tempest-AttachVolumeShelveTestJSON-1252849694-project-member] Using /dev/sd instead of None {{(pid=61923) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 801.374432] env[61923]: DEBUG nova.compute.manager [None req-998fcb75-b821-4586-ae48-2fe4d02a9864 tempest-AttachVolumeShelveTestJSON-1252849694 tempest-AttachVolumeShelveTestJSON-1252849694-project-member] [instance: 870602dd-1b85-4c97-9506-f08e9f1bade6] Allocating IP information in the background. {{(pid=61923) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 801.374606] env[61923]: DEBUG nova.network.neutron [None req-998fcb75-b821-4586-ae48-2fe4d02a9864 tempest-AttachVolumeShelveTestJSON-1252849694 tempest-AttachVolumeShelveTestJSON-1252849694-project-member] [instance: 870602dd-1b85-4c97-9506-f08e9f1bade6] allocate_for_instance() {{(pid=61923) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 801.427744] env[61923]: DEBUG nova.policy [None req-998fcb75-b821-4586-ae48-2fe4d02a9864 tempest-AttachVolumeShelveTestJSON-1252849694 tempest-AttachVolumeShelveTestJSON-1252849694-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'da35e21b822b4ccba98f1a19892fe8f5', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f0dfb7e2632b47f48da6dc458d8f155c', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61923) authorize /opt/stack/nova/nova/policy.py:201}} [ 801.603018] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d19eb65b-dcd8-407b-be0b-4655ff8da4bd {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.611046] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2130a7a-2fc6-486e-affa-3720bb217a54 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.642525] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9d518e8-d2ed-43e5-8bc3-9ec204057caf {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.650368] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af62f745-fda4-4e6e-a2dc-293be6be964a {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.663661] env[61923]: DEBUG nova.compute.provider_tree [None req-f8f53256-e468-4f43-98b4-c171269c77f6 tempest-ServerDiagnosticsV248Test-1822646244 tempest-ServerDiagnosticsV248Test-1822646244-project-member] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 801.690989] env[61923]: DEBUG oslo_concurrency.lockutils [None req-28282894-8344-49f7-8696-a5bcafc57c69 tempest-ServerRescueTestJSONUnderV235-1784422495 tempest-ServerRescueTestJSONUnderV235-1784422495-project-member] Releasing lock "refresh_cache-b2d2b34d-827f-4f96-af8e-936135ba8700" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 801.691364] env[61923]: DEBUG nova.compute.manager [None req-28282894-8344-49f7-8696-a5bcafc57c69 tempest-ServerRescueTestJSONUnderV235-1784422495 tempest-ServerRescueTestJSONUnderV235-1784422495-project-member] [instance: b2d2b34d-827f-4f96-af8e-936135ba8700] Start destroying the instance on the hypervisor. {{(pid=61923) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 801.691552] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-28282894-8344-49f7-8696-a5bcafc57c69 tempest-ServerRescueTestJSONUnderV235-1784422495 tempest-ServerRescueTestJSONUnderV235-1784422495-project-member] [instance: b2d2b34d-827f-4f96-af8e-936135ba8700] Destroying instance {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 801.691820] env[61923]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-2a96f210-4f55-401f-97e4-eb2c743b92cf {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.700507] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c0236c1-2097-4bb9-9a62-c680ad912483 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.724491] env[61923]: WARNING nova.virt.vmwareapi.vmops [None req-28282894-8344-49f7-8696-a5bcafc57c69 tempest-ServerRescueTestJSONUnderV235-1784422495 tempest-ServerRescueTestJSONUnderV235-1784422495-project-member] [instance: b2d2b34d-827f-4f96-af8e-936135ba8700] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance b2d2b34d-827f-4f96-af8e-936135ba8700 could not be found. [ 801.724757] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-28282894-8344-49f7-8696-a5bcafc57c69 tempest-ServerRescueTestJSONUnderV235-1784422495 tempest-ServerRescueTestJSONUnderV235-1784422495-project-member] [instance: b2d2b34d-827f-4f96-af8e-936135ba8700] Instance destroyed {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 801.724963] env[61923]: INFO nova.compute.manager [None req-28282894-8344-49f7-8696-a5bcafc57c69 tempest-ServerRescueTestJSONUnderV235-1784422495 tempest-ServerRescueTestJSONUnderV235-1784422495-project-member] [instance: b2d2b34d-827f-4f96-af8e-936135ba8700] Took 0.03 seconds to destroy the instance on the hypervisor. [ 801.725248] env[61923]: DEBUG oslo.service.loopingcall [None req-28282894-8344-49f7-8696-a5bcafc57c69 tempest-ServerRescueTestJSONUnderV235-1784422495 tempest-ServerRescueTestJSONUnderV235-1784422495-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61923) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 801.725497] env[61923]: DEBUG nova.compute.manager [-] [instance: b2d2b34d-827f-4f96-af8e-936135ba8700] Deallocating network for instance {{(pid=61923) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 801.725606] env[61923]: DEBUG nova.network.neutron [-] [instance: b2d2b34d-827f-4f96-af8e-936135ba8700] deallocate_for_instance() {{(pid=61923) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 801.857613] env[61923]: DEBUG nova.network.neutron [-] [instance: b2d2b34d-827f-4f96-af8e-936135ba8700] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 801.875637] env[61923]: DEBUG nova.compute.manager [None req-998fcb75-b821-4586-ae48-2fe4d02a9864 tempest-AttachVolumeShelveTestJSON-1252849694 tempest-AttachVolumeShelveTestJSON-1252849694-project-member] [instance: 870602dd-1b85-4c97-9506-f08e9f1bade6] Start building block device mappings for instance. {{(pid=61923) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 801.906617] env[61923]: DEBUG nova.network.neutron [None req-998fcb75-b821-4586-ae48-2fe4d02a9864 tempest-AttachVolumeShelveTestJSON-1252849694 tempest-AttachVolumeShelveTestJSON-1252849694-project-member] [instance: 870602dd-1b85-4c97-9506-f08e9f1bade6] Successfully created port: eb1ea65c-43f8-4f04-ae2d-6429b469985d {{(pid=61923) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 801.946908] env[61923]: DEBUG nova.compute.manager [req-aa76e431-20a0-4924-b6f1-21cb327c1668 req-3a53a892-cd50-43cb-8d0f-2d5c9d83511a service nova] [instance: b2d2b34d-827f-4f96-af8e-936135ba8700] Received event network-changed-85462a2f-eba7-481a-8f13-c572e92d6634 {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 801.947050] env[61923]: DEBUG nova.compute.manager [req-aa76e431-20a0-4924-b6f1-21cb327c1668 req-3a53a892-cd50-43cb-8d0f-2d5c9d83511a service nova] [instance: b2d2b34d-827f-4f96-af8e-936135ba8700] Refreshing instance network info cache due to event network-changed-85462a2f-eba7-481a-8f13-c572e92d6634. {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 801.947248] env[61923]: DEBUG oslo_concurrency.lockutils [req-aa76e431-20a0-4924-b6f1-21cb327c1668 req-3a53a892-cd50-43cb-8d0f-2d5c9d83511a service nova] Acquiring lock "refresh_cache-b2d2b34d-827f-4f96-af8e-936135ba8700" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 801.947413] env[61923]: DEBUG oslo_concurrency.lockutils [req-aa76e431-20a0-4924-b6f1-21cb327c1668 req-3a53a892-cd50-43cb-8d0f-2d5c9d83511a service nova] Acquired lock "refresh_cache-b2d2b34d-827f-4f96-af8e-936135ba8700" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 801.947543] env[61923]: DEBUG nova.network.neutron [req-aa76e431-20a0-4924-b6f1-21cb327c1668 req-3a53a892-cd50-43cb-8d0f-2d5c9d83511a service nova] [instance: b2d2b34d-827f-4f96-af8e-936135ba8700] Refreshing network info cache for port 85462a2f-eba7-481a-8f13-c572e92d6634 {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 802.167539] env[61923]: DEBUG nova.scheduler.client.report [None req-f8f53256-e468-4f43-98b4-c171269c77f6 tempest-ServerDiagnosticsV248Test-1822646244 tempest-ServerDiagnosticsV248Test-1822646244-project-member] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 802.360400] env[61923]: DEBUG nova.network.neutron [-] [instance: b2d2b34d-827f-4f96-af8e-936135ba8700] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 802.470603] env[61923]: DEBUG nova.network.neutron [req-aa76e431-20a0-4924-b6f1-21cb327c1668 req-3a53a892-cd50-43cb-8d0f-2d5c9d83511a service nova] [instance: b2d2b34d-827f-4f96-af8e-936135ba8700] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 802.620617] env[61923]: DEBUG nova.network.neutron [req-aa76e431-20a0-4924-b6f1-21cb327c1668 req-3a53a892-cd50-43cb-8d0f-2d5c9d83511a service nova] [instance: b2d2b34d-827f-4f96-af8e-936135ba8700] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 802.672611] env[61923]: DEBUG oslo_concurrency.lockutils [None req-f8f53256-e468-4f43-98b4-c171269c77f6 tempest-ServerDiagnosticsV248Test-1822646244 tempest-ServerDiagnosticsV248Test-1822646244-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.806s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 802.674940] env[61923]: DEBUG oslo_concurrency.lockutils [None req-6d07cbaf-b976-4d41-9fed-82d41c8e5c61 tempest-ServerShowV254Test-921228709 tempest-ServerShowV254Test-921228709-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 23.598s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 802.675467] env[61923]: DEBUG nova.objects.instance [None req-6d07cbaf-b976-4d41-9fed-82d41c8e5c61 tempest-ServerShowV254Test-921228709 tempest-ServerShowV254Test-921228709-project-member] [instance: ecedb77f-4a2a-4047-964f-80796954d847] Trying to apply a migration context that does not seem to be set for this instance {{(pid=61923) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 802.697819] env[61923]: INFO nova.scheduler.client.report [None req-f8f53256-e468-4f43-98b4-c171269c77f6 tempest-ServerDiagnosticsV248Test-1822646244 tempest-ServerDiagnosticsV248Test-1822646244-project-member] Deleted allocations for instance 655b01ae-76a5-4c09-9245-e629fec8fd4b [ 802.863506] env[61923]: INFO nova.compute.manager [-] [instance: b2d2b34d-827f-4f96-af8e-936135ba8700] Took 1.14 seconds to deallocate network for instance. [ 802.866071] env[61923]: DEBUG nova.compute.claims [None req-28282894-8344-49f7-8696-a5bcafc57c69 tempest-ServerRescueTestJSONUnderV235-1784422495 tempest-ServerRescueTestJSONUnderV235-1784422495-project-member] [instance: b2d2b34d-827f-4f96-af8e-936135ba8700] Aborting claim: {{(pid=61923) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 802.867328] env[61923]: DEBUG oslo_concurrency.lockutils [None req-28282894-8344-49f7-8696-a5bcafc57c69 tempest-ServerRescueTestJSONUnderV235-1784422495 tempest-ServerRescueTestJSONUnderV235-1784422495-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 802.888801] env[61923]: DEBUG nova.compute.manager [None req-998fcb75-b821-4586-ae48-2fe4d02a9864 tempest-AttachVolumeShelveTestJSON-1252849694 tempest-AttachVolumeShelveTestJSON-1252849694-project-member] [instance: 870602dd-1b85-4c97-9506-f08e9f1bade6] Start spawning the instance on the hypervisor. {{(pid=61923) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 802.913755] env[61923]: DEBUG nova.virt.hardware [None req-998fcb75-b821-4586-ae48-2fe4d02a9864 tempest-AttachVolumeShelveTestJSON-1252849694 tempest-AttachVolumeShelveTestJSON-1252849694-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-29T20:07:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-29T20:07:35Z,direct_url=,disk_format='vmdk',id=0f153f63-ae0a-45b1-b7f5-7f9b673c947f,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='4e7b5e3b3c3443c8bd5c70f8a15739f5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-29T20:07:36Z,virtual_size=,visibility=), allow threads: False {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 802.914010] env[61923]: DEBUG nova.virt.hardware [None req-998fcb75-b821-4586-ae48-2fe4d02a9864 tempest-AttachVolumeShelveTestJSON-1252849694 tempest-AttachVolumeShelveTestJSON-1252849694-project-member] Flavor limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 802.914174] env[61923]: DEBUG nova.virt.hardware [None req-998fcb75-b821-4586-ae48-2fe4d02a9864 tempest-AttachVolumeShelveTestJSON-1252849694 tempest-AttachVolumeShelveTestJSON-1252849694-project-member] Image limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 802.914354] env[61923]: DEBUG nova.virt.hardware [None req-998fcb75-b821-4586-ae48-2fe4d02a9864 tempest-AttachVolumeShelveTestJSON-1252849694 tempest-AttachVolumeShelveTestJSON-1252849694-project-member] Flavor pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 802.914497] env[61923]: DEBUG nova.virt.hardware [None req-998fcb75-b821-4586-ae48-2fe4d02a9864 tempest-AttachVolumeShelveTestJSON-1252849694 tempest-AttachVolumeShelveTestJSON-1252849694-project-member] Image pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 802.914641] env[61923]: DEBUG nova.virt.hardware [None req-998fcb75-b821-4586-ae48-2fe4d02a9864 tempest-AttachVolumeShelveTestJSON-1252849694 tempest-AttachVolumeShelveTestJSON-1252849694-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 802.914846] env[61923]: DEBUG nova.virt.hardware [None req-998fcb75-b821-4586-ae48-2fe4d02a9864 tempest-AttachVolumeShelveTestJSON-1252849694 tempest-AttachVolumeShelveTestJSON-1252849694-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 802.915017] env[61923]: DEBUG nova.virt.hardware [None req-998fcb75-b821-4586-ae48-2fe4d02a9864 tempest-AttachVolumeShelveTestJSON-1252849694 tempest-AttachVolumeShelveTestJSON-1252849694-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 802.916560] env[61923]: DEBUG nova.virt.hardware [None req-998fcb75-b821-4586-ae48-2fe4d02a9864 tempest-AttachVolumeShelveTestJSON-1252849694 tempest-AttachVolumeShelveTestJSON-1252849694-project-member] Got 1 possible topologies {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 802.916794] env[61923]: DEBUG nova.virt.hardware [None req-998fcb75-b821-4586-ae48-2fe4d02a9864 tempest-AttachVolumeShelveTestJSON-1252849694 tempest-AttachVolumeShelveTestJSON-1252849694-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 802.916982] env[61923]: DEBUG nova.virt.hardware [None req-998fcb75-b821-4586-ae48-2fe4d02a9864 tempest-AttachVolumeShelveTestJSON-1252849694 tempest-AttachVolumeShelveTestJSON-1252849694-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 802.917847] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1148c6f3-f8d8-407e-9a1f-89153d53aea1 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.926392] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-afe38f36-1cae-4a32-b872-78fb21fe6f2b {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.045260] env[61923]: ERROR nova.compute.manager [None req-998fcb75-b821-4586-ae48-2fe4d02a9864 tempest-AttachVolumeShelveTestJSON-1252849694 tempest-AttachVolumeShelveTestJSON-1252849694-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port eb1ea65c-43f8-4f04-ae2d-6429b469985d, please check neutron logs for more information. [ 803.045260] env[61923]: ERROR nova.compute.manager Traceback (most recent call last): [ 803.045260] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 803.045260] env[61923]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 803.045260] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 803.045260] env[61923]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 803.045260] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 803.045260] env[61923]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 803.045260] env[61923]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 803.045260] env[61923]: ERROR nova.compute.manager self.force_reraise() [ 803.045260] env[61923]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 803.045260] env[61923]: ERROR nova.compute.manager raise self.value [ 803.045260] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 803.045260] env[61923]: ERROR nova.compute.manager updated_port = self._update_port( [ 803.045260] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 803.045260] env[61923]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 803.045730] env[61923]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 803.045730] env[61923]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 803.045730] env[61923]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port eb1ea65c-43f8-4f04-ae2d-6429b469985d, please check neutron logs for more information. [ 803.045730] env[61923]: ERROR nova.compute.manager [ 803.045730] env[61923]: Traceback (most recent call last): [ 803.045730] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 803.045730] env[61923]: listener.cb(fileno) [ 803.045730] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 803.045730] env[61923]: result = function(*args, **kwargs) [ 803.045730] env[61923]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 803.045730] env[61923]: return func(*args, **kwargs) [ 803.045730] env[61923]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 803.045730] env[61923]: raise e [ 803.045730] env[61923]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 803.045730] env[61923]: nwinfo = self.network_api.allocate_for_instance( [ 803.045730] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 803.045730] env[61923]: created_port_ids = self._update_ports_for_instance( [ 803.045730] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 803.045730] env[61923]: with excutils.save_and_reraise_exception(): [ 803.045730] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 803.045730] env[61923]: self.force_reraise() [ 803.045730] env[61923]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 803.045730] env[61923]: raise self.value [ 803.045730] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 803.045730] env[61923]: updated_port = self._update_port( [ 803.045730] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 803.045730] env[61923]: _ensure_no_port_binding_failure(port) [ 803.045730] env[61923]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 803.045730] env[61923]: raise exception.PortBindingFailed(port_id=port['id']) [ 803.046458] env[61923]: nova.exception.PortBindingFailed: Binding failed for port eb1ea65c-43f8-4f04-ae2d-6429b469985d, please check neutron logs for more information. [ 803.046458] env[61923]: Removing descriptor: 17 [ 803.046458] env[61923]: ERROR nova.compute.manager [None req-998fcb75-b821-4586-ae48-2fe4d02a9864 tempest-AttachVolumeShelveTestJSON-1252849694 tempest-AttachVolumeShelveTestJSON-1252849694-project-member] [instance: 870602dd-1b85-4c97-9506-f08e9f1bade6] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port eb1ea65c-43f8-4f04-ae2d-6429b469985d, please check neutron logs for more information. [ 803.046458] env[61923]: ERROR nova.compute.manager [instance: 870602dd-1b85-4c97-9506-f08e9f1bade6] Traceback (most recent call last): [ 803.046458] env[61923]: ERROR nova.compute.manager [instance: 870602dd-1b85-4c97-9506-f08e9f1bade6] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 803.046458] env[61923]: ERROR nova.compute.manager [instance: 870602dd-1b85-4c97-9506-f08e9f1bade6] yield resources [ 803.046458] env[61923]: ERROR nova.compute.manager [instance: 870602dd-1b85-4c97-9506-f08e9f1bade6] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 803.046458] env[61923]: ERROR nova.compute.manager [instance: 870602dd-1b85-4c97-9506-f08e9f1bade6] self.driver.spawn(context, instance, image_meta, [ 803.046458] env[61923]: ERROR nova.compute.manager [instance: 870602dd-1b85-4c97-9506-f08e9f1bade6] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 803.046458] env[61923]: ERROR nova.compute.manager [instance: 870602dd-1b85-4c97-9506-f08e9f1bade6] self._vmops.spawn(context, instance, image_meta, injected_files, [ 803.046458] env[61923]: ERROR nova.compute.manager [instance: 870602dd-1b85-4c97-9506-f08e9f1bade6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 803.046458] env[61923]: ERROR nova.compute.manager [instance: 870602dd-1b85-4c97-9506-f08e9f1bade6] vm_ref = self.build_virtual_machine(instance, [ 803.046757] env[61923]: ERROR nova.compute.manager [instance: 870602dd-1b85-4c97-9506-f08e9f1bade6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 803.046757] env[61923]: ERROR nova.compute.manager [instance: 870602dd-1b85-4c97-9506-f08e9f1bade6] vif_infos = vmwarevif.get_vif_info(self._session, [ 803.046757] env[61923]: ERROR nova.compute.manager [instance: 870602dd-1b85-4c97-9506-f08e9f1bade6] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 803.046757] env[61923]: ERROR nova.compute.manager [instance: 870602dd-1b85-4c97-9506-f08e9f1bade6] for vif in network_info: [ 803.046757] env[61923]: ERROR nova.compute.manager [instance: 870602dd-1b85-4c97-9506-f08e9f1bade6] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 803.046757] env[61923]: ERROR nova.compute.manager [instance: 870602dd-1b85-4c97-9506-f08e9f1bade6] return self._sync_wrapper(fn, *args, **kwargs) [ 803.046757] env[61923]: ERROR nova.compute.manager [instance: 870602dd-1b85-4c97-9506-f08e9f1bade6] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 803.046757] env[61923]: ERROR nova.compute.manager [instance: 870602dd-1b85-4c97-9506-f08e9f1bade6] self.wait() [ 803.046757] env[61923]: ERROR nova.compute.manager [instance: 870602dd-1b85-4c97-9506-f08e9f1bade6] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 803.046757] env[61923]: ERROR nova.compute.manager [instance: 870602dd-1b85-4c97-9506-f08e9f1bade6] self[:] = self._gt.wait() [ 803.046757] env[61923]: ERROR nova.compute.manager [instance: 870602dd-1b85-4c97-9506-f08e9f1bade6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 803.046757] env[61923]: ERROR nova.compute.manager [instance: 870602dd-1b85-4c97-9506-f08e9f1bade6] return self._exit_event.wait() [ 803.046757] env[61923]: ERROR nova.compute.manager [instance: 870602dd-1b85-4c97-9506-f08e9f1bade6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 803.047060] env[61923]: ERROR nova.compute.manager [instance: 870602dd-1b85-4c97-9506-f08e9f1bade6] result = hub.switch() [ 803.047060] env[61923]: ERROR nova.compute.manager [instance: 870602dd-1b85-4c97-9506-f08e9f1bade6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 803.047060] env[61923]: ERROR nova.compute.manager [instance: 870602dd-1b85-4c97-9506-f08e9f1bade6] return self.greenlet.switch() [ 803.047060] env[61923]: ERROR nova.compute.manager [instance: 870602dd-1b85-4c97-9506-f08e9f1bade6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 803.047060] env[61923]: ERROR nova.compute.manager [instance: 870602dd-1b85-4c97-9506-f08e9f1bade6] result = function(*args, **kwargs) [ 803.047060] env[61923]: ERROR nova.compute.manager [instance: 870602dd-1b85-4c97-9506-f08e9f1bade6] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 803.047060] env[61923]: ERROR nova.compute.manager [instance: 870602dd-1b85-4c97-9506-f08e9f1bade6] return func(*args, **kwargs) [ 803.047060] env[61923]: ERROR nova.compute.manager [instance: 870602dd-1b85-4c97-9506-f08e9f1bade6] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 803.047060] env[61923]: ERROR nova.compute.manager [instance: 870602dd-1b85-4c97-9506-f08e9f1bade6] raise e [ 803.047060] env[61923]: ERROR nova.compute.manager [instance: 870602dd-1b85-4c97-9506-f08e9f1bade6] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 803.047060] env[61923]: ERROR nova.compute.manager [instance: 870602dd-1b85-4c97-9506-f08e9f1bade6] nwinfo = self.network_api.allocate_for_instance( [ 803.047060] env[61923]: ERROR nova.compute.manager [instance: 870602dd-1b85-4c97-9506-f08e9f1bade6] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 803.047060] env[61923]: ERROR nova.compute.manager [instance: 870602dd-1b85-4c97-9506-f08e9f1bade6] created_port_ids = self._update_ports_for_instance( [ 803.047336] env[61923]: ERROR nova.compute.manager [instance: 870602dd-1b85-4c97-9506-f08e9f1bade6] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 803.047336] env[61923]: ERROR nova.compute.manager [instance: 870602dd-1b85-4c97-9506-f08e9f1bade6] with excutils.save_and_reraise_exception(): [ 803.047336] env[61923]: ERROR nova.compute.manager [instance: 870602dd-1b85-4c97-9506-f08e9f1bade6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 803.047336] env[61923]: ERROR nova.compute.manager [instance: 870602dd-1b85-4c97-9506-f08e9f1bade6] self.force_reraise() [ 803.047336] env[61923]: ERROR nova.compute.manager [instance: 870602dd-1b85-4c97-9506-f08e9f1bade6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 803.047336] env[61923]: ERROR nova.compute.manager [instance: 870602dd-1b85-4c97-9506-f08e9f1bade6] raise self.value [ 803.047336] env[61923]: ERROR nova.compute.manager [instance: 870602dd-1b85-4c97-9506-f08e9f1bade6] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 803.047336] env[61923]: ERROR nova.compute.manager [instance: 870602dd-1b85-4c97-9506-f08e9f1bade6] updated_port = self._update_port( [ 803.047336] env[61923]: ERROR nova.compute.manager [instance: 870602dd-1b85-4c97-9506-f08e9f1bade6] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 803.047336] env[61923]: ERROR nova.compute.manager [instance: 870602dd-1b85-4c97-9506-f08e9f1bade6] _ensure_no_port_binding_failure(port) [ 803.047336] env[61923]: ERROR nova.compute.manager [instance: 870602dd-1b85-4c97-9506-f08e9f1bade6] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 803.047336] env[61923]: ERROR nova.compute.manager [instance: 870602dd-1b85-4c97-9506-f08e9f1bade6] raise exception.PortBindingFailed(port_id=port['id']) [ 803.047677] env[61923]: ERROR nova.compute.manager [instance: 870602dd-1b85-4c97-9506-f08e9f1bade6] nova.exception.PortBindingFailed: Binding failed for port eb1ea65c-43f8-4f04-ae2d-6429b469985d, please check neutron logs for more information. [ 803.047677] env[61923]: ERROR nova.compute.manager [instance: 870602dd-1b85-4c97-9506-f08e9f1bade6] [ 803.047677] env[61923]: INFO nova.compute.manager [None req-998fcb75-b821-4586-ae48-2fe4d02a9864 tempest-AttachVolumeShelveTestJSON-1252849694 tempest-AttachVolumeShelveTestJSON-1252849694-project-member] [instance: 870602dd-1b85-4c97-9506-f08e9f1bade6] Terminating instance [ 803.049022] env[61923]: DEBUG oslo_concurrency.lockutils [None req-998fcb75-b821-4586-ae48-2fe4d02a9864 tempest-AttachVolumeShelveTestJSON-1252849694 tempest-AttachVolumeShelveTestJSON-1252849694-project-member] Acquiring lock "refresh_cache-870602dd-1b85-4c97-9506-f08e9f1bade6" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 803.049022] env[61923]: DEBUG oslo_concurrency.lockutils [None req-998fcb75-b821-4586-ae48-2fe4d02a9864 tempest-AttachVolumeShelveTestJSON-1252849694 tempest-AttachVolumeShelveTestJSON-1252849694-project-member] Acquired lock "refresh_cache-870602dd-1b85-4c97-9506-f08e9f1bade6" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 803.049022] env[61923]: DEBUG nova.network.neutron [None req-998fcb75-b821-4586-ae48-2fe4d02a9864 tempest-AttachVolumeShelveTestJSON-1252849694 tempest-AttachVolumeShelveTestJSON-1252849694-project-member] [instance: 870602dd-1b85-4c97-9506-f08e9f1bade6] Building network info cache for instance {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 803.123438] env[61923]: DEBUG oslo_concurrency.lockutils [req-aa76e431-20a0-4924-b6f1-21cb327c1668 req-3a53a892-cd50-43cb-8d0f-2d5c9d83511a service nova] Releasing lock "refresh_cache-b2d2b34d-827f-4f96-af8e-936135ba8700" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 803.123769] env[61923]: DEBUG nova.compute.manager [req-aa76e431-20a0-4924-b6f1-21cb327c1668 req-3a53a892-cd50-43cb-8d0f-2d5c9d83511a service nova] [instance: b2d2b34d-827f-4f96-af8e-936135ba8700] Received event network-vif-deleted-85462a2f-eba7-481a-8f13-c572e92d6634 {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 803.208627] env[61923]: DEBUG oslo_concurrency.lockutils [None req-f8f53256-e468-4f43-98b4-c171269c77f6 tempest-ServerDiagnosticsV248Test-1822646244 tempest-ServerDiagnosticsV248Test-1822646244-project-member] Lock "655b01ae-76a5-4c09-9245-e629fec8fd4b" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 34.492s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 803.580440] env[61923]: DEBUG nova.network.neutron [None req-998fcb75-b821-4586-ae48-2fe4d02a9864 tempest-AttachVolumeShelveTestJSON-1252849694 tempest-AttachVolumeShelveTestJSON-1252849694-project-member] [instance: 870602dd-1b85-4c97-9506-f08e9f1bade6] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 803.685970] env[61923]: DEBUG oslo_concurrency.lockutils [None req-6d07cbaf-b976-4d41-9fed-82d41c8e5c61 tempest-ServerShowV254Test-921228709 tempest-ServerShowV254Test-921228709-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.011s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 803.687918] env[61923]: DEBUG nova.network.neutron [None req-998fcb75-b821-4586-ae48-2fe4d02a9864 tempest-AttachVolumeShelveTestJSON-1252849694 tempest-AttachVolumeShelveTestJSON-1252849694-project-member] [instance: 870602dd-1b85-4c97-9506-f08e9f1bade6] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 803.688839] env[61923]: DEBUG oslo_concurrency.lockutils [None req-33807679-7c2d-4470-9448-1b4dbb76ab45 tempest-ServerShowV254Test-921228709 tempest-ServerShowV254Test-921228709-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 19.326s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 803.689073] env[61923]: DEBUG nova.objects.instance [None req-33807679-7c2d-4470-9448-1b4dbb76ab45 tempest-ServerShowV254Test-921228709 tempest-ServerShowV254Test-921228709-project-member] Lazy-loading 'resources' on Instance uuid ecedb77f-4a2a-4047-964f-80796954d847 {{(pid=61923) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 804.049377] env[61923]: DEBUG nova.compute.manager [req-31bd1e17-4592-470a-aa8d-77f595521fcc req-4cdb3106-8813-412d-8bc7-0614637fffab service nova] [instance: 870602dd-1b85-4c97-9506-f08e9f1bade6] Received event network-changed-eb1ea65c-43f8-4f04-ae2d-6429b469985d {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 804.049377] env[61923]: DEBUG nova.compute.manager [req-31bd1e17-4592-470a-aa8d-77f595521fcc req-4cdb3106-8813-412d-8bc7-0614637fffab service nova] [instance: 870602dd-1b85-4c97-9506-f08e9f1bade6] Refreshing instance network info cache due to event network-changed-eb1ea65c-43f8-4f04-ae2d-6429b469985d. {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 804.049865] env[61923]: DEBUG oslo_concurrency.lockutils [req-31bd1e17-4592-470a-aa8d-77f595521fcc req-4cdb3106-8813-412d-8bc7-0614637fffab service nova] Acquiring lock "refresh_cache-870602dd-1b85-4c97-9506-f08e9f1bade6" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 804.191567] env[61923]: DEBUG oslo_concurrency.lockutils [None req-998fcb75-b821-4586-ae48-2fe4d02a9864 tempest-AttachVolumeShelveTestJSON-1252849694 tempest-AttachVolumeShelveTestJSON-1252849694-project-member] Releasing lock "refresh_cache-870602dd-1b85-4c97-9506-f08e9f1bade6" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 804.191998] env[61923]: DEBUG nova.compute.manager [None req-998fcb75-b821-4586-ae48-2fe4d02a9864 tempest-AttachVolumeShelveTestJSON-1252849694 tempest-AttachVolumeShelveTestJSON-1252849694-project-member] [instance: 870602dd-1b85-4c97-9506-f08e9f1bade6] Start destroying the instance on the hypervisor. {{(pid=61923) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 804.192211] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-998fcb75-b821-4586-ae48-2fe4d02a9864 tempest-AttachVolumeShelveTestJSON-1252849694 tempest-AttachVolumeShelveTestJSON-1252849694-project-member] [instance: 870602dd-1b85-4c97-9506-f08e9f1bade6] Destroying instance {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 804.195057] env[61923]: DEBUG oslo_concurrency.lockutils [req-31bd1e17-4592-470a-aa8d-77f595521fcc req-4cdb3106-8813-412d-8bc7-0614637fffab service nova] Acquired lock "refresh_cache-870602dd-1b85-4c97-9506-f08e9f1bade6" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 804.195248] env[61923]: DEBUG nova.network.neutron [req-31bd1e17-4592-470a-aa8d-77f595521fcc req-4cdb3106-8813-412d-8bc7-0614637fffab service nova] [instance: 870602dd-1b85-4c97-9506-f08e9f1bade6] Refreshing network info cache for port eb1ea65c-43f8-4f04-ae2d-6429b469985d {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 804.196241] env[61923]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-485b5838-eef2-4952-9f16-98b4de110e5f {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.205917] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98cb5af4-a543-40d0-b2df-7bd7aafba91a {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.231737] env[61923]: WARNING nova.virt.vmwareapi.vmops [None req-998fcb75-b821-4586-ae48-2fe4d02a9864 tempest-AttachVolumeShelveTestJSON-1252849694 tempest-AttachVolumeShelveTestJSON-1252849694-project-member] [instance: 870602dd-1b85-4c97-9506-f08e9f1bade6] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 870602dd-1b85-4c97-9506-f08e9f1bade6 could not be found. [ 804.231960] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-998fcb75-b821-4586-ae48-2fe4d02a9864 tempest-AttachVolumeShelveTestJSON-1252849694 tempest-AttachVolumeShelveTestJSON-1252849694-project-member] [instance: 870602dd-1b85-4c97-9506-f08e9f1bade6] Instance destroyed {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 804.232157] env[61923]: INFO nova.compute.manager [None req-998fcb75-b821-4586-ae48-2fe4d02a9864 tempest-AttachVolumeShelveTestJSON-1252849694 tempest-AttachVolumeShelveTestJSON-1252849694-project-member] [instance: 870602dd-1b85-4c97-9506-f08e9f1bade6] Took 0.04 seconds to destroy the instance on the hypervisor. [ 804.232399] env[61923]: DEBUG oslo.service.loopingcall [None req-998fcb75-b821-4586-ae48-2fe4d02a9864 tempest-AttachVolumeShelveTestJSON-1252849694 tempest-AttachVolumeShelveTestJSON-1252849694-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61923) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 804.234768] env[61923]: DEBUG nova.compute.manager [-] [instance: 870602dd-1b85-4c97-9506-f08e9f1bade6] Deallocating network for instance {{(pid=61923) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 804.234869] env[61923]: DEBUG nova.network.neutron [-] [instance: 870602dd-1b85-4c97-9506-f08e9f1bade6] deallocate_for_instance() {{(pid=61923) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 804.255930] env[61923]: DEBUG nova.network.neutron [-] [instance: 870602dd-1b85-4c97-9506-f08e9f1bade6] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 804.432951] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9d0a5d8-f49b-4b68-9202-1307d3b77346 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.442035] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f8da6ca-bb57-48df-881a-03d6303a88d6 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.475032] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5ee496f-6abe-4721-9bb6-a76b6e93b43d {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.481949] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-047e461e-dcda-4040-b02b-7a4bb9146b0c {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.497028] env[61923]: DEBUG nova.compute.provider_tree [None req-33807679-7c2d-4470-9448-1b4dbb76ab45 tempest-ServerShowV254Test-921228709 tempest-ServerShowV254Test-921228709-project-member] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 804.722727] env[61923]: DEBUG nova.network.neutron [req-31bd1e17-4592-470a-aa8d-77f595521fcc req-4cdb3106-8813-412d-8bc7-0614637fffab service nova] [instance: 870602dd-1b85-4c97-9506-f08e9f1bade6] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 804.758677] env[61923]: DEBUG nova.network.neutron [-] [instance: 870602dd-1b85-4c97-9506-f08e9f1bade6] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 804.857999] env[61923]: DEBUG nova.network.neutron [req-31bd1e17-4592-470a-aa8d-77f595521fcc req-4cdb3106-8813-412d-8bc7-0614637fffab service nova] [instance: 870602dd-1b85-4c97-9506-f08e9f1bade6] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 804.999749] env[61923]: DEBUG nova.scheduler.client.report [None req-33807679-7c2d-4470-9448-1b4dbb76ab45 tempest-ServerShowV254Test-921228709 tempest-ServerShowV254Test-921228709-project-member] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 805.261708] env[61923]: INFO nova.compute.manager [-] [instance: 870602dd-1b85-4c97-9506-f08e9f1bade6] Took 1.03 seconds to deallocate network for instance. [ 805.264116] env[61923]: DEBUG nova.compute.claims [None req-998fcb75-b821-4586-ae48-2fe4d02a9864 tempest-AttachVolumeShelveTestJSON-1252849694 tempest-AttachVolumeShelveTestJSON-1252849694-project-member] [instance: 870602dd-1b85-4c97-9506-f08e9f1bade6] Aborting claim: {{(pid=61923) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 805.264301] env[61923]: DEBUG oslo_concurrency.lockutils [None req-998fcb75-b821-4586-ae48-2fe4d02a9864 tempest-AttachVolumeShelveTestJSON-1252849694 tempest-AttachVolumeShelveTestJSON-1252849694-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 805.360649] env[61923]: DEBUG oslo_concurrency.lockutils [req-31bd1e17-4592-470a-aa8d-77f595521fcc req-4cdb3106-8813-412d-8bc7-0614637fffab service nova] Releasing lock "refresh_cache-870602dd-1b85-4c97-9506-f08e9f1bade6" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 805.360943] env[61923]: DEBUG nova.compute.manager [req-31bd1e17-4592-470a-aa8d-77f595521fcc req-4cdb3106-8813-412d-8bc7-0614637fffab service nova] [instance: 870602dd-1b85-4c97-9506-f08e9f1bade6] Received event network-vif-deleted-eb1ea65c-43f8-4f04-ae2d-6429b469985d {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 805.506063] env[61923]: DEBUG oslo_concurrency.lockutils [None req-33807679-7c2d-4470-9448-1b4dbb76ab45 tempest-ServerShowV254Test-921228709 tempest-ServerShowV254Test-921228709-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.817s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 805.511775] env[61923]: DEBUG oslo_concurrency.lockutils [None req-d3d04fc8-2bde-4c80-b2d6-dd8057edd757 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 20.104s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 805.538952] env[61923]: INFO nova.scheduler.client.report [None req-33807679-7c2d-4470-9448-1b4dbb76ab45 tempest-ServerShowV254Test-921228709 tempest-ServerShowV254Test-921228709-project-member] Deleted allocations for instance ecedb77f-4a2a-4047-964f-80796954d847 [ 806.050158] env[61923]: DEBUG oslo_concurrency.lockutils [None req-33807679-7c2d-4470-9448-1b4dbb76ab45 tempest-ServerShowV254Test-921228709 tempest-ServerShowV254Test-921228709-project-member] Lock "ecedb77f-4a2a-4047-964f-80796954d847" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 25.836s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 806.257196] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf1abe6c-f122-4dbd-bf16-da492927f955 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.265370] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c64a688-525e-4d9c-a8fc-426d6fb4d90f {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.296136] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0306c33d-e50b-49ca-ad53-16e11d91924a {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.304040] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a00e29f-42d7-418f-bca8-3279e2fbb7ef {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.318893] env[61923]: DEBUG nova.compute.provider_tree [None req-d3d04fc8-2bde-4c80-b2d6-dd8057edd757 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 806.820971] env[61923]: DEBUG nova.scheduler.client.report [None req-d3d04fc8-2bde-4c80-b2d6-dd8057edd757 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 807.327460] env[61923]: DEBUG oslo_concurrency.lockutils [None req-d3d04fc8-2bde-4c80-b2d6-dd8057edd757 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.816s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 807.328109] env[61923]: ERROR nova.compute.manager [None req-d3d04fc8-2bde-4c80-b2d6-dd8057edd757 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] [instance: b497f3f3-001f-4e5e-b6b7-8e2006116856] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port fc48dbbd-d086-4537-ab13-11a91a9ac0fc, please check neutron logs for more information. [ 807.328109] env[61923]: ERROR nova.compute.manager [instance: b497f3f3-001f-4e5e-b6b7-8e2006116856] Traceback (most recent call last): [ 807.328109] env[61923]: ERROR nova.compute.manager [instance: b497f3f3-001f-4e5e-b6b7-8e2006116856] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 807.328109] env[61923]: ERROR nova.compute.manager [instance: b497f3f3-001f-4e5e-b6b7-8e2006116856] self.driver.spawn(context, instance, image_meta, [ 807.328109] env[61923]: ERROR nova.compute.manager [instance: b497f3f3-001f-4e5e-b6b7-8e2006116856] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 807.328109] env[61923]: ERROR nova.compute.manager [instance: b497f3f3-001f-4e5e-b6b7-8e2006116856] self._vmops.spawn(context, instance, image_meta, injected_files, [ 807.328109] env[61923]: ERROR nova.compute.manager [instance: b497f3f3-001f-4e5e-b6b7-8e2006116856] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 807.328109] env[61923]: ERROR nova.compute.manager [instance: b497f3f3-001f-4e5e-b6b7-8e2006116856] vm_ref = self.build_virtual_machine(instance, [ 807.328109] env[61923]: ERROR nova.compute.manager [instance: b497f3f3-001f-4e5e-b6b7-8e2006116856] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 807.328109] env[61923]: ERROR nova.compute.manager [instance: b497f3f3-001f-4e5e-b6b7-8e2006116856] vif_infos = vmwarevif.get_vif_info(self._session, [ 807.328109] env[61923]: ERROR nova.compute.manager [instance: b497f3f3-001f-4e5e-b6b7-8e2006116856] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 807.328403] env[61923]: ERROR nova.compute.manager [instance: b497f3f3-001f-4e5e-b6b7-8e2006116856] for vif in network_info: [ 807.328403] env[61923]: ERROR nova.compute.manager [instance: b497f3f3-001f-4e5e-b6b7-8e2006116856] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 807.328403] env[61923]: ERROR nova.compute.manager [instance: b497f3f3-001f-4e5e-b6b7-8e2006116856] return self._sync_wrapper(fn, *args, **kwargs) [ 807.328403] env[61923]: ERROR nova.compute.manager [instance: b497f3f3-001f-4e5e-b6b7-8e2006116856] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 807.328403] env[61923]: ERROR nova.compute.manager [instance: b497f3f3-001f-4e5e-b6b7-8e2006116856] self.wait() [ 807.328403] env[61923]: ERROR nova.compute.manager [instance: b497f3f3-001f-4e5e-b6b7-8e2006116856] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 807.328403] env[61923]: ERROR nova.compute.manager [instance: b497f3f3-001f-4e5e-b6b7-8e2006116856] self[:] = self._gt.wait() [ 807.328403] env[61923]: ERROR nova.compute.manager [instance: b497f3f3-001f-4e5e-b6b7-8e2006116856] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 807.328403] env[61923]: ERROR nova.compute.manager [instance: b497f3f3-001f-4e5e-b6b7-8e2006116856] return self._exit_event.wait() [ 807.328403] env[61923]: ERROR nova.compute.manager [instance: b497f3f3-001f-4e5e-b6b7-8e2006116856] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 807.328403] env[61923]: ERROR nova.compute.manager [instance: b497f3f3-001f-4e5e-b6b7-8e2006116856] result = hub.switch() [ 807.328403] env[61923]: ERROR nova.compute.manager [instance: b497f3f3-001f-4e5e-b6b7-8e2006116856] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 807.328403] env[61923]: ERROR nova.compute.manager [instance: b497f3f3-001f-4e5e-b6b7-8e2006116856] return self.greenlet.switch() [ 807.328771] env[61923]: ERROR nova.compute.manager [instance: b497f3f3-001f-4e5e-b6b7-8e2006116856] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 807.328771] env[61923]: ERROR nova.compute.manager [instance: b497f3f3-001f-4e5e-b6b7-8e2006116856] result = function(*args, **kwargs) [ 807.328771] env[61923]: ERROR nova.compute.manager [instance: b497f3f3-001f-4e5e-b6b7-8e2006116856] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 807.328771] env[61923]: ERROR nova.compute.manager [instance: b497f3f3-001f-4e5e-b6b7-8e2006116856] return func(*args, **kwargs) [ 807.328771] env[61923]: ERROR nova.compute.manager [instance: b497f3f3-001f-4e5e-b6b7-8e2006116856] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 807.328771] env[61923]: ERROR nova.compute.manager [instance: b497f3f3-001f-4e5e-b6b7-8e2006116856] raise e [ 807.328771] env[61923]: ERROR nova.compute.manager [instance: b497f3f3-001f-4e5e-b6b7-8e2006116856] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 807.328771] env[61923]: ERROR nova.compute.manager [instance: b497f3f3-001f-4e5e-b6b7-8e2006116856] nwinfo = self.network_api.allocate_for_instance( [ 807.328771] env[61923]: ERROR nova.compute.manager [instance: b497f3f3-001f-4e5e-b6b7-8e2006116856] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 807.328771] env[61923]: ERROR nova.compute.manager [instance: b497f3f3-001f-4e5e-b6b7-8e2006116856] created_port_ids = self._update_ports_for_instance( [ 807.328771] env[61923]: ERROR nova.compute.manager [instance: b497f3f3-001f-4e5e-b6b7-8e2006116856] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 807.328771] env[61923]: ERROR nova.compute.manager [instance: b497f3f3-001f-4e5e-b6b7-8e2006116856] with excutils.save_and_reraise_exception(): [ 807.328771] env[61923]: ERROR nova.compute.manager [instance: b497f3f3-001f-4e5e-b6b7-8e2006116856] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 807.329073] env[61923]: ERROR nova.compute.manager [instance: b497f3f3-001f-4e5e-b6b7-8e2006116856] self.force_reraise() [ 807.329073] env[61923]: ERROR nova.compute.manager [instance: b497f3f3-001f-4e5e-b6b7-8e2006116856] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 807.329073] env[61923]: ERROR nova.compute.manager [instance: b497f3f3-001f-4e5e-b6b7-8e2006116856] raise self.value [ 807.329073] env[61923]: ERROR nova.compute.manager [instance: b497f3f3-001f-4e5e-b6b7-8e2006116856] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 807.329073] env[61923]: ERROR nova.compute.manager [instance: b497f3f3-001f-4e5e-b6b7-8e2006116856] updated_port = self._update_port( [ 807.329073] env[61923]: ERROR nova.compute.manager [instance: b497f3f3-001f-4e5e-b6b7-8e2006116856] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 807.329073] env[61923]: ERROR nova.compute.manager [instance: b497f3f3-001f-4e5e-b6b7-8e2006116856] _ensure_no_port_binding_failure(port) [ 807.329073] env[61923]: ERROR nova.compute.manager [instance: b497f3f3-001f-4e5e-b6b7-8e2006116856] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 807.329073] env[61923]: ERROR nova.compute.manager [instance: b497f3f3-001f-4e5e-b6b7-8e2006116856] raise exception.PortBindingFailed(port_id=port['id']) [ 807.329073] env[61923]: ERROR nova.compute.manager [instance: b497f3f3-001f-4e5e-b6b7-8e2006116856] nova.exception.PortBindingFailed: Binding failed for port fc48dbbd-d086-4537-ab13-11a91a9ac0fc, please check neutron logs for more information. [ 807.329073] env[61923]: ERROR nova.compute.manager [instance: b497f3f3-001f-4e5e-b6b7-8e2006116856] [ 807.329315] env[61923]: DEBUG nova.compute.utils [None req-d3d04fc8-2bde-4c80-b2d6-dd8057edd757 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] [instance: b497f3f3-001f-4e5e-b6b7-8e2006116856] Binding failed for port fc48dbbd-d086-4537-ab13-11a91a9ac0fc, please check neutron logs for more information. {{(pid=61923) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 807.330314] env[61923]: DEBUG oslo_concurrency.lockutils [None req-1d78b71c-3449-4286-a17a-dc5396f589fe tempest-ServersAdminTestJSON-678434436 tempest-ServersAdminTestJSON-678434436-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 21.129s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 807.336460] env[61923]: DEBUG nova.compute.manager [None req-d3d04fc8-2bde-4c80-b2d6-dd8057edd757 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] [instance: b497f3f3-001f-4e5e-b6b7-8e2006116856] Build of instance b497f3f3-001f-4e5e-b6b7-8e2006116856 was re-scheduled: Binding failed for port fc48dbbd-d086-4537-ab13-11a91a9ac0fc, please check neutron logs for more information. {{(pid=61923) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 807.336941] env[61923]: DEBUG nova.compute.manager [None req-d3d04fc8-2bde-4c80-b2d6-dd8057edd757 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] [instance: b497f3f3-001f-4e5e-b6b7-8e2006116856] Unplugging VIFs for instance {{(pid=61923) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 807.337184] env[61923]: DEBUG oslo_concurrency.lockutils [None req-d3d04fc8-2bde-4c80-b2d6-dd8057edd757 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] Acquiring lock "refresh_cache-b497f3f3-001f-4e5e-b6b7-8e2006116856" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 807.337327] env[61923]: DEBUG oslo_concurrency.lockutils [None req-d3d04fc8-2bde-4c80-b2d6-dd8057edd757 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] Acquired lock "refresh_cache-b497f3f3-001f-4e5e-b6b7-8e2006116856" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 807.337485] env[61923]: DEBUG nova.network.neutron [None req-d3d04fc8-2bde-4c80-b2d6-dd8057edd757 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] [instance: b497f3f3-001f-4e5e-b6b7-8e2006116856] Building network info cache for instance {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 807.873015] env[61923]: DEBUG nova.network.neutron [None req-d3d04fc8-2bde-4c80-b2d6-dd8057edd757 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] [instance: b497f3f3-001f-4e5e-b6b7-8e2006116856] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 808.084998] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-763952f5-5aec-4aff-a0a2-ca56a6a712b8 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.092270] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3add8e7e-3196-4bdf-996e-e0c4e497e797 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.096640] env[61923]: DEBUG nova.network.neutron [None req-d3d04fc8-2bde-4c80-b2d6-dd8057edd757 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] [instance: b497f3f3-001f-4e5e-b6b7-8e2006116856] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 808.130062] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35826de9-9b69-4280-a697-3d15e6229bba {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.138243] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40098fad-dc83-459e-a496-af818ec23716 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.153522] env[61923]: DEBUG nova.compute.provider_tree [None req-1d78b71c-3449-4286-a17a-dc5396f589fe tempest-ServersAdminTestJSON-678434436 tempest-ServersAdminTestJSON-678434436-project-member] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 808.601516] env[61923]: DEBUG oslo_concurrency.lockutils [None req-d3d04fc8-2bde-4c80-b2d6-dd8057edd757 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] Releasing lock "refresh_cache-b497f3f3-001f-4e5e-b6b7-8e2006116856" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 808.601805] env[61923]: DEBUG nova.compute.manager [None req-d3d04fc8-2bde-4c80-b2d6-dd8057edd757 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61923) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 808.601882] env[61923]: DEBUG nova.compute.manager [None req-d3d04fc8-2bde-4c80-b2d6-dd8057edd757 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] [instance: b497f3f3-001f-4e5e-b6b7-8e2006116856] Deallocating network for instance {{(pid=61923) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 808.602057] env[61923]: DEBUG nova.network.neutron [None req-d3d04fc8-2bde-4c80-b2d6-dd8057edd757 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] [instance: b497f3f3-001f-4e5e-b6b7-8e2006116856] deallocate_for_instance() {{(pid=61923) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 808.618319] env[61923]: DEBUG nova.network.neutron [None req-d3d04fc8-2bde-4c80-b2d6-dd8057edd757 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] [instance: b497f3f3-001f-4e5e-b6b7-8e2006116856] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 808.659064] env[61923]: DEBUG nova.scheduler.client.report [None req-1d78b71c-3449-4286-a17a-dc5396f589fe tempest-ServersAdminTestJSON-678434436 tempest-ServersAdminTestJSON-678434436-project-member] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 809.120945] env[61923]: DEBUG nova.network.neutron [None req-d3d04fc8-2bde-4c80-b2d6-dd8057edd757 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] [instance: b497f3f3-001f-4e5e-b6b7-8e2006116856] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 809.162934] env[61923]: DEBUG oslo_concurrency.lockutils [None req-1d78b71c-3449-4286-a17a-dc5396f589fe tempest-ServersAdminTestJSON-678434436 tempest-ServersAdminTestJSON-678434436-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.832s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 809.163587] env[61923]: ERROR nova.compute.manager [None req-1d78b71c-3449-4286-a17a-dc5396f589fe tempest-ServersAdminTestJSON-678434436 tempest-ServersAdminTestJSON-678434436-project-member] [instance: faff83a1-3532-4396-ac3f-600f86653749] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 556dc345-8ad3-4e4f-8dd5-1d38e265adae, please check neutron logs for more information. [ 809.163587] env[61923]: ERROR nova.compute.manager [instance: faff83a1-3532-4396-ac3f-600f86653749] Traceback (most recent call last): [ 809.163587] env[61923]: ERROR nova.compute.manager [instance: faff83a1-3532-4396-ac3f-600f86653749] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 809.163587] env[61923]: ERROR nova.compute.manager [instance: faff83a1-3532-4396-ac3f-600f86653749] self.driver.spawn(context, instance, image_meta, [ 809.163587] env[61923]: ERROR nova.compute.manager [instance: faff83a1-3532-4396-ac3f-600f86653749] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 809.163587] env[61923]: ERROR nova.compute.manager [instance: faff83a1-3532-4396-ac3f-600f86653749] self._vmops.spawn(context, instance, image_meta, injected_files, [ 809.163587] env[61923]: ERROR nova.compute.manager [instance: faff83a1-3532-4396-ac3f-600f86653749] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 809.163587] env[61923]: ERROR nova.compute.manager [instance: faff83a1-3532-4396-ac3f-600f86653749] vm_ref = self.build_virtual_machine(instance, [ 809.163587] env[61923]: ERROR nova.compute.manager [instance: faff83a1-3532-4396-ac3f-600f86653749] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 809.163587] env[61923]: ERROR nova.compute.manager [instance: faff83a1-3532-4396-ac3f-600f86653749] vif_infos = vmwarevif.get_vif_info(self._session, [ 809.163587] env[61923]: ERROR nova.compute.manager [instance: faff83a1-3532-4396-ac3f-600f86653749] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 809.163969] env[61923]: ERROR nova.compute.manager [instance: faff83a1-3532-4396-ac3f-600f86653749] for vif in network_info: [ 809.163969] env[61923]: ERROR nova.compute.manager [instance: faff83a1-3532-4396-ac3f-600f86653749] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 809.163969] env[61923]: ERROR nova.compute.manager [instance: faff83a1-3532-4396-ac3f-600f86653749] return self._sync_wrapper(fn, *args, **kwargs) [ 809.163969] env[61923]: ERROR nova.compute.manager [instance: faff83a1-3532-4396-ac3f-600f86653749] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 809.163969] env[61923]: ERROR nova.compute.manager [instance: faff83a1-3532-4396-ac3f-600f86653749] self.wait() [ 809.163969] env[61923]: ERROR nova.compute.manager [instance: faff83a1-3532-4396-ac3f-600f86653749] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 809.163969] env[61923]: ERROR nova.compute.manager [instance: faff83a1-3532-4396-ac3f-600f86653749] self[:] = self._gt.wait() [ 809.163969] env[61923]: ERROR nova.compute.manager [instance: faff83a1-3532-4396-ac3f-600f86653749] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 809.163969] env[61923]: ERROR nova.compute.manager [instance: faff83a1-3532-4396-ac3f-600f86653749] return self._exit_event.wait() [ 809.163969] env[61923]: ERROR nova.compute.manager [instance: faff83a1-3532-4396-ac3f-600f86653749] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 809.163969] env[61923]: ERROR nova.compute.manager [instance: faff83a1-3532-4396-ac3f-600f86653749] current.throw(*self._exc) [ 809.163969] env[61923]: ERROR nova.compute.manager [instance: faff83a1-3532-4396-ac3f-600f86653749] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 809.163969] env[61923]: ERROR nova.compute.manager [instance: faff83a1-3532-4396-ac3f-600f86653749] result = function(*args, **kwargs) [ 809.164329] env[61923]: ERROR nova.compute.manager [instance: faff83a1-3532-4396-ac3f-600f86653749] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 809.164329] env[61923]: ERROR nova.compute.manager [instance: faff83a1-3532-4396-ac3f-600f86653749] return func(*args, **kwargs) [ 809.164329] env[61923]: ERROR nova.compute.manager [instance: faff83a1-3532-4396-ac3f-600f86653749] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 809.164329] env[61923]: ERROR nova.compute.manager [instance: faff83a1-3532-4396-ac3f-600f86653749] raise e [ 809.164329] env[61923]: ERROR nova.compute.manager [instance: faff83a1-3532-4396-ac3f-600f86653749] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 809.164329] env[61923]: ERROR nova.compute.manager [instance: faff83a1-3532-4396-ac3f-600f86653749] nwinfo = self.network_api.allocate_for_instance( [ 809.164329] env[61923]: ERROR nova.compute.manager [instance: faff83a1-3532-4396-ac3f-600f86653749] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 809.164329] env[61923]: ERROR nova.compute.manager [instance: faff83a1-3532-4396-ac3f-600f86653749] created_port_ids = self._update_ports_for_instance( [ 809.164329] env[61923]: ERROR nova.compute.manager [instance: faff83a1-3532-4396-ac3f-600f86653749] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 809.164329] env[61923]: ERROR nova.compute.manager [instance: faff83a1-3532-4396-ac3f-600f86653749] with excutils.save_and_reraise_exception(): [ 809.164329] env[61923]: ERROR nova.compute.manager [instance: faff83a1-3532-4396-ac3f-600f86653749] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 809.164329] env[61923]: ERROR nova.compute.manager [instance: faff83a1-3532-4396-ac3f-600f86653749] self.force_reraise() [ 809.164329] env[61923]: ERROR nova.compute.manager [instance: faff83a1-3532-4396-ac3f-600f86653749] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 809.164719] env[61923]: ERROR nova.compute.manager [instance: faff83a1-3532-4396-ac3f-600f86653749] raise self.value [ 809.164719] env[61923]: ERROR nova.compute.manager [instance: faff83a1-3532-4396-ac3f-600f86653749] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 809.164719] env[61923]: ERROR nova.compute.manager [instance: faff83a1-3532-4396-ac3f-600f86653749] updated_port = self._update_port( [ 809.164719] env[61923]: ERROR nova.compute.manager [instance: faff83a1-3532-4396-ac3f-600f86653749] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 809.164719] env[61923]: ERROR nova.compute.manager [instance: faff83a1-3532-4396-ac3f-600f86653749] _ensure_no_port_binding_failure(port) [ 809.164719] env[61923]: ERROR nova.compute.manager [instance: faff83a1-3532-4396-ac3f-600f86653749] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 809.164719] env[61923]: ERROR nova.compute.manager [instance: faff83a1-3532-4396-ac3f-600f86653749] raise exception.PortBindingFailed(port_id=port['id']) [ 809.164719] env[61923]: ERROR nova.compute.manager [instance: faff83a1-3532-4396-ac3f-600f86653749] nova.exception.PortBindingFailed: Binding failed for port 556dc345-8ad3-4e4f-8dd5-1d38e265adae, please check neutron logs for more information. [ 809.164719] env[61923]: ERROR nova.compute.manager [instance: faff83a1-3532-4396-ac3f-600f86653749] [ 809.164719] env[61923]: DEBUG nova.compute.utils [None req-1d78b71c-3449-4286-a17a-dc5396f589fe tempest-ServersAdminTestJSON-678434436 tempest-ServersAdminTestJSON-678434436-project-member] [instance: faff83a1-3532-4396-ac3f-600f86653749] Binding failed for port 556dc345-8ad3-4e4f-8dd5-1d38e265adae, please check neutron logs for more information. {{(pid=61923) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 809.165889] env[61923]: DEBUG oslo_concurrency.lockutils [None req-6e6344ac-0cd2-410e-863e-3289b0c1fe3f tempest-ServersAdminTestJSON-678434436 tempest-ServersAdminTestJSON-678434436-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 20.553s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 809.168985] env[61923]: DEBUG nova.compute.manager [None req-1d78b71c-3449-4286-a17a-dc5396f589fe tempest-ServersAdminTestJSON-678434436 tempest-ServersAdminTestJSON-678434436-project-member] [instance: faff83a1-3532-4396-ac3f-600f86653749] Build of instance faff83a1-3532-4396-ac3f-600f86653749 was re-scheduled: Binding failed for port 556dc345-8ad3-4e4f-8dd5-1d38e265adae, please check neutron logs for more information. {{(pid=61923) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 809.169468] env[61923]: DEBUG nova.compute.manager [None req-1d78b71c-3449-4286-a17a-dc5396f589fe tempest-ServersAdminTestJSON-678434436 tempest-ServersAdminTestJSON-678434436-project-member] [instance: faff83a1-3532-4396-ac3f-600f86653749] Unplugging VIFs for instance {{(pid=61923) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 809.169669] env[61923]: DEBUG oslo_concurrency.lockutils [None req-1d78b71c-3449-4286-a17a-dc5396f589fe tempest-ServersAdminTestJSON-678434436 tempest-ServersAdminTestJSON-678434436-project-member] Acquiring lock "refresh_cache-faff83a1-3532-4396-ac3f-600f86653749" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 809.169791] env[61923]: DEBUG oslo_concurrency.lockutils [None req-1d78b71c-3449-4286-a17a-dc5396f589fe tempest-ServersAdminTestJSON-678434436 tempest-ServersAdminTestJSON-678434436-project-member] Acquired lock "refresh_cache-faff83a1-3532-4396-ac3f-600f86653749" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 809.169951] env[61923]: DEBUG nova.network.neutron [None req-1d78b71c-3449-4286-a17a-dc5396f589fe tempest-ServersAdminTestJSON-678434436 tempest-ServersAdminTestJSON-678434436-project-member] [instance: faff83a1-3532-4396-ac3f-600f86653749] Building network info cache for instance {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 809.624868] env[61923]: INFO nova.compute.manager [None req-d3d04fc8-2bde-4c80-b2d6-dd8057edd757 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] [instance: b497f3f3-001f-4e5e-b6b7-8e2006116856] Took 1.02 seconds to deallocate network for instance. [ 809.691042] env[61923]: DEBUG nova.network.neutron [None req-1d78b71c-3449-4286-a17a-dc5396f589fe tempest-ServersAdminTestJSON-678434436 tempest-ServersAdminTestJSON-678434436-project-member] [instance: faff83a1-3532-4396-ac3f-600f86653749] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 809.802527] env[61923]: DEBUG nova.network.neutron [None req-1d78b71c-3449-4286-a17a-dc5396f589fe tempest-ServersAdminTestJSON-678434436 tempest-ServersAdminTestJSON-678434436-project-member] [instance: faff83a1-3532-4396-ac3f-600f86653749] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 809.879414] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-549eb966-8f4a-44aa-a6c8-3f06c06ae62d {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.886939] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-485ffb2a-cf65-437b-8053-3b3357e77842 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.917241] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-524edfd5-f165-4bc5-9c5f-026cce7b9a2e {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.924174] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af94dacc-0542-43f2-b1db-f7093e4fcdec {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.936671] env[61923]: DEBUG nova.compute.provider_tree [None req-6e6344ac-0cd2-410e-863e-3289b0c1fe3f tempest-ServersAdminTestJSON-678434436 tempest-ServersAdminTestJSON-678434436-project-member] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 810.307552] env[61923]: DEBUG oslo_concurrency.lockutils [None req-1d78b71c-3449-4286-a17a-dc5396f589fe tempest-ServersAdminTestJSON-678434436 tempest-ServersAdminTestJSON-678434436-project-member] Releasing lock "refresh_cache-faff83a1-3532-4396-ac3f-600f86653749" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 810.307784] env[61923]: DEBUG nova.compute.manager [None req-1d78b71c-3449-4286-a17a-dc5396f589fe tempest-ServersAdminTestJSON-678434436 tempest-ServersAdminTestJSON-678434436-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61923) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 810.307972] env[61923]: DEBUG nova.compute.manager [None req-1d78b71c-3449-4286-a17a-dc5396f589fe tempest-ServersAdminTestJSON-678434436 tempest-ServersAdminTestJSON-678434436-project-member] [instance: faff83a1-3532-4396-ac3f-600f86653749] Deallocating network for instance {{(pid=61923) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 810.308160] env[61923]: DEBUG nova.network.neutron [None req-1d78b71c-3449-4286-a17a-dc5396f589fe tempest-ServersAdminTestJSON-678434436 tempest-ServersAdminTestJSON-678434436-project-member] [instance: faff83a1-3532-4396-ac3f-600f86653749] deallocate_for_instance() {{(pid=61923) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 810.322855] env[61923]: DEBUG nova.network.neutron [None req-1d78b71c-3449-4286-a17a-dc5396f589fe tempest-ServersAdminTestJSON-678434436 tempest-ServersAdminTestJSON-678434436-project-member] [instance: faff83a1-3532-4396-ac3f-600f86653749] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 810.440070] env[61923]: DEBUG nova.scheduler.client.report [None req-6e6344ac-0cd2-410e-863e-3289b0c1fe3f tempest-ServersAdminTestJSON-678434436 tempest-ServersAdminTestJSON-678434436-project-member] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 810.653610] env[61923]: INFO nova.scheduler.client.report [None req-d3d04fc8-2bde-4c80-b2d6-dd8057edd757 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] Deleted allocations for instance b497f3f3-001f-4e5e-b6b7-8e2006116856 [ 810.825566] env[61923]: DEBUG nova.network.neutron [None req-1d78b71c-3449-4286-a17a-dc5396f589fe tempest-ServersAdminTestJSON-678434436 tempest-ServersAdminTestJSON-678434436-project-member] [instance: faff83a1-3532-4396-ac3f-600f86653749] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 810.946231] env[61923]: DEBUG oslo_concurrency.lockutils [None req-6e6344ac-0cd2-410e-863e-3289b0c1fe3f tempest-ServersAdminTestJSON-678434436 tempest-ServersAdminTestJSON-678434436-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.780s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 810.947099] env[61923]: ERROR nova.compute.manager [None req-6e6344ac-0cd2-410e-863e-3289b0c1fe3f tempest-ServersAdminTestJSON-678434436 tempest-ServersAdminTestJSON-678434436-project-member] [instance: 3f298e29-69ff-46ac-8018-ceb3990b9848] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port bfaa52bf-dae4-491e-bd1e-681526f95731, please check neutron logs for more information. [ 810.947099] env[61923]: ERROR nova.compute.manager [instance: 3f298e29-69ff-46ac-8018-ceb3990b9848] Traceback (most recent call last): [ 810.947099] env[61923]: ERROR nova.compute.manager [instance: 3f298e29-69ff-46ac-8018-ceb3990b9848] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 810.947099] env[61923]: ERROR nova.compute.manager [instance: 3f298e29-69ff-46ac-8018-ceb3990b9848] self.driver.spawn(context, instance, image_meta, [ 810.947099] env[61923]: ERROR nova.compute.manager [instance: 3f298e29-69ff-46ac-8018-ceb3990b9848] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 810.947099] env[61923]: ERROR nova.compute.manager [instance: 3f298e29-69ff-46ac-8018-ceb3990b9848] self._vmops.spawn(context, instance, image_meta, injected_files, [ 810.947099] env[61923]: ERROR nova.compute.manager [instance: 3f298e29-69ff-46ac-8018-ceb3990b9848] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 810.947099] env[61923]: ERROR nova.compute.manager [instance: 3f298e29-69ff-46ac-8018-ceb3990b9848] vm_ref = self.build_virtual_machine(instance, [ 810.947099] env[61923]: ERROR nova.compute.manager [instance: 3f298e29-69ff-46ac-8018-ceb3990b9848] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 810.947099] env[61923]: ERROR nova.compute.manager [instance: 3f298e29-69ff-46ac-8018-ceb3990b9848] vif_infos = vmwarevif.get_vif_info(self._session, [ 810.947099] env[61923]: ERROR nova.compute.manager [instance: 3f298e29-69ff-46ac-8018-ceb3990b9848] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 810.947411] env[61923]: ERROR nova.compute.manager [instance: 3f298e29-69ff-46ac-8018-ceb3990b9848] for vif in network_info: [ 810.947411] env[61923]: ERROR nova.compute.manager [instance: 3f298e29-69ff-46ac-8018-ceb3990b9848] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 810.947411] env[61923]: ERROR nova.compute.manager [instance: 3f298e29-69ff-46ac-8018-ceb3990b9848] return self._sync_wrapper(fn, *args, **kwargs) [ 810.947411] env[61923]: ERROR nova.compute.manager [instance: 3f298e29-69ff-46ac-8018-ceb3990b9848] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 810.947411] env[61923]: ERROR nova.compute.manager [instance: 3f298e29-69ff-46ac-8018-ceb3990b9848] self.wait() [ 810.947411] env[61923]: ERROR nova.compute.manager [instance: 3f298e29-69ff-46ac-8018-ceb3990b9848] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 810.947411] env[61923]: ERROR nova.compute.manager [instance: 3f298e29-69ff-46ac-8018-ceb3990b9848] self[:] = self._gt.wait() [ 810.947411] env[61923]: ERROR nova.compute.manager [instance: 3f298e29-69ff-46ac-8018-ceb3990b9848] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 810.947411] env[61923]: ERROR nova.compute.manager [instance: 3f298e29-69ff-46ac-8018-ceb3990b9848] return self._exit_event.wait() [ 810.947411] env[61923]: ERROR nova.compute.manager [instance: 3f298e29-69ff-46ac-8018-ceb3990b9848] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 810.947411] env[61923]: ERROR nova.compute.manager [instance: 3f298e29-69ff-46ac-8018-ceb3990b9848] current.throw(*self._exc) [ 810.947411] env[61923]: ERROR nova.compute.manager [instance: 3f298e29-69ff-46ac-8018-ceb3990b9848] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 810.947411] env[61923]: ERROR nova.compute.manager [instance: 3f298e29-69ff-46ac-8018-ceb3990b9848] result = function(*args, **kwargs) [ 810.947721] env[61923]: ERROR nova.compute.manager [instance: 3f298e29-69ff-46ac-8018-ceb3990b9848] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 810.947721] env[61923]: ERROR nova.compute.manager [instance: 3f298e29-69ff-46ac-8018-ceb3990b9848] return func(*args, **kwargs) [ 810.947721] env[61923]: ERROR nova.compute.manager [instance: 3f298e29-69ff-46ac-8018-ceb3990b9848] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 810.947721] env[61923]: ERROR nova.compute.manager [instance: 3f298e29-69ff-46ac-8018-ceb3990b9848] raise e [ 810.947721] env[61923]: ERROR nova.compute.manager [instance: 3f298e29-69ff-46ac-8018-ceb3990b9848] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 810.947721] env[61923]: ERROR nova.compute.manager [instance: 3f298e29-69ff-46ac-8018-ceb3990b9848] nwinfo = self.network_api.allocate_for_instance( [ 810.947721] env[61923]: ERROR nova.compute.manager [instance: 3f298e29-69ff-46ac-8018-ceb3990b9848] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 810.947721] env[61923]: ERROR nova.compute.manager [instance: 3f298e29-69ff-46ac-8018-ceb3990b9848] created_port_ids = self._update_ports_for_instance( [ 810.947721] env[61923]: ERROR nova.compute.manager [instance: 3f298e29-69ff-46ac-8018-ceb3990b9848] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 810.947721] env[61923]: ERROR nova.compute.manager [instance: 3f298e29-69ff-46ac-8018-ceb3990b9848] with excutils.save_and_reraise_exception(): [ 810.947721] env[61923]: ERROR nova.compute.manager [instance: 3f298e29-69ff-46ac-8018-ceb3990b9848] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 810.947721] env[61923]: ERROR nova.compute.manager [instance: 3f298e29-69ff-46ac-8018-ceb3990b9848] self.force_reraise() [ 810.947721] env[61923]: ERROR nova.compute.manager [instance: 3f298e29-69ff-46ac-8018-ceb3990b9848] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 810.948066] env[61923]: ERROR nova.compute.manager [instance: 3f298e29-69ff-46ac-8018-ceb3990b9848] raise self.value [ 810.948066] env[61923]: ERROR nova.compute.manager [instance: 3f298e29-69ff-46ac-8018-ceb3990b9848] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 810.948066] env[61923]: ERROR nova.compute.manager [instance: 3f298e29-69ff-46ac-8018-ceb3990b9848] updated_port = self._update_port( [ 810.948066] env[61923]: ERROR nova.compute.manager [instance: 3f298e29-69ff-46ac-8018-ceb3990b9848] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 810.948066] env[61923]: ERROR nova.compute.manager [instance: 3f298e29-69ff-46ac-8018-ceb3990b9848] _ensure_no_port_binding_failure(port) [ 810.948066] env[61923]: ERROR nova.compute.manager [instance: 3f298e29-69ff-46ac-8018-ceb3990b9848] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 810.948066] env[61923]: ERROR nova.compute.manager [instance: 3f298e29-69ff-46ac-8018-ceb3990b9848] raise exception.PortBindingFailed(port_id=port['id']) [ 810.948066] env[61923]: ERROR nova.compute.manager [instance: 3f298e29-69ff-46ac-8018-ceb3990b9848] nova.exception.PortBindingFailed: Binding failed for port bfaa52bf-dae4-491e-bd1e-681526f95731, please check neutron logs for more information. [ 810.948066] env[61923]: ERROR nova.compute.manager [instance: 3f298e29-69ff-46ac-8018-ceb3990b9848] [ 810.948066] env[61923]: DEBUG nova.compute.utils [None req-6e6344ac-0cd2-410e-863e-3289b0c1fe3f tempest-ServersAdminTestJSON-678434436 tempest-ServersAdminTestJSON-678434436-project-member] [instance: 3f298e29-69ff-46ac-8018-ceb3990b9848] Binding failed for port bfaa52bf-dae4-491e-bd1e-681526f95731, please check neutron logs for more information. {{(pid=61923) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 810.949460] env[61923]: DEBUG oslo_concurrency.lockutils [None req-37409b66-7d98-4871-81c4-a83dd0296328 tempest-ServerAddressesNegativeTestJSON-675891034 tempest-ServerAddressesNegativeTestJSON-675891034-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 19.892s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 810.953021] env[61923]: DEBUG nova.compute.manager [None req-6e6344ac-0cd2-410e-863e-3289b0c1fe3f tempest-ServersAdminTestJSON-678434436 tempest-ServersAdminTestJSON-678434436-project-member] [instance: 3f298e29-69ff-46ac-8018-ceb3990b9848] Build of instance 3f298e29-69ff-46ac-8018-ceb3990b9848 was re-scheduled: Binding failed for port bfaa52bf-dae4-491e-bd1e-681526f95731, please check neutron logs for more information. {{(pid=61923) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 810.953155] env[61923]: DEBUG nova.compute.manager [None req-6e6344ac-0cd2-410e-863e-3289b0c1fe3f tempest-ServersAdminTestJSON-678434436 tempest-ServersAdminTestJSON-678434436-project-member] [instance: 3f298e29-69ff-46ac-8018-ceb3990b9848] Unplugging VIFs for instance {{(pid=61923) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 810.953396] env[61923]: DEBUG oslo_concurrency.lockutils [None req-6e6344ac-0cd2-410e-863e-3289b0c1fe3f tempest-ServersAdminTestJSON-678434436 tempest-ServersAdminTestJSON-678434436-project-member] Acquiring lock "refresh_cache-3f298e29-69ff-46ac-8018-ceb3990b9848" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 810.953562] env[61923]: DEBUG oslo_concurrency.lockutils [None req-6e6344ac-0cd2-410e-863e-3289b0c1fe3f tempest-ServersAdminTestJSON-678434436 tempest-ServersAdminTestJSON-678434436-project-member] Acquired lock "refresh_cache-3f298e29-69ff-46ac-8018-ceb3990b9848" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 810.953747] env[61923]: DEBUG nova.network.neutron [None req-6e6344ac-0cd2-410e-863e-3289b0c1fe3f tempest-ServersAdminTestJSON-678434436 tempest-ServersAdminTestJSON-678434436-project-member] [instance: 3f298e29-69ff-46ac-8018-ceb3990b9848] Building network info cache for instance {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 811.162552] env[61923]: DEBUG oslo_concurrency.lockutils [None req-d3d04fc8-2bde-4c80-b2d6-dd8057edd757 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] Lock "b497f3f3-001f-4e5e-b6b7-8e2006116856" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 147.254s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 811.328057] env[61923]: INFO nova.compute.manager [None req-1d78b71c-3449-4286-a17a-dc5396f589fe tempest-ServersAdminTestJSON-678434436 tempest-ServersAdminTestJSON-678434436-project-member] [instance: faff83a1-3532-4396-ac3f-600f86653749] Took 1.02 seconds to deallocate network for instance. [ 811.473744] env[61923]: DEBUG nova.network.neutron [None req-6e6344ac-0cd2-410e-863e-3289b0c1fe3f tempest-ServersAdminTestJSON-678434436 tempest-ServersAdminTestJSON-678434436-project-member] [instance: 3f298e29-69ff-46ac-8018-ceb3990b9848] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 811.581018] env[61923]: DEBUG nova.network.neutron [None req-6e6344ac-0cd2-410e-863e-3289b0c1fe3f tempest-ServersAdminTestJSON-678434436 tempest-ServersAdminTestJSON-678434436-project-member] [instance: 3f298e29-69ff-46ac-8018-ceb3990b9848] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 811.664725] env[61923]: DEBUG nova.compute.manager [None req-c06a566a-3828-47dc-91c4-25d15d717c22 tempest-ServerShowV257Test-986496257 tempest-ServerShowV257Test-986496257-project-member] [instance: b8c60af7-3561-468c-a9a5-1c645080fc69] Starting instance... {{(pid=61923) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 811.752705] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25a89b07-696b-40a7-9211-32bde4b810e5 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.760653] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9cb5aa0b-65c3-4590-abaf-0a84d84e8b79 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.790979] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-560d3b03-f00a-4d0b-b9d1-05fc26325f76 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.798241] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b04ce13-d49e-4243-bfcc-7f01b965c550 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.812078] env[61923]: DEBUG nova.compute.provider_tree [None req-37409b66-7d98-4871-81c4-a83dd0296328 tempest-ServerAddressesNegativeTestJSON-675891034 tempest-ServerAddressesNegativeTestJSON-675891034-project-member] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 812.083215] env[61923]: DEBUG oslo_concurrency.lockutils [None req-6e6344ac-0cd2-410e-863e-3289b0c1fe3f tempest-ServersAdminTestJSON-678434436 tempest-ServersAdminTestJSON-678434436-project-member] Releasing lock "refresh_cache-3f298e29-69ff-46ac-8018-ceb3990b9848" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 812.083475] env[61923]: DEBUG nova.compute.manager [None req-6e6344ac-0cd2-410e-863e-3289b0c1fe3f tempest-ServersAdminTestJSON-678434436 tempest-ServersAdminTestJSON-678434436-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61923) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 812.083661] env[61923]: DEBUG nova.compute.manager [None req-6e6344ac-0cd2-410e-863e-3289b0c1fe3f tempest-ServersAdminTestJSON-678434436 tempest-ServersAdminTestJSON-678434436-project-member] [instance: 3f298e29-69ff-46ac-8018-ceb3990b9848] Deallocating network for instance {{(pid=61923) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 812.083950] env[61923]: DEBUG nova.network.neutron [None req-6e6344ac-0cd2-410e-863e-3289b0c1fe3f tempest-ServersAdminTestJSON-678434436 tempest-ServersAdminTestJSON-678434436-project-member] [instance: 3f298e29-69ff-46ac-8018-ceb3990b9848] deallocate_for_instance() {{(pid=61923) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 812.099175] env[61923]: DEBUG nova.network.neutron [None req-6e6344ac-0cd2-410e-863e-3289b0c1fe3f tempest-ServersAdminTestJSON-678434436 tempest-ServersAdminTestJSON-678434436-project-member] [instance: 3f298e29-69ff-46ac-8018-ceb3990b9848] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 812.191320] env[61923]: DEBUG oslo_concurrency.lockutils [None req-c06a566a-3828-47dc-91c4-25d15d717c22 tempest-ServerShowV257Test-986496257 tempest-ServerShowV257Test-986496257-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 812.315227] env[61923]: DEBUG nova.scheduler.client.report [None req-37409b66-7d98-4871-81c4-a83dd0296328 tempest-ServerAddressesNegativeTestJSON-675891034 tempest-ServerAddressesNegativeTestJSON-675891034-project-member] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 812.357128] env[61923]: INFO nova.scheduler.client.report [None req-1d78b71c-3449-4286-a17a-dc5396f589fe tempest-ServersAdminTestJSON-678434436 tempest-ServersAdminTestJSON-678434436-project-member] Deleted allocations for instance faff83a1-3532-4396-ac3f-600f86653749 [ 812.605039] env[61923]: DEBUG nova.network.neutron [None req-6e6344ac-0cd2-410e-863e-3289b0c1fe3f tempest-ServersAdminTestJSON-678434436 tempest-ServersAdminTestJSON-678434436-project-member] [instance: 3f298e29-69ff-46ac-8018-ceb3990b9848] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 812.820007] env[61923]: DEBUG oslo_concurrency.lockutils [None req-37409b66-7d98-4871-81c4-a83dd0296328 tempest-ServerAddressesNegativeTestJSON-675891034 tempest-ServerAddressesNegativeTestJSON-675891034-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.870s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 812.820670] env[61923]: ERROR nova.compute.manager [None req-37409b66-7d98-4871-81c4-a83dd0296328 tempest-ServerAddressesNegativeTestJSON-675891034 tempest-ServerAddressesNegativeTestJSON-675891034-project-member] [instance: 6fe76341-d578-4cb3-9dce-c025f0d798b8] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 4f871414-2810-4b6d-aaf8-2f56db898ba0, please check neutron logs for more information. [ 812.820670] env[61923]: ERROR nova.compute.manager [instance: 6fe76341-d578-4cb3-9dce-c025f0d798b8] Traceback (most recent call last): [ 812.820670] env[61923]: ERROR nova.compute.manager [instance: 6fe76341-d578-4cb3-9dce-c025f0d798b8] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 812.820670] env[61923]: ERROR nova.compute.manager [instance: 6fe76341-d578-4cb3-9dce-c025f0d798b8] self.driver.spawn(context, instance, image_meta, [ 812.820670] env[61923]: ERROR nova.compute.manager [instance: 6fe76341-d578-4cb3-9dce-c025f0d798b8] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 812.820670] env[61923]: ERROR nova.compute.manager [instance: 6fe76341-d578-4cb3-9dce-c025f0d798b8] self._vmops.spawn(context, instance, image_meta, injected_files, [ 812.820670] env[61923]: ERROR nova.compute.manager [instance: 6fe76341-d578-4cb3-9dce-c025f0d798b8] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 812.820670] env[61923]: ERROR nova.compute.manager [instance: 6fe76341-d578-4cb3-9dce-c025f0d798b8] vm_ref = self.build_virtual_machine(instance, [ 812.820670] env[61923]: ERROR nova.compute.manager [instance: 6fe76341-d578-4cb3-9dce-c025f0d798b8] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 812.820670] env[61923]: ERROR nova.compute.manager [instance: 6fe76341-d578-4cb3-9dce-c025f0d798b8] vif_infos = vmwarevif.get_vif_info(self._session, [ 812.820670] env[61923]: ERROR nova.compute.manager [instance: 6fe76341-d578-4cb3-9dce-c025f0d798b8] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 812.820936] env[61923]: ERROR nova.compute.manager [instance: 6fe76341-d578-4cb3-9dce-c025f0d798b8] for vif in network_info: [ 812.820936] env[61923]: ERROR nova.compute.manager [instance: 6fe76341-d578-4cb3-9dce-c025f0d798b8] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 812.820936] env[61923]: ERROR nova.compute.manager [instance: 6fe76341-d578-4cb3-9dce-c025f0d798b8] return self._sync_wrapper(fn, *args, **kwargs) [ 812.820936] env[61923]: ERROR nova.compute.manager [instance: 6fe76341-d578-4cb3-9dce-c025f0d798b8] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 812.820936] env[61923]: ERROR nova.compute.manager [instance: 6fe76341-d578-4cb3-9dce-c025f0d798b8] self.wait() [ 812.820936] env[61923]: ERROR nova.compute.manager [instance: 6fe76341-d578-4cb3-9dce-c025f0d798b8] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 812.820936] env[61923]: ERROR nova.compute.manager [instance: 6fe76341-d578-4cb3-9dce-c025f0d798b8] self[:] = self._gt.wait() [ 812.820936] env[61923]: ERROR nova.compute.manager [instance: 6fe76341-d578-4cb3-9dce-c025f0d798b8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 812.820936] env[61923]: ERROR nova.compute.manager [instance: 6fe76341-d578-4cb3-9dce-c025f0d798b8] return self._exit_event.wait() [ 812.820936] env[61923]: ERROR nova.compute.manager [instance: 6fe76341-d578-4cb3-9dce-c025f0d798b8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 812.820936] env[61923]: ERROR nova.compute.manager [instance: 6fe76341-d578-4cb3-9dce-c025f0d798b8] result = hub.switch() [ 812.820936] env[61923]: ERROR nova.compute.manager [instance: 6fe76341-d578-4cb3-9dce-c025f0d798b8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 812.820936] env[61923]: ERROR nova.compute.manager [instance: 6fe76341-d578-4cb3-9dce-c025f0d798b8] return self.greenlet.switch() [ 812.821239] env[61923]: ERROR nova.compute.manager [instance: 6fe76341-d578-4cb3-9dce-c025f0d798b8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 812.821239] env[61923]: ERROR nova.compute.manager [instance: 6fe76341-d578-4cb3-9dce-c025f0d798b8] result = function(*args, **kwargs) [ 812.821239] env[61923]: ERROR nova.compute.manager [instance: 6fe76341-d578-4cb3-9dce-c025f0d798b8] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 812.821239] env[61923]: ERROR nova.compute.manager [instance: 6fe76341-d578-4cb3-9dce-c025f0d798b8] return func(*args, **kwargs) [ 812.821239] env[61923]: ERROR nova.compute.manager [instance: 6fe76341-d578-4cb3-9dce-c025f0d798b8] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 812.821239] env[61923]: ERROR nova.compute.manager [instance: 6fe76341-d578-4cb3-9dce-c025f0d798b8] raise e [ 812.821239] env[61923]: ERROR nova.compute.manager [instance: 6fe76341-d578-4cb3-9dce-c025f0d798b8] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 812.821239] env[61923]: ERROR nova.compute.manager [instance: 6fe76341-d578-4cb3-9dce-c025f0d798b8] nwinfo = self.network_api.allocate_for_instance( [ 812.821239] env[61923]: ERROR nova.compute.manager [instance: 6fe76341-d578-4cb3-9dce-c025f0d798b8] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 812.821239] env[61923]: ERROR nova.compute.manager [instance: 6fe76341-d578-4cb3-9dce-c025f0d798b8] created_port_ids = self._update_ports_for_instance( [ 812.821239] env[61923]: ERROR nova.compute.manager [instance: 6fe76341-d578-4cb3-9dce-c025f0d798b8] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 812.821239] env[61923]: ERROR nova.compute.manager [instance: 6fe76341-d578-4cb3-9dce-c025f0d798b8] with excutils.save_and_reraise_exception(): [ 812.821239] env[61923]: ERROR nova.compute.manager [instance: 6fe76341-d578-4cb3-9dce-c025f0d798b8] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 812.821545] env[61923]: ERROR nova.compute.manager [instance: 6fe76341-d578-4cb3-9dce-c025f0d798b8] self.force_reraise() [ 812.821545] env[61923]: ERROR nova.compute.manager [instance: 6fe76341-d578-4cb3-9dce-c025f0d798b8] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 812.821545] env[61923]: ERROR nova.compute.manager [instance: 6fe76341-d578-4cb3-9dce-c025f0d798b8] raise self.value [ 812.821545] env[61923]: ERROR nova.compute.manager [instance: 6fe76341-d578-4cb3-9dce-c025f0d798b8] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 812.821545] env[61923]: ERROR nova.compute.manager [instance: 6fe76341-d578-4cb3-9dce-c025f0d798b8] updated_port = self._update_port( [ 812.821545] env[61923]: ERROR nova.compute.manager [instance: 6fe76341-d578-4cb3-9dce-c025f0d798b8] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 812.821545] env[61923]: ERROR nova.compute.manager [instance: 6fe76341-d578-4cb3-9dce-c025f0d798b8] _ensure_no_port_binding_failure(port) [ 812.821545] env[61923]: ERROR nova.compute.manager [instance: 6fe76341-d578-4cb3-9dce-c025f0d798b8] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 812.821545] env[61923]: ERROR nova.compute.manager [instance: 6fe76341-d578-4cb3-9dce-c025f0d798b8] raise exception.PortBindingFailed(port_id=port['id']) [ 812.821545] env[61923]: ERROR nova.compute.manager [instance: 6fe76341-d578-4cb3-9dce-c025f0d798b8] nova.exception.PortBindingFailed: Binding failed for port 4f871414-2810-4b6d-aaf8-2f56db898ba0, please check neutron logs for more information. [ 812.821545] env[61923]: ERROR nova.compute.manager [instance: 6fe76341-d578-4cb3-9dce-c025f0d798b8] [ 812.821849] env[61923]: DEBUG nova.compute.utils [None req-37409b66-7d98-4871-81c4-a83dd0296328 tempest-ServerAddressesNegativeTestJSON-675891034 tempest-ServerAddressesNegativeTestJSON-675891034-project-member] [instance: 6fe76341-d578-4cb3-9dce-c025f0d798b8] Binding failed for port 4f871414-2810-4b6d-aaf8-2f56db898ba0, please check neutron logs for more information. {{(pid=61923) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 812.822647] env[61923]: DEBUG oslo_concurrency.lockutils [None req-aed8d24a-43d8-4fbe-ac67-0dd6c37614c4 tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 19.471s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 812.825442] env[61923]: DEBUG nova.compute.manager [None req-37409b66-7d98-4871-81c4-a83dd0296328 tempest-ServerAddressesNegativeTestJSON-675891034 tempest-ServerAddressesNegativeTestJSON-675891034-project-member] [instance: 6fe76341-d578-4cb3-9dce-c025f0d798b8] Build of instance 6fe76341-d578-4cb3-9dce-c025f0d798b8 was re-scheduled: Binding failed for port 4f871414-2810-4b6d-aaf8-2f56db898ba0, please check neutron logs for more information. {{(pid=61923) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 812.825870] env[61923]: DEBUG nova.compute.manager [None req-37409b66-7d98-4871-81c4-a83dd0296328 tempest-ServerAddressesNegativeTestJSON-675891034 tempest-ServerAddressesNegativeTestJSON-675891034-project-member] [instance: 6fe76341-d578-4cb3-9dce-c025f0d798b8] Unplugging VIFs for instance {{(pid=61923) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 812.826267] env[61923]: DEBUG oslo_concurrency.lockutils [None req-37409b66-7d98-4871-81c4-a83dd0296328 tempest-ServerAddressesNegativeTestJSON-675891034 tempest-ServerAddressesNegativeTestJSON-675891034-project-member] Acquiring lock "refresh_cache-6fe76341-d578-4cb3-9dce-c025f0d798b8" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 812.826441] env[61923]: DEBUG oslo_concurrency.lockutils [None req-37409b66-7d98-4871-81c4-a83dd0296328 tempest-ServerAddressesNegativeTestJSON-675891034 tempest-ServerAddressesNegativeTestJSON-675891034-project-member] Acquired lock "refresh_cache-6fe76341-d578-4cb3-9dce-c025f0d798b8" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 812.826628] env[61923]: DEBUG nova.network.neutron [None req-37409b66-7d98-4871-81c4-a83dd0296328 tempest-ServerAddressesNegativeTestJSON-675891034 tempest-ServerAddressesNegativeTestJSON-675891034-project-member] [instance: 6fe76341-d578-4cb3-9dce-c025f0d798b8] Building network info cache for instance {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 812.868616] env[61923]: DEBUG oslo_concurrency.lockutils [None req-1d78b71c-3449-4286-a17a-dc5396f589fe tempest-ServersAdminTestJSON-678434436 tempest-ServersAdminTestJSON-678434436-project-member] Lock "faff83a1-3532-4396-ac3f-600f86653749" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 127.172s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 813.107760] env[61923]: INFO nova.compute.manager [None req-6e6344ac-0cd2-410e-863e-3289b0c1fe3f tempest-ServersAdminTestJSON-678434436 tempest-ServersAdminTestJSON-678434436-project-member] [instance: 3f298e29-69ff-46ac-8018-ceb3990b9848] Took 1.02 seconds to deallocate network for instance. [ 813.347814] env[61923]: DEBUG nova.network.neutron [None req-37409b66-7d98-4871-81c4-a83dd0296328 tempest-ServerAddressesNegativeTestJSON-675891034 tempest-ServerAddressesNegativeTestJSON-675891034-project-member] [instance: 6fe76341-d578-4cb3-9dce-c025f0d798b8] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 813.371649] env[61923]: DEBUG nova.compute.manager [None req-8e8efa20-a2ed-4da9-89f1-5ff799c0382c tempest-ServersNegativeTestMultiTenantJSON-1497120864 tempest-ServersNegativeTestMultiTenantJSON-1497120864-project-member] [instance: 1d1b1dc4-c701-42b2-beba-344d622aef41] Starting instance... {{(pid=61923) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 813.420748] env[61923]: DEBUG oslo_concurrency.lockutils [None req-1214e428-bf00-4cbe-9e1e-7c8be1612c36 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] Acquiring lock "af3c317a-4007-4cea-a060-1e7dde5ce49e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 813.420840] env[61923]: DEBUG oslo_concurrency.lockutils [None req-1214e428-bf00-4cbe-9e1e-7c8be1612c36 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] Lock "af3c317a-4007-4cea-a060-1e7dde5ce49e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 813.453490] env[61923]: DEBUG nova.network.neutron [None req-37409b66-7d98-4871-81c4-a83dd0296328 tempest-ServerAddressesNegativeTestJSON-675891034 tempest-ServerAddressesNegativeTestJSON-675891034-project-member] [instance: 6fe76341-d578-4cb3-9dce-c025f0d798b8] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 813.536535] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6877e896-9940-4424-90a7-21f0f3300fb7 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.544019] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8bc38933-148c-4371-abf3-8f8d759c849a {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.573514] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-570a6575-986d-4e02-9a66-d0339506d798 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.580537] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e70cceaf-6cb1-48dc-b30e-211a0160b02e {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.593230] env[61923]: DEBUG nova.compute.provider_tree [None req-aed8d24a-43d8-4fbe-ac67-0dd6c37614c4 tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 813.898795] env[61923]: DEBUG oslo_concurrency.lockutils [None req-8e8efa20-a2ed-4da9-89f1-5ff799c0382c tempest-ServersNegativeTestMultiTenantJSON-1497120864 tempest-ServersNegativeTestMultiTenantJSON-1497120864-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 813.956147] env[61923]: DEBUG oslo_concurrency.lockutils [None req-37409b66-7d98-4871-81c4-a83dd0296328 tempest-ServerAddressesNegativeTestJSON-675891034 tempest-ServerAddressesNegativeTestJSON-675891034-project-member] Releasing lock "refresh_cache-6fe76341-d578-4cb3-9dce-c025f0d798b8" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 813.956369] env[61923]: DEBUG nova.compute.manager [None req-37409b66-7d98-4871-81c4-a83dd0296328 tempest-ServerAddressesNegativeTestJSON-675891034 tempest-ServerAddressesNegativeTestJSON-675891034-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61923) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 813.956576] env[61923]: DEBUG nova.compute.manager [None req-37409b66-7d98-4871-81c4-a83dd0296328 tempest-ServerAddressesNegativeTestJSON-675891034 tempest-ServerAddressesNegativeTestJSON-675891034-project-member] [instance: 6fe76341-d578-4cb3-9dce-c025f0d798b8] Deallocating network for instance {{(pid=61923) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 813.956762] env[61923]: DEBUG nova.network.neutron [None req-37409b66-7d98-4871-81c4-a83dd0296328 tempest-ServerAddressesNegativeTestJSON-675891034 tempest-ServerAddressesNegativeTestJSON-675891034-project-member] [instance: 6fe76341-d578-4cb3-9dce-c025f0d798b8] deallocate_for_instance() {{(pid=61923) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 813.972993] env[61923]: DEBUG nova.network.neutron [None req-37409b66-7d98-4871-81c4-a83dd0296328 tempest-ServerAddressesNegativeTestJSON-675891034 tempest-ServerAddressesNegativeTestJSON-675891034-project-member] [instance: 6fe76341-d578-4cb3-9dce-c025f0d798b8] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 814.096446] env[61923]: DEBUG nova.scheduler.client.report [None req-aed8d24a-43d8-4fbe-ac67-0dd6c37614c4 tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 814.135043] env[61923]: INFO nova.scheduler.client.report [None req-6e6344ac-0cd2-410e-863e-3289b0c1fe3f tempest-ServersAdminTestJSON-678434436 tempest-ServersAdminTestJSON-678434436-project-member] Deleted allocations for instance 3f298e29-69ff-46ac-8018-ceb3990b9848 [ 814.475756] env[61923]: DEBUG nova.network.neutron [None req-37409b66-7d98-4871-81c4-a83dd0296328 tempest-ServerAddressesNegativeTestJSON-675891034 tempest-ServerAddressesNegativeTestJSON-675891034-project-member] [instance: 6fe76341-d578-4cb3-9dce-c025f0d798b8] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 814.601560] env[61923]: DEBUG oslo_concurrency.lockutils [None req-aed8d24a-43d8-4fbe-ac67-0dd6c37614c4 tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.779s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 814.602234] env[61923]: ERROR nova.compute.manager [None req-aed8d24a-43d8-4fbe-ac67-0dd6c37614c4 tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] [instance: 4d709a3a-65af-4adf-9ad0-4d830a9518c2] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 89b347c2-d7a8-4e9a-bc5c-1468e948003b, please check neutron logs for more information. [ 814.602234] env[61923]: ERROR nova.compute.manager [instance: 4d709a3a-65af-4adf-9ad0-4d830a9518c2] Traceback (most recent call last): [ 814.602234] env[61923]: ERROR nova.compute.manager [instance: 4d709a3a-65af-4adf-9ad0-4d830a9518c2] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 814.602234] env[61923]: ERROR nova.compute.manager [instance: 4d709a3a-65af-4adf-9ad0-4d830a9518c2] self.driver.spawn(context, instance, image_meta, [ 814.602234] env[61923]: ERROR nova.compute.manager [instance: 4d709a3a-65af-4adf-9ad0-4d830a9518c2] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 814.602234] env[61923]: ERROR nova.compute.manager [instance: 4d709a3a-65af-4adf-9ad0-4d830a9518c2] self._vmops.spawn(context, instance, image_meta, injected_files, [ 814.602234] env[61923]: ERROR nova.compute.manager [instance: 4d709a3a-65af-4adf-9ad0-4d830a9518c2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 814.602234] env[61923]: ERROR nova.compute.manager [instance: 4d709a3a-65af-4adf-9ad0-4d830a9518c2] vm_ref = self.build_virtual_machine(instance, [ 814.602234] env[61923]: ERROR nova.compute.manager [instance: 4d709a3a-65af-4adf-9ad0-4d830a9518c2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 814.602234] env[61923]: ERROR nova.compute.manager [instance: 4d709a3a-65af-4adf-9ad0-4d830a9518c2] vif_infos = vmwarevif.get_vif_info(self._session, [ 814.602234] env[61923]: ERROR nova.compute.manager [instance: 4d709a3a-65af-4adf-9ad0-4d830a9518c2] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 814.602523] env[61923]: ERROR nova.compute.manager [instance: 4d709a3a-65af-4adf-9ad0-4d830a9518c2] for vif in network_info: [ 814.602523] env[61923]: ERROR nova.compute.manager [instance: 4d709a3a-65af-4adf-9ad0-4d830a9518c2] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 814.602523] env[61923]: ERROR nova.compute.manager [instance: 4d709a3a-65af-4adf-9ad0-4d830a9518c2] return self._sync_wrapper(fn, *args, **kwargs) [ 814.602523] env[61923]: ERROR nova.compute.manager [instance: 4d709a3a-65af-4adf-9ad0-4d830a9518c2] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 814.602523] env[61923]: ERROR nova.compute.manager [instance: 4d709a3a-65af-4adf-9ad0-4d830a9518c2] self.wait() [ 814.602523] env[61923]: ERROR nova.compute.manager [instance: 4d709a3a-65af-4adf-9ad0-4d830a9518c2] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 814.602523] env[61923]: ERROR nova.compute.manager [instance: 4d709a3a-65af-4adf-9ad0-4d830a9518c2] self[:] = self._gt.wait() [ 814.602523] env[61923]: ERROR nova.compute.manager [instance: 4d709a3a-65af-4adf-9ad0-4d830a9518c2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 814.602523] env[61923]: ERROR nova.compute.manager [instance: 4d709a3a-65af-4adf-9ad0-4d830a9518c2] return self._exit_event.wait() [ 814.602523] env[61923]: ERROR nova.compute.manager [instance: 4d709a3a-65af-4adf-9ad0-4d830a9518c2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 814.602523] env[61923]: ERROR nova.compute.manager [instance: 4d709a3a-65af-4adf-9ad0-4d830a9518c2] current.throw(*self._exc) [ 814.602523] env[61923]: ERROR nova.compute.manager [instance: 4d709a3a-65af-4adf-9ad0-4d830a9518c2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 814.602523] env[61923]: ERROR nova.compute.manager [instance: 4d709a3a-65af-4adf-9ad0-4d830a9518c2] result = function(*args, **kwargs) [ 814.602810] env[61923]: ERROR nova.compute.manager [instance: 4d709a3a-65af-4adf-9ad0-4d830a9518c2] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 814.602810] env[61923]: ERROR nova.compute.manager [instance: 4d709a3a-65af-4adf-9ad0-4d830a9518c2] return func(*args, **kwargs) [ 814.602810] env[61923]: ERROR nova.compute.manager [instance: 4d709a3a-65af-4adf-9ad0-4d830a9518c2] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 814.602810] env[61923]: ERROR nova.compute.manager [instance: 4d709a3a-65af-4adf-9ad0-4d830a9518c2] raise e [ 814.602810] env[61923]: ERROR nova.compute.manager [instance: 4d709a3a-65af-4adf-9ad0-4d830a9518c2] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 814.602810] env[61923]: ERROR nova.compute.manager [instance: 4d709a3a-65af-4adf-9ad0-4d830a9518c2] nwinfo = self.network_api.allocate_for_instance( [ 814.602810] env[61923]: ERROR nova.compute.manager [instance: 4d709a3a-65af-4adf-9ad0-4d830a9518c2] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 814.602810] env[61923]: ERROR nova.compute.manager [instance: 4d709a3a-65af-4adf-9ad0-4d830a9518c2] created_port_ids = self._update_ports_for_instance( [ 814.602810] env[61923]: ERROR nova.compute.manager [instance: 4d709a3a-65af-4adf-9ad0-4d830a9518c2] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 814.602810] env[61923]: ERROR nova.compute.manager [instance: 4d709a3a-65af-4adf-9ad0-4d830a9518c2] with excutils.save_and_reraise_exception(): [ 814.602810] env[61923]: ERROR nova.compute.manager [instance: 4d709a3a-65af-4adf-9ad0-4d830a9518c2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 814.602810] env[61923]: ERROR nova.compute.manager [instance: 4d709a3a-65af-4adf-9ad0-4d830a9518c2] self.force_reraise() [ 814.602810] env[61923]: ERROR nova.compute.manager [instance: 4d709a3a-65af-4adf-9ad0-4d830a9518c2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 814.603107] env[61923]: ERROR nova.compute.manager [instance: 4d709a3a-65af-4adf-9ad0-4d830a9518c2] raise self.value [ 814.603107] env[61923]: ERROR nova.compute.manager [instance: 4d709a3a-65af-4adf-9ad0-4d830a9518c2] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 814.603107] env[61923]: ERROR nova.compute.manager [instance: 4d709a3a-65af-4adf-9ad0-4d830a9518c2] updated_port = self._update_port( [ 814.603107] env[61923]: ERROR nova.compute.manager [instance: 4d709a3a-65af-4adf-9ad0-4d830a9518c2] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 814.603107] env[61923]: ERROR nova.compute.manager [instance: 4d709a3a-65af-4adf-9ad0-4d830a9518c2] _ensure_no_port_binding_failure(port) [ 814.603107] env[61923]: ERROR nova.compute.manager [instance: 4d709a3a-65af-4adf-9ad0-4d830a9518c2] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 814.603107] env[61923]: ERROR nova.compute.manager [instance: 4d709a3a-65af-4adf-9ad0-4d830a9518c2] raise exception.PortBindingFailed(port_id=port['id']) [ 814.603107] env[61923]: ERROR nova.compute.manager [instance: 4d709a3a-65af-4adf-9ad0-4d830a9518c2] nova.exception.PortBindingFailed: Binding failed for port 89b347c2-d7a8-4e9a-bc5c-1468e948003b, please check neutron logs for more information. [ 814.603107] env[61923]: ERROR nova.compute.manager [instance: 4d709a3a-65af-4adf-9ad0-4d830a9518c2] [ 814.603107] env[61923]: DEBUG nova.compute.utils [None req-aed8d24a-43d8-4fbe-ac67-0dd6c37614c4 tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] [instance: 4d709a3a-65af-4adf-9ad0-4d830a9518c2] Binding failed for port 89b347c2-d7a8-4e9a-bc5c-1468e948003b, please check neutron logs for more information. {{(pid=61923) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 814.604309] env[61923]: DEBUG oslo_concurrency.lockutils [None req-aed8d24a-43d8-4fbe-ac67-0dd6c37614c4 tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 18.926s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 814.607160] env[61923]: DEBUG nova.compute.manager [None req-aed8d24a-43d8-4fbe-ac67-0dd6c37614c4 tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] [instance: 4d709a3a-65af-4adf-9ad0-4d830a9518c2] Build of instance 4d709a3a-65af-4adf-9ad0-4d830a9518c2 was re-scheduled: Binding failed for port 89b347c2-d7a8-4e9a-bc5c-1468e948003b, please check neutron logs for more information. {{(pid=61923) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 814.607567] env[61923]: DEBUG nova.compute.manager [None req-aed8d24a-43d8-4fbe-ac67-0dd6c37614c4 tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] [instance: 4d709a3a-65af-4adf-9ad0-4d830a9518c2] Unplugging VIFs for instance {{(pid=61923) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 814.607785] env[61923]: DEBUG oslo_concurrency.lockutils [None req-aed8d24a-43d8-4fbe-ac67-0dd6c37614c4 tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Acquiring lock "refresh_cache-4d709a3a-65af-4adf-9ad0-4d830a9518c2" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 814.607930] env[61923]: DEBUG oslo_concurrency.lockutils [None req-aed8d24a-43d8-4fbe-ac67-0dd6c37614c4 tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Acquired lock "refresh_cache-4d709a3a-65af-4adf-9ad0-4d830a9518c2" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 814.608098] env[61923]: DEBUG nova.network.neutron [None req-aed8d24a-43d8-4fbe-ac67-0dd6c37614c4 tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] [instance: 4d709a3a-65af-4adf-9ad0-4d830a9518c2] Building network info cache for instance {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 814.647073] env[61923]: DEBUG oslo_concurrency.lockutils [None req-6e6344ac-0cd2-410e-863e-3289b0c1fe3f tempest-ServersAdminTestJSON-678434436 tempest-ServersAdminTestJSON-678434436-project-member] Lock "3f298e29-69ff-46ac-8018-ceb3990b9848" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 128.024s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 814.980872] env[61923]: INFO nova.compute.manager [None req-37409b66-7d98-4871-81c4-a83dd0296328 tempest-ServerAddressesNegativeTestJSON-675891034 tempest-ServerAddressesNegativeTestJSON-675891034-project-member] [instance: 6fe76341-d578-4cb3-9dce-c025f0d798b8] Took 1.02 seconds to deallocate network for instance. [ 815.151416] env[61923]: DEBUG nova.compute.manager [None req-458eed44-c3ca-4049-bcef-fbcaf60c50ed tempest-ServersTestManualDisk-2012652654 tempest-ServersTestManualDisk-2012652654-project-member] [instance: f1839f61-4314-48fe-8ab6-14b5e49d569d] Starting instance... {{(pid=61923) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 815.176849] env[61923]: DEBUG nova.network.neutron [None req-aed8d24a-43d8-4fbe-ac67-0dd6c37614c4 tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] [instance: 4d709a3a-65af-4adf-9ad0-4d830a9518c2] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 815.327725] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c2d91ff-0979-4585-aa78-90a8a73b4aad {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.331466] env[61923]: DEBUG nova.network.neutron [None req-aed8d24a-43d8-4fbe-ac67-0dd6c37614c4 tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] [instance: 4d709a3a-65af-4adf-9ad0-4d830a9518c2] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 815.338142] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b31ec29-ef95-468f-9dd0-59288e9c59f5 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.377633] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81799c3b-d274-4783-8d28-eb3ad6b2b8dd {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.386999] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d06dbd98-fd43-4e3a-b88e-5828376f827d {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.401648] env[61923]: DEBUG nova.compute.provider_tree [None req-aed8d24a-43d8-4fbe-ac67-0dd6c37614c4 tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 815.417995] env[61923]: DEBUG oslo_concurrency.lockutils [None req-52eadf86-cc45-4cae-af2a-645ac3cb8658 tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Acquiring lock "87ae37dc-d0d0-4b76-afdd-0ba3e8f6ce0b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 815.418239] env[61923]: DEBUG oslo_concurrency.lockutils [None req-52eadf86-cc45-4cae-af2a-645ac3cb8658 tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Lock "87ae37dc-d0d0-4b76-afdd-0ba3e8f6ce0b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 815.673577] env[61923]: DEBUG oslo_concurrency.lockutils [None req-5fbf0568-2a85-4e3c-9b63-7d91ecc77593 tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Acquiring lock "2719569b-8572-4199-8158-7bb367d17dc5" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 815.673807] env[61923]: DEBUG oslo_concurrency.lockutils [None req-5fbf0568-2a85-4e3c-9b63-7d91ecc77593 tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Lock "2719569b-8572-4199-8158-7bb367d17dc5" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 815.681651] env[61923]: DEBUG oslo_concurrency.lockutils [None req-458eed44-c3ca-4049-bcef-fbcaf60c50ed tempest-ServersTestManualDisk-2012652654 tempest-ServersTestManualDisk-2012652654-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 815.833773] env[61923]: DEBUG oslo_concurrency.lockutils [None req-aed8d24a-43d8-4fbe-ac67-0dd6c37614c4 tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Releasing lock "refresh_cache-4d709a3a-65af-4adf-9ad0-4d830a9518c2" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 815.834040] env[61923]: DEBUG nova.compute.manager [None req-aed8d24a-43d8-4fbe-ac67-0dd6c37614c4 tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61923) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 815.834229] env[61923]: DEBUG nova.compute.manager [None req-aed8d24a-43d8-4fbe-ac67-0dd6c37614c4 tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] [instance: 4d709a3a-65af-4adf-9ad0-4d830a9518c2] Deallocating network for instance {{(pid=61923) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 815.834412] env[61923]: DEBUG nova.network.neutron [None req-aed8d24a-43d8-4fbe-ac67-0dd6c37614c4 tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] [instance: 4d709a3a-65af-4adf-9ad0-4d830a9518c2] deallocate_for_instance() {{(pid=61923) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 815.850021] env[61923]: DEBUG nova.network.neutron [None req-aed8d24a-43d8-4fbe-ac67-0dd6c37614c4 tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] [instance: 4d709a3a-65af-4adf-9ad0-4d830a9518c2] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 815.905221] env[61923]: DEBUG nova.scheduler.client.report [None req-aed8d24a-43d8-4fbe-ac67-0dd6c37614c4 tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 816.012017] env[61923]: INFO nova.scheduler.client.report [None req-37409b66-7d98-4871-81c4-a83dd0296328 tempest-ServerAddressesNegativeTestJSON-675891034 tempest-ServerAddressesNegativeTestJSON-675891034-project-member] Deleted allocations for instance 6fe76341-d578-4cb3-9dce-c025f0d798b8 [ 816.353007] env[61923]: DEBUG nova.network.neutron [None req-aed8d24a-43d8-4fbe-ac67-0dd6c37614c4 tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] [instance: 4d709a3a-65af-4adf-9ad0-4d830a9518c2] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 816.410567] env[61923]: DEBUG oslo_concurrency.lockutils [None req-aed8d24a-43d8-4fbe-ac67-0dd6c37614c4 tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.806s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 816.411199] env[61923]: ERROR nova.compute.manager [None req-aed8d24a-43d8-4fbe-ac67-0dd6c37614c4 tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] [instance: fe981195-1e2c-4c8d-8c53-327851f845c7] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 5b85f4c9-ccc5-45a2-a383-7d3b903c5c61, please check neutron logs for more information. [ 816.411199] env[61923]: ERROR nova.compute.manager [instance: fe981195-1e2c-4c8d-8c53-327851f845c7] Traceback (most recent call last): [ 816.411199] env[61923]: ERROR nova.compute.manager [instance: fe981195-1e2c-4c8d-8c53-327851f845c7] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 816.411199] env[61923]: ERROR nova.compute.manager [instance: fe981195-1e2c-4c8d-8c53-327851f845c7] self.driver.spawn(context, instance, image_meta, [ 816.411199] env[61923]: ERROR nova.compute.manager [instance: fe981195-1e2c-4c8d-8c53-327851f845c7] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 816.411199] env[61923]: ERROR nova.compute.manager [instance: fe981195-1e2c-4c8d-8c53-327851f845c7] self._vmops.spawn(context, instance, image_meta, injected_files, [ 816.411199] env[61923]: ERROR nova.compute.manager [instance: fe981195-1e2c-4c8d-8c53-327851f845c7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 816.411199] env[61923]: ERROR nova.compute.manager [instance: fe981195-1e2c-4c8d-8c53-327851f845c7] vm_ref = self.build_virtual_machine(instance, [ 816.411199] env[61923]: ERROR nova.compute.manager [instance: fe981195-1e2c-4c8d-8c53-327851f845c7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 816.411199] env[61923]: ERROR nova.compute.manager [instance: fe981195-1e2c-4c8d-8c53-327851f845c7] vif_infos = vmwarevif.get_vif_info(self._session, [ 816.411199] env[61923]: ERROR nova.compute.manager [instance: fe981195-1e2c-4c8d-8c53-327851f845c7] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 816.411670] env[61923]: ERROR nova.compute.manager [instance: fe981195-1e2c-4c8d-8c53-327851f845c7] for vif in network_info: [ 816.411670] env[61923]: ERROR nova.compute.manager [instance: fe981195-1e2c-4c8d-8c53-327851f845c7] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 816.411670] env[61923]: ERROR nova.compute.manager [instance: fe981195-1e2c-4c8d-8c53-327851f845c7] return self._sync_wrapper(fn, *args, **kwargs) [ 816.411670] env[61923]: ERROR nova.compute.manager [instance: fe981195-1e2c-4c8d-8c53-327851f845c7] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 816.411670] env[61923]: ERROR nova.compute.manager [instance: fe981195-1e2c-4c8d-8c53-327851f845c7] self.wait() [ 816.411670] env[61923]: ERROR nova.compute.manager [instance: fe981195-1e2c-4c8d-8c53-327851f845c7] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 816.411670] env[61923]: ERROR nova.compute.manager [instance: fe981195-1e2c-4c8d-8c53-327851f845c7] self[:] = self._gt.wait() [ 816.411670] env[61923]: ERROR nova.compute.manager [instance: fe981195-1e2c-4c8d-8c53-327851f845c7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 816.411670] env[61923]: ERROR nova.compute.manager [instance: fe981195-1e2c-4c8d-8c53-327851f845c7] return self._exit_event.wait() [ 816.411670] env[61923]: ERROR nova.compute.manager [instance: fe981195-1e2c-4c8d-8c53-327851f845c7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 816.411670] env[61923]: ERROR nova.compute.manager [instance: fe981195-1e2c-4c8d-8c53-327851f845c7] current.throw(*self._exc) [ 816.411670] env[61923]: ERROR nova.compute.manager [instance: fe981195-1e2c-4c8d-8c53-327851f845c7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 816.411670] env[61923]: ERROR nova.compute.manager [instance: fe981195-1e2c-4c8d-8c53-327851f845c7] result = function(*args, **kwargs) [ 816.412048] env[61923]: ERROR nova.compute.manager [instance: fe981195-1e2c-4c8d-8c53-327851f845c7] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 816.412048] env[61923]: ERROR nova.compute.manager [instance: fe981195-1e2c-4c8d-8c53-327851f845c7] return func(*args, **kwargs) [ 816.412048] env[61923]: ERROR nova.compute.manager [instance: fe981195-1e2c-4c8d-8c53-327851f845c7] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 816.412048] env[61923]: ERROR nova.compute.manager [instance: fe981195-1e2c-4c8d-8c53-327851f845c7] raise e [ 816.412048] env[61923]: ERROR nova.compute.manager [instance: fe981195-1e2c-4c8d-8c53-327851f845c7] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 816.412048] env[61923]: ERROR nova.compute.manager [instance: fe981195-1e2c-4c8d-8c53-327851f845c7] nwinfo = self.network_api.allocate_for_instance( [ 816.412048] env[61923]: ERROR nova.compute.manager [instance: fe981195-1e2c-4c8d-8c53-327851f845c7] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 816.412048] env[61923]: ERROR nova.compute.manager [instance: fe981195-1e2c-4c8d-8c53-327851f845c7] created_port_ids = self._update_ports_for_instance( [ 816.412048] env[61923]: ERROR nova.compute.manager [instance: fe981195-1e2c-4c8d-8c53-327851f845c7] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 816.412048] env[61923]: ERROR nova.compute.manager [instance: fe981195-1e2c-4c8d-8c53-327851f845c7] with excutils.save_and_reraise_exception(): [ 816.412048] env[61923]: ERROR nova.compute.manager [instance: fe981195-1e2c-4c8d-8c53-327851f845c7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 816.412048] env[61923]: ERROR nova.compute.manager [instance: fe981195-1e2c-4c8d-8c53-327851f845c7] self.force_reraise() [ 816.412048] env[61923]: ERROR nova.compute.manager [instance: fe981195-1e2c-4c8d-8c53-327851f845c7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 816.412336] env[61923]: ERROR nova.compute.manager [instance: fe981195-1e2c-4c8d-8c53-327851f845c7] raise self.value [ 816.412336] env[61923]: ERROR nova.compute.manager [instance: fe981195-1e2c-4c8d-8c53-327851f845c7] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 816.412336] env[61923]: ERROR nova.compute.manager [instance: fe981195-1e2c-4c8d-8c53-327851f845c7] updated_port = self._update_port( [ 816.412336] env[61923]: ERROR nova.compute.manager [instance: fe981195-1e2c-4c8d-8c53-327851f845c7] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 816.412336] env[61923]: ERROR nova.compute.manager [instance: fe981195-1e2c-4c8d-8c53-327851f845c7] _ensure_no_port_binding_failure(port) [ 816.412336] env[61923]: ERROR nova.compute.manager [instance: fe981195-1e2c-4c8d-8c53-327851f845c7] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 816.412336] env[61923]: ERROR nova.compute.manager [instance: fe981195-1e2c-4c8d-8c53-327851f845c7] raise exception.PortBindingFailed(port_id=port['id']) [ 816.412336] env[61923]: ERROR nova.compute.manager [instance: fe981195-1e2c-4c8d-8c53-327851f845c7] nova.exception.PortBindingFailed: Binding failed for port 5b85f4c9-ccc5-45a2-a383-7d3b903c5c61, please check neutron logs for more information. [ 816.412336] env[61923]: ERROR nova.compute.manager [instance: fe981195-1e2c-4c8d-8c53-327851f845c7] [ 816.412336] env[61923]: DEBUG nova.compute.utils [None req-aed8d24a-43d8-4fbe-ac67-0dd6c37614c4 tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] [instance: fe981195-1e2c-4c8d-8c53-327851f845c7] Binding failed for port 5b85f4c9-ccc5-45a2-a383-7d3b903c5c61, please check neutron logs for more information. {{(pid=61923) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 816.413117] env[61923]: DEBUG oslo_concurrency.lockutils [None req-a9ea97fa-71e5-4463-a1f3-ab69743d14a4 tempest-ServerActionsTestJSON-1295690556 tempest-ServerActionsTestJSON-1295690556-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 17.356s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 816.416012] env[61923]: DEBUG nova.compute.manager [None req-aed8d24a-43d8-4fbe-ac67-0dd6c37614c4 tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] [instance: fe981195-1e2c-4c8d-8c53-327851f845c7] Build of instance fe981195-1e2c-4c8d-8c53-327851f845c7 was re-scheduled: Binding failed for port 5b85f4c9-ccc5-45a2-a383-7d3b903c5c61, please check neutron logs for more information. {{(pid=61923) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 816.416439] env[61923]: DEBUG nova.compute.manager [None req-aed8d24a-43d8-4fbe-ac67-0dd6c37614c4 tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] [instance: fe981195-1e2c-4c8d-8c53-327851f845c7] Unplugging VIFs for instance {{(pid=61923) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 816.416802] env[61923]: DEBUG oslo_concurrency.lockutils [None req-aed8d24a-43d8-4fbe-ac67-0dd6c37614c4 tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Acquiring lock "refresh_cache-fe981195-1e2c-4c8d-8c53-327851f845c7" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 816.417029] env[61923]: DEBUG oslo_concurrency.lockutils [None req-aed8d24a-43d8-4fbe-ac67-0dd6c37614c4 tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Acquired lock "refresh_cache-fe981195-1e2c-4c8d-8c53-327851f845c7" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 816.417213] env[61923]: DEBUG nova.network.neutron [None req-aed8d24a-43d8-4fbe-ac67-0dd6c37614c4 tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] [instance: fe981195-1e2c-4c8d-8c53-327851f845c7] Building network info cache for instance {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 816.520429] env[61923]: DEBUG oslo_concurrency.lockutils [None req-37409b66-7d98-4871-81c4-a83dd0296328 tempest-ServerAddressesNegativeTestJSON-675891034 tempest-ServerAddressesNegativeTestJSON-675891034-project-member] Lock "6fe76341-d578-4cb3-9dce-c025f0d798b8" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 128.858s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 816.855675] env[61923]: INFO nova.compute.manager [None req-aed8d24a-43d8-4fbe-ac67-0dd6c37614c4 tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] [instance: 4d709a3a-65af-4adf-9ad0-4d830a9518c2] Took 1.02 seconds to deallocate network for instance. [ 816.984524] env[61923]: DEBUG nova.network.neutron [None req-aed8d24a-43d8-4fbe-ac67-0dd6c37614c4 tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] [instance: fe981195-1e2c-4c8d-8c53-327851f845c7] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 817.022496] env[61923]: DEBUG nova.compute.manager [None req-9ff8b121-368d-4886-9882-962e7ef81bcf tempest-ServerTagsTestJSON-19215696 tempest-ServerTagsTestJSON-19215696-project-member] [instance: 48825592-dcdc-41d5-9fbf-500d1f31017b] Starting instance... {{(pid=61923) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 817.136650] env[61923]: DEBUG nova.network.neutron [None req-aed8d24a-43d8-4fbe-ac67-0dd6c37614c4 tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] [instance: fe981195-1e2c-4c8d-8c53-327851f845c7] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 817.179735] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea6415a3-827e-4714-a98e-529a3808e548 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.189882] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e81fd720-c67d-4056-9282-40e4d27600e0 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.219712] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9abb11ca-a474-462c-860d-6e067239c235 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.227484] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08aff049-9c6f-441d-930f-59eab99a88d4 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.242508] env[61923]: DEBUG nova.compute.provider_tree [None req-a9ea97fa-71e5-4463-a1f3-ab69743d14a4 tempest-ServerActionsTestJSON-1295690556 tempest-ServerActionsTestJSON-1295690556-project-member] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 817.545431] env[61923]: DEBUG oslo_concurrency.lockutils [None req-9ff8b121-368d-4886-9882-962e7ef81bcf tempest-ServerTagsTestJSON-19215696 tempest-ServerTagsTestJSON-19215696-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 817.639411] env[61923]: DEBUG oslo_concurrency.lockutils [None req-aed8d24a-43d8-4fbe-ac67-0dd6c37614c4 tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Releasing lock "refresh_cache-fe981195-1e2c-4c8d-8c53-327851f845c7" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 817.639702] env[61923]: DEBUG nova.compute.manager [None req-aed8d24a-43d8-4fbe-ac67-0dd6c37614c4 tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61923) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 817.639905] env[61923]: DEBUG nova.compute.manager [None req-aed8d24a-43d8-4fbe-ac67-0dd6c37614c4 tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] [instance: fe981195-1e2c-4c8d-8c53-327851f845c7] Deallocating network for instance {{(pid=61923) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 817.640132] env[61923]: DEBUG nova.network.neutron [None req-aed8d24a-43d8-4fbe-ac67-0dd6c37614c4 tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] [instance: fe981195-1e2c-4c8d-8c53-327851f845c7] deallocate_for_instance() {{(pid=61923) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 817.666582] env[61923]: DEBUG nova.network.neutron [None req-aed8d24a-43d8-4fbe-ac67-0dd6c37614c4 tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] [instance: fe981195-1e2c-4c8d-8c53-327851f845c7] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 817.749018] env[61923]: DEBUG nova.scheduler.client.report [None req-a9ea97fa-71e5-4463-a1f3-ab69743d14a4 tempest-ServerActionsTestJSON-1295690556 tempest-ServerActionsTestJSON-1295690556-project-member] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 817.890819] env[61923]: INFO nova.scheduler.client.report [None req-aed8d24a-43d8-4fbe-ac67-0dd6c37614c4 tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Deleted allocations for instance 4d709a3a-65af-4adf-9ad0-4d830a9518c2 [ 818.170442] env[61923]: DEBUG nova.network.neutron [None req-aed8d24a-43d8-4fbe-ac67-0dd6c37614c4 tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] [instance: fe981195-1e2c-4c8d-8c53-327851f845c7] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 818.252358] env[61923]: DEBUG oslo_concurrency.lockutils [None req-a9ea97fa-71e5-4463-a1f3-ab69743d14a4 tempest-ServerActionsTestJSON-1295690556 tempest-ServerActionsTestJSON-1295690556-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.839s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 818.253007] env[61923]: ERROR nova.compute.manager [None req-a9ea97fa-71e5-4463-a1f3-ab69743d14a4 tempest-ServerActionsTestJSON-1295690556 tempest-ServerActionsTestJSON-1295690556-project-member] [instance: ad3b3933-5bd9-444a-af7b-c81e868e0622] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port d7bef7b0-fae5-4d49-917d-36c20321c890, please check neutron logs for more information. [ 818.253007] env[61923]: ERROR nova.compute.manager [instance: ad3b3933-5bd9-444a-af7b-c81e868e0622] Traceback (most recent call last): [ 818.253007] env[61923]: ERROR nova.compute.manager [instance: ad3b3933-5bd9-444a-af7b-c81e868e0622] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 818.253007] env[61923]: ERROR nova.compute.manager [instance: ad3b3933-5bd9-444a-af7b-c81e868e0622] self.driver.spawn(context, instance, image_meta, [ 818.253007] env[61923]: ERROR nova.compute.manager [instance: ad3b3933-5bd9-444a-af7b-c81e868e0622] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 818.253007] env[61923]: ERROR nova.compute.manager [instance: ad3b3933-5bd9-444a-af7b-c81e868e0622] self._vmops.spawn(context, instance, image_meta, injected_files, [ 818.253007] env[61923]: ERROR nova.compute.manager [instance: ad3b3933-5bd9-444a-af7b-c81e868e0622] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 818.253007] env[61923]: ERROR nova.compute.manager [instance: ad3b3933-5bd9-444a-af7b-c81e868e0622] vm_ref = self.build_virtual_machine(instance, [ 818.253007] env[61923]: ERROR nova.compute.manager [instance: ad3b3933-5bd9-444a-af7b-c81e868e0622] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 818.253007] env[61923]: ERROR nova.compute.manager [instance: ad3b3933-5bd9-444a-af7b-c81e868e0622] vif_infos = vmwarevif.get_vif_info(self._session, [ 818.253007] env[61923]: ERROR nova.compute.manager [instance: ad3b3933-5bd9-444a-af7b-c81e868e0622] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 818.253391] env[61923]: ERROR nova.compute.manager [instance: ad3b3933-5bd9-444a-af7b-c81e868e0622] for vif in network_info: [ 818.253391] env[61923]: ERROR nova.compute.manager [instance: ad3b3933-5bd9-444a-af7b-c81e868e0622] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 818.253391] env[61923]: ERROR nova.compute.manager [instance: ad3b3933-5bd9-444a-af7b-c81e868e0622] return self._sync_wrapper(fn, *args, **kwargs) [ 818.253391] env[61923]: ERROR nova.compute.manager [instance: ad3b3933-5bd9-444a-af7b-c81e868e0622] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 818.253391] env[61923]: ERROR nova.compute.manager [instance: ad3b3933-5bd9-444a-af7b-c81e868e0622] self.wait() [ 818.253391] env[61923]: ERROR nova.compute.manager [instance: ad3b3933-5bd9-444a-af7b-c81e868e0622] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 818.253391] env[61923]: ERROR nova.compute.manager [instance: ad3b3933-5bd9-444a-af7b-c81e868e0622] self[:] = self._gt.wait() [ 818.253391] env[61923]: ERROR nova.compute.manager [instance: ad3b3933-5bd9-444a-af7b-c81e868e0622] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 818.253391] env[61923]: ERROR nova.compute.manager [instance: ad3b3933-5bd9-444a-af7b-c81e868e0622] return self._exit_event.wait() [ 818.253391] env[61923]: ERROR nova.compute.manager [instance: ad3b3933-5bd9-444a-af7b-c81e868e0622] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 818.253391] env[61923]: ERROR nova.compute.manager [instance: ad3b3933-5bd9-444a-af7b-c81e868e0622] result = hub.switch() [ 818.253391] env[61923]: ERROR nova.compute.manager [instance: ad3b3933-5bd9-444a-af7b-c81e868e0622] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 818.253391] env[61923]: ERROR nova.compute.manager [instance: ad3b3933-5bd9-444a-af7b-c81e868e0622] return self.greenlet.switch() [ 818.253669] env[61923]: ERROR nova.compute.manager [instance: ad3b3933-5bd9-444a-af7b-c81e868e0622] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 818.253669] env[61923]: ERROR nova.compute.manager [instance: ad3b3933-5bd9-444a-af7b-c81e868e0622] result = function(*args, **kwargs) [ 818.253669] env[61923]: ERROR nova.compute.manager [instance: ad3b3933-5bd9-444a-af7b-c81e868e0622] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 818.253669] env[61923]: ERROR nova.compute.manager [instance: ad3b3933-5bd9-444a-af7b-c81e868e0622] return func(*args, **kwargs) [ 818.253669] env[61923]: ERROR nova.compute.manager [instance: ad3b3933-5bd9-444a-af7b-c81e868e0622] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 818.253669] env[61923]: ERROR nova.compute.manager [instance: ad3b3933-5bd9-444a-af7b-c81e868e0622] raise e [ 818.253669] env[61923]: ERROR nova.compute.manager [instance: ad3b3933-5bd9-444a-af7b-c81e868e0622] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 818.253669] env[61923]: ERROR nova.compute.manager [instance: ad3b3933-5bd9-444a-af7b-c81e868e0622] nwinfo = self.network_api.allocate_for_instance( [ 818.253669] env[61923]: ERROR nova.compute.manager [instance: ad3b3933-5bd9-444a-af7b-c81e868e0622] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 818.253669] env[61923]: ERROR nova.compute.manager [instance: ad3b3933-5bd9-444a-af7b-c81e868e0622] created_port_ids = self._update_ports_for_instance( [ 818.253669] env[61923]: ERROR nova.compute.manager [instance: ad3b3933-5bd9-444a-af7b-c81e868e0622] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 818.253669] env[61923]: ERROR nova.compute.manager [instance: ad3b3933-5bd9-444a-af7b-c81e868e0622] with excutils.save_and_reraise_exception(): [ 818.253669] env[61923]: ERROR nova.compute.manager [instance: ad3b3933-5bd9-444a-af7b-c81e868e0622] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 818.253947] env[61923]: ERROR nova.compute.manager [instance: ad3b3933-5bd9-444a-af7b-c81e868e0622] self.force_reraise() [ 818.253947] env[61923]: ERROR nova.compute.manager [instance: ad3b3933-5bd9-444a-af7b-c81e868e0622] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 818.253947] env[61923]: ERROR nova.compute.manager [instance: ad3b3933-5bd9-444a-af7b-c81e868e0622] raise self.value [ 818.253947] env[61923]: ERROR nova.compute.manager [instance: ad3b3933-5bd9-444a-af7b-c81e868e0622] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 818.253947] env[61923]: ERROR nova.compute.manager [instance: ad3b3933-5bd9-444a-af7b-c81e868e0622] updated_port = self._update_port( [ 818.253947] env[61923]: ERROR nova.compute.manager [instance: ad3b3933-5bd9-444a-af7b-c81e868e0622] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 818.253947] env[61923]: ERROR nova.compute.manager [instance: ad3b3933-5bd9-444a-af7b-c81e868e0622] _ensure_no_port_binding_failure(port) [ 818.253947] env[61923]: ERROR nova.compute.manager [instance: ad3b3933-5bd9-444a-af7b-c81e868e0622] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 818.253947] env[61923]: ERROR nova.compute.manager [instance: ad3b3933-5bd9-444a-af7b-c81e868e0622] raise exception.PortBindingFailed(port_id=port['id']) [ 818.253947] env[61923]: ERROR nova.compute.manager [instance: ad3b3933-5bd9-444a-af7b-c81e868e0622] nova.exception.PortBindingFailed: Binding failed for port d7bef7b0-fae5-4d49-917d-36c20321c890, please check neutron logs for more information. [ 818.253947] env[61923]: ERROR nova.compute.manager [instance: ad3b3933-5bd9-444a-af7b-c81e868e0622] [ 818.254206] env[61923]: DEBUG nova.compute.utils [None req-a9ea97fa-71e5-4463-a1f3-ab69743d14a4 tempest-ServerActionsTestJSON-1295690556 tempest-ServerActionsTestJSON-1295690556-project-member] [instance: ad3b3933-5bd9-444a-af7b-c81e868e0622] Binding failed for port d7bef7b0-fae5-4d49-917d-36c20321c890, please check neutron logs for more information. {{(pid=61923) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 818.254945] env[61923]: DEBUG oslo_concurrency.lockutils [None req-0adbd41a-bb08-4b5a-bc10-ab9778329dd8 tempest-VolumesAdminNegativeTest-1422120048 tempest-VolumesAdminNegativeTest-1422120048-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 17.885s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 818.257805] env[61923]: DEBUG nova.compute.manager [None req-a9ea97fa-71e5-4463-a1f3-ab69743d14a4 tempest-ServerActionsTestJSON-1295690556 tempest-ServerActionsTestJSON-1295690556-project-member] [instance: ad3b3933-5bd9-444a-af7b-c81e868e0622] Build of instance ad3b3933-5bd9-444a-af7b-c81e868e0622 was re-scheduled: Binding failed for port d7bef7b0-fae5-4d49-917d-36c20321c890, please check neutron logs for more information. {{(pid=61923) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 818.258246] env[61923]: DEBUG nova.compute.manager [None req-a9ea97fa-71e5-4463-a1f3-ab69743d14a4 tempest-ServerActionsTestJSON-1295690556 tempest-ServerActionsTestJSON-1295690556-project-member] [instance: ad3b3933-5bd9-444a-af7b-c81e868e0622] Unplugging VIFs for instance {{(pid=61923) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 818.258608] env[61923]: DEBUG oslo_concurrency.lockutils [None req-a9ea97fa-71e5-4463-a1f3-ab69743d14a4 tempest-ServerActionsTestJSON-1295690556 tempest-ServerActionsTestJSON-1295690556-project-member] Acquiring lock "refresh_cache-ad3b3933-5bd9-444a-af7b-c81e868e0622" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 818.258608] env[61923]: DEBUG oslo_concurrency.lockutils [None req-a9ea97fa-71e5-4463-a1f3-ab69743d14a4 tempest-ServerActionsTestJSON-1295690556 tempest-ServerActionsTestJSON-1295690556-project-member] Acquired lock "refresh_cache-ad3b3933-5bd9-444a-af7b-c81e868e0622" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 818.259374] env[61923]: DEBUG nova.network.neutron [None req-a9ea97fa-71e5-4463-a1f3-ab69743d14a4 tempest-ServerActionsTestJSON-1295690556 tempest-ServerActionsTestJSON-1295690556-project-member] [instance: ad3b3933-5bd9-444a-af7b-c81e868e0622] Building network info cache for instance {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 818.404822] env[61923]: DEBUG oslo_concurrency.lockutils [None req-aed8d24a-43d8-4fbe-ac67-0dd6c37614c4 tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Lock "4d709a3a-65af-4adf-9ad0-4d830a9518c2" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 128.797s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 818.676112] env[61923]: INFO nova.compute.manager [None req-aed8d24a-43d8-4fbe-ac67-0dd6c37614c4 tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] [instance: fe981195-1e2c-4c8d-8c53-327851f845c7] Took 1.04 seconds to deallocate network for instance. [ 818.785438] env[61923]: DEBUG nova.network.neutron [None req-a9ea97fa-71e5-4463-a1f3-ab69743d14a4 tempest-ServerActionsTestJSON-1295690556 tempest-ServerActionsTestJSON-1295690556-project-member] [instance: ad3b3933-5bd9-444a-af7b-c81e868e0622] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 818.891100] env[61923]: DEBUG nova.network.neutron [None req-a9ea97fa-71e5-4463-a1f3-ab69743d14a4 tempest-ServerActionsTestJSON-1295690556 tempest-ServerActionsTestJSON-1295690556-project-member] [instance: ad3b3933-5bd9-444a-af7b-c81e868e0622] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 818.907470] env[61923]: DEBUG nova.compute.manager [None req-950e07c2-502c-43fa-8598-da5f189b07e3 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] [instance: 422d06d6-6932-46c3-bb25-841e6f65c028] Starting instance... {{(pid=61923) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 819.045213] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e22e33fa-dba6-4da7-9d8d-b9016418a7c6 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.053038] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d059fd36-63ed-4c87-8ade-d12cc73c2ba0 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.083835] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f1117af-dbec-4c65-a119-334565b1a560 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.091206] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65197f32-f10f-494b-9d09-996067102fa7 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.104786] env[61923]: DEBUG nova.compute.provider_tree [None req-0adbd41a-bb08-4b5a-bc10-ab9778329dd8 tempest-VolumesAdminNegativeTest-1422120048 tempest-VolumesAdminNegativeTest-1422120048-project-member] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 819.394241] env[61923]: DEBUG oslo_concurrency.lockutils [None req-a9ea97fa-71e5-4463-a1f3-ab69743d14a4 tempest-ServerActionsTestJSON-1295690556 tempest-ServerActionsTestJSON-1295690556-project-member] Releasing lock "refresh_cache-ad3b3933-5bd9-444a-af7b-c81e868e0622" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 819.394241] env[61923]: DEBUG nova.compute.manager [None req-a9ea97fa-71e5-4463-a1f3-ab69743d14a4 tempest-ServerActionsTestJSON-1295690556 tempest-ServerActionsTestJSON-1295690556-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61923) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 819.394241] env[61923]: DEBUG nova.compute.manager [None req-a9ea97fa-71e5-4463-a1f3-ab69743d14a4 tempest-ServerActionsTestJSON-1295690556 tempest-ServerActionsTestJSON-1295690556-project-member] [instance: ad3b3933-5bd9-444a-af7b-c81e868e0622] Deallocating network for instance {{(pid=61923) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 819.394241] env[61923]: DEBUG nova.network.neutron [None req-a9ea97fa-71e5-4463-a1f3-ab69743d14a4 tempest-ServerActionsTestJSON-1295690556 tempest-ServerActionsTestJSON-1295690556-project-member] [instance: ad3b3933-5bd9-444a-af7b-c81e868e0622] deallocate_for_instance() {{(pid=61923) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 819.428259] env[61923]: DEBUG nova.network.neutron [None req-a9ea97fa-71e5-4463-a1f3-ab69743d14a4 tempest-ServerActionsTestJSON-1295690556 tempest-ServerActionsTestJSON-1295690556-project-member] [instance: ad3b3933-5bd9-444a-af7b-c81e868e0622] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 819.430252] env[61923]: DEBUG oslo_concurrency.lockutils [None req-950e07c2-502c-43fa-8598-da5f189b07e3 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 819.608078] env[61923]: DEBUG nova.scheduler.client.report [None req-0adbd41a-bb08-4b5a-bc10-ab9778329dd8 tempest-VolumesAdminNegativeTest-1422120048 tempest-VolumesAdminNegativeTest-1422120048-project-member] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 819.707496] env[61923]: INFO nova.scheduler.client.report [None req-aed8d24a-43d8-4fbe-ac67-0dd6c37614c4 tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Deleted allocations for instance fe981195-1e2c-4c8d-8c53-327851f845c7 [ 819.929437] env[61923]: DEBUG nova.network.neutron [None req-a9ea97fa-71e5-4463-a1f3-ab69743d14a4 tempest-ServerActionsTestJSON-1295690556 tempest-ServerActionsTestJSON-1295690556-project-member] [instance: ad3b3933-5bd9-444a-af7b-c81e868e0622] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 820.113145] env[61923]: DEBUG oslo_concurrency.lockutils [None req-0adbd41a-bb08-4b5a-bc10-ab9778329dd8 tempest-VolumesAdminNegativeTest-1422120048 tempest-VolumesAdminNegativeTest-1422120048-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.858s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 820.114349] env[61923]: ERROR nova.compute.manager [None req-0adbd41a-bb08-4b5a-bc10-ab9778329dd8 tempest-VolumesAdminNegativeTest-1422120048 tempest-VolumesAdminNegativeTest-1422120048-project-member] [instance: 7bd3204b-5754-4ff2-b0ff-f30556dd72a1] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 9712b870-b8eb-4048-8ebf-0be8f9c5dfc2, please check neutron logs for more information. [ 820.114349] env[61923]: ERROR nova.compute.manager [instance: 7bd3204b-5754-4ff2-b0ff-f30556dd72a1] Traceback (most recent call last): [ 820.114349] env[61923]: ERROR nova.compute.manager [instance: 7bd3204b-5754-4ff2-b0ff-f30556dd72a1] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 820.114349] env[61923]: ERROR nova.compute.manager [instance: 7bd3204b-5754-4ff2-b0ff-f30556dd72a1] self.driver.spawn(context, instance, image_meta, [ 820.114349] env[61923]: ERROR nova.compute.manager [instance: 7bd3204b-5754-4ff2-b0ff-f30556dd72a1] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 820.114349] env[61923]: ERROR nova.compute.manager [instance: 7bd3204b-5754-4ff2-b0ff-f30556dd72a1] self._vmops.spawn(context, instance, image_meta, injected_files, [ 820.114349] env[61923]: ERROR nova.compute.manager [instance: 7bd3204b-5754-4ff2-b0ff-f30556dd72a1] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 820.114349] env[61923]: ERROR nova.compute.manager [instance: 7bd3204b-5754-4ff2-b0ff-f30556dd72a1] vm_ref = self.build_virtual_machine(instance, [ 820.114349] env[61923]: ERROR nova.compute.manager [instance: 7bd3204b-5754-4ff2-b0ff-f30556dd72a1] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 820.114349] env[61923]: ERROR nova.compute.manager [instance: 7bd3204b-5754-4ff2-b0ff-f30556dd72a1] vif_infos = vmwarevif.get_vif_info(self._session, [ 820.114349] env[61923]: ERROR nova.compute.manager [instance: 7bd3204b-5754-4ff2-b0ff-f30556dd72a1] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 820.114949] env[61923]: ERROR nova.compute.manager [instance: 7bd3204b-5754-4ff2-b0ff-f30556dd72a1] for vif in network_info: [ 820.114949] env[61923]: ERROR nova.compute.manager [instance: 7bd3204b-5754-4ff2-b0ff-f30556dd72a1] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 820.114949] env[61923]: ERROR nova.compute.manager [instance: 7bd3204b-5754-4ff2-b0ff-f30556dd72a1] return self._sync_wrapper(fn, *args, **kwargs) [ 820.114949] env[61923]: ERROR nova.compute.manager [instance: 7bd3204b-5754-4ff2-b0ff-f30556dd72a1] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 820.114949] env[61923]: ERROR nova.compute.manager [instance: 7bd3204b-5754-4ff2-b0ff-f30556dd72a1] self.wait() [ 820.114949] env[61923]: ERROR nova.compute.manager [instance: 7bd3204b-5754-4ff2-b0ff-f30556dd72a1] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 820.114949] env[61923]: ERROR nova.compute.manager [instance: 7bd3204b-5754-4ff2-b0ff-f30556dd72a1] self[:] = self._gt.wait() [ 820.114949] env[61923]: ERROR nova.compute.manager [instance: 7bd3204b-5754-4ff2-b0ff-f30556dd72a1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 820.114949] env[61923]: ERROR nova.compute.manager [instance: 7bd3204b-5754-4ff2-b0ff-f30556dd72a1] return self._exit_event.wait() [ 820.114949] env[61923]: ERROR nova.compute.manager [instance: 7bd3204b-5754-4ff2-b0ff-f30556dd72a1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 820.114949] env[61923]: ERROR nova.compute.manager [instance: 7bd3204b-5754-4ff2-b0ff-f30556dd72a1] current.throw(*self._exc) [ 820.114949] env[61923]: ERROR nova.compute.manager [instance: 7bd3204b-5754-4ff2-b0ff-f30556dd72a1] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 820.114949] env[61923]: ERROR nova.compute.manager [instance: 7bd3204b-5754-4ff2-b0ff-f30556dd72a1] result = function(*args, **kwargs) [ 820.115292] env[61923]: ERROR nova.compute.manager [instance: 7bd3204b-5754-4ff2-b0ff-f30556dd72a1] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 820.115292] env[61923]: ERROR nova.compute.manager [instance: 7bd3204b-5754-4ff2-b0ff-f30556dd72a1] return func(*args, **kwargs) [ 820.115292] env[61923]: ERROR nova.compute.manager [instance: 7bd3204b-5754-4ff2-b0ff-f30556dd72a1] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 820.115292] env[61923]: ERROR nova.compute.manager [instance: 7bd3204b-5754-4ff2-b0ff-f30556dd72a1] raise e [ 820.115292] env[61923]: ERROR nova.compute.manager [instance: 7bd3204b-5754-4ff2-b0ff-f30556dd72a1] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 820.115292] env[61923]: ERROR nova.compute.manager [instance: 7bd3204b-5754-4ff2-b0ff-f30556dd72a1] nwinfo = self.network_api.allocate_for_instance( [ 820.115292] env[61923]: ERROR nova.compute.manager [instance: 7bd3204b-5754-4ff2-b0ff-f30556dd72a1] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 820.115292] env[61923]: ERROR nova.compute.manager [instance: 7bd3204b-5754-4ff2-b0ff-f30556dd72a1] created_port_ids = self._update_ports_for_instance( [ 820.115292] env[61923]: ERROR nova.compute.manager [instance: 7bd3204b-5754-4ff2-b0ff-f30556dd72a1] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 820.115292] env[61923]: ERROR nova.compute.manager [instance: 7bd3204b-5754-4ff2-b0ff-f30556dd72a1] with excutils.save_and_reraise_exception(): [ 820.115292] env[61923]: ERROR nova.compute.manager [instance: 7bd3204b-5754-4ff2-b0ff-f30556dd72a1] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 820.115292] env[61923]: ERROR nova.compute.manager [instance: 7bd3204b-5754-4ff2-b0ff-f30556dd72a1] self.force_reraise() [ 820.115292] env[61923]: ERROR nova.compute.manager [instance: 7bd3204b-5754-4ff2-b0ff-f30556dd72a1] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 820.115603] env[61923]: ERROR nova.compute.manager [instance: 7bd3204b-5754-4ff2-b0ff-f30556dd72a1] raise self.value [ 820.115603] env[61923]: ERROR nova.compute.manager [instance: 7bd3204b-5754-4ff2-b0ff-f30556dd72a1] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 820.115603] env[61923]: ERROR nova.compute.manager [instance: 7bd3204b-5754-4ff2-b0ff-f30556dd72a1] updated_port = self._update_port( [ 820.115603] env[61923]: ERROR nova.compute.manager [instance: 7bd3204b-5754-4ff2-b0ff-f30556dd72a1] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 820.115603] env[61923]: ERROR nova.compute.manager [instance: 7bd3204b-5754-4ff2-b0ff-f30556dd72a1] _ensure_no_port_binding_failure(port) [ 820.115603] env[61923]: ERROR nova.compute.manager [instance: 7bd3204b-5754-4ff2-b0ff-f30556dd72a1] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 820.115603] env[61923]: ERROR nova.compute.manager [instance: 7bd3204b-5754-4ff2-b0ff-f30556dd72a1] raise exception.PortBindingFailed(port_id=port['id']) [ 820.115603] env[61923]: ERROR nova.compute.manager [instance: 7bd3204b-5754-4ff2-b0ff-f30556dd72a1] nova.exception.PortBindingFailed: Binding failed for port 9712b870-b8eb-4048-8ebf-0be8f9c5dfc2, please check neutron logs for more information. [ 820.115603] env[61923]: ERROR nova.compute.manager [instance: 7bd3204b-5754-4ff2-b0ff-f30556dd72a1] [ 820.115603] env[61923]: DEBUG nova.compute.utils [None req-0adbd41a-bb08-4b5a-bc10-ab9778329dd8 tempest-VolumesAdminNegativeTest-1422120048 tempest-VolumesAdminNegativeTest-1422120048-project-member] [instance: 7bd3204b-5754-4ff2-b0ff-f30556dd72a1] Binding failed for port 9712b870-b8eb-4048-8ebf-0be8f9c5dfc2, please check neutron logs for more information. {{(pid=61923) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 820.117590] env[61923]: DEBUG oslo_concurrency.lockutils [None req-28282894-8344-49f7-8696-a5bcafc57c69 tempest-ServerRescueTestJSONUnderV235-1784422495 tempest-ServerRescueTestJSONUnderV235-1784422495-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 17.250s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 820.119404] env[61923]: DEBUG nova.compute.manager [None req-0adbd41a-bb08-4b5a-bc10-ab9778329dd8 tempest-VolumesAdminNegativeTest-1422120048 tempest-VolumesAdminNegativeTest-1422120048-project-member] [instance: 7bd3204b-5754-4ff2-b0ff-f30556dd72a1] Build of instance 7bd3204b-5754-4ff2-b0ff-f30556dd72a1 was re-scheduled: Binding failed for port 9712b870-b8eb-4048-8ebf-0be8f9c5dfc2, please check neutron logs for more information. {{(pid=61923) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 820.120655] env[61923]: DEBUG nova.compute.manager [None req-0adbd41a-bb08-4b5a-bc10-ab9778329dd8 tempest-VolumesAdminNegativeTest-1422120048 tempest-VolumesAdminNegativeTest-1422120048-project-member] [instance: 7bd3204b-5754-4ff2-b0ff-f30556dd72a1] Unplugging VIFs for instance {{(pid=61923) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 820.120655] env[61923]: DEBUG oslo_concurrency.lockutils [None req-0adbd41a-bb08-4b5a-bc10-ab9778329dd8 tempest-VolumesAdminNegativeTest-1422120048 tempest-VolumesAdminNegativeTest-1422120048-project-member] Acquiring lock "refresh_cache-7bd3204b-5754-4ff2-b0ff-f30556dd72a1" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 820.120655] env[61923]: DEBUG oslo_concurrency.lockutils [None req-0adbd41a-bb08-4b5a-bc10-ab9778329dd8 tempest-VolumesAdminNegativeTest-1422120048 tempest-VolumesAdminNegativeTest-1422120048-project-member] Acquired lock "refresh_cache-7bd3204b-5754-4ff2-b0ff-f30556dd72a1" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 820.120655] env[61923]: DEBUG nova.network.neutron [None req-0adbd41a-bb08-4b5a-bc10-ab9778329dd8 tempest-VolumesAdminNegativeTest-1422120048 tempest-VolumesAdminNegativeTest-1422120048-project-member] [instance: 7bd3204b-5754-4ff2-b0ff-f30556dd72a1] Building network info cache for instance {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 820.216566] env[61923]: DEBUG oslo_concurrency.lockutils [None req-aed8d24a-43d8-4fbe-ac67-0dd6c37614c4 tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Lock "fe981195-1e2c-4c8d-8c53-327851f845c7" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 130.571s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 820.432627] env[61923]: INFO nova.compute.manager [None req-a9ea97fa-71e5-4463-a1f3-ab69743d14a4 tempest-ServerActionsTestJSON-1295690556 tempest-ServerActionsTestJSON-1295690556-project-member] [instance: ad3b3933-5bd9-444a-af7b-c81e868e0622] Took 1.04 seconds to deallocate network for instance. [ 820.646023] env[61923]: DEBUG nova.network.neutron [None req-0adbd41a-bb08-4b5a-bc10-ab9778329dd8 tempest-VolumesAdminNegativeTest-1422120048 tempest-VolumesAdminNegativeTest-1422120048-project-member] [instance: 7bd3204b-5754-4ff2-b0ff-f30556dd72a1] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 820.720941] env[61923]: DEBUG nova.compute.manager [None req-e62df7da-42b5-4263-bcf8-001a2d06427a tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: 73692517-1816-4e9b-ab2d-8265c683e83d] Starting instance... {{(pid=61923) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 820.810094] env[61923]: DEBUG nova.network.neutron [None req-0adbd41a-bb08-4b5a-bc10-ab9778329dd8 tempest-VolumesAdminNegativeTest-1422120048 tempest-VolumesAdminNegativeTest-1422120048-project-member] [instance: 7bd3204b-5754-4ff2-b0ff-f30556dd72a1] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 820.882251] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-657c1cd3-8c3a-43aa-9464-c16896c7323d {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.893649] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5a5ff98-bd05-4fd6-9876-b614f1235eab {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.934201] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa8922c7-9780-484b-befa-94c16ac3ecea {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.949174] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b963fb01-08b7-408b-8d47-42a28f2c664b {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.965929] env[61923]: DEBUG nova.compute.provider_tree [None req-28282894-8344-49f7-8696-a5bcafc57c69 tempest-ServerRescueTestJSONUnderV235-1784422495 tempest-ServerRescueTestJSONUnderV235-1784422495-project-member] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 821.248139] env[61923]: DEBUG oslo_concurrency.lockutils [None req-e62df7da-42b5-4263-bcf8-001a2d06427a tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 821.314419] env[61923]: DEBUG oslo_concurrency.lockutils [None req-0adbd41a-bb08-4b5a-bc10-ab9778329dd8 tempest-VolumesAdminNegativeTest-1422120048 tempest-VolumesAdminNegativeTest-1422120048-project-member] Releasing lock "refresh_cache-7bd3204b-5754-4ff2-b0ff-f30556dd72a1" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 821.314747] env[61923]: DEBUG nova.compute.manager [None req-0adbd41a-bb08-4b5a-bc10-ab9778329dd8 tempest-VolumesAdminNegativeTest-1422120048 tempest-VolumesAdminNegativeTest-1422120048-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61923) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 821.314883] env[61923]: DEBUG nova.compute.manager [None req-0adbd41a-bb08-4b5a-bc10-ab9778329dd8 tempest-VolumesAdminNegativeTest-1422120048 tempest-VolumesAdminNegativeTest-1422120048-project-member] [instance: 7bd3204b-5754-4ff2-b0ff-f30556dd72a1] Deallocating network for instance {{(pid=61923) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 821.315087] env[61923]: DEBUG nova.network.neutron [None req-0adbd41a-bb08-4b5a-bc10-ab9778329dd8 tempest-VolumesAdminNegativeTest-1422120048 tempest-VolumesAdminNegativeTest-1422120048-project-member] [instance: 7bd3204b-5754-4ff2-b0ff-f30556dd72a1] deallocate_for_instance() {{(pid=61923) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 821.337216] env[61923]: DEBUG nova.network.neutron [None req-0adbd41a-bb08-4b5a-bc10-ab9778329dd8 tempest-VolumesAdminNegativeTest-1422120048 tempest-VolumesAdminNegativeTest-1422120048-project-member] [instance: 7bd3204b-5754-4ff2-b0ff-f30556dd72a1] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 821.468566] env[61923]: INFO nova.scheduler.client.report [None req-a9ea97fa-71e5-4463-a1f3-ab69743d14a4 tempest-ServerActionsTestJSON-1295690556 tempest-ServerActionsTestJSON-1295690556-project-member] Deleted allocations for instance ad3b3933-5bd9-444a-af7b-c81e868e0622 [ 821.477301] env[61923]: DEBUG nova.scheduler.client.report [None req-28282894-8344-49f7-8696-a5bcafc57c69 tempest-ServerRescueTestJSONUnderV235-1784422495 tempest-ServerRescueTestJSONUnderV235-1784422495-project-member] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 821.555704] env[61923]: DEBUG oslo_concurrency.lockutils [None req-cdbe0145-cd3c-4551-af87-0c7f8b53f26c tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Acquiring lock "444affa5-a7ed-4a17-9015-9fd5724aab64" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 821.555973] env[61923]: DEBUG oslo_concurrency.lockutils [None req-cdbe0145-cd3c-4551-af87-0c7f8b53f26c tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Lock "444affa5-a7ed-4a17-9015-9fd5724aab64" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 821.583694] env[61923]: DEBUG oslo_concurrency.lockutils [None req-cdbe0145-cd3c-4551-af87-0c7f8b53f26c tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Acquiring lock "e413c45d-cd89-44d4-9102-3d188907e7bb" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 821.584055] env[61923]: DEBUG oslo_concurrency.lockutils [None req-cdbe0145-cd3c-4551-af87-0c7f8b53f26c tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Lock "e413c45d-cd89-44d4-9102-3d188907e7bb" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 821.840244] env[61923]: DEBUG nova.network.neutron [None req-0adbd41a-bb08-4b5a-bc10-ab9778329dd8 tempest-VolumesAdminNegativeTest-1422120048 tempest-VolumesAdminNegativeTest-1422120048-project-member] [instance: 7bd3204b-5754-4ff2-b0ff-f30556dd72a1] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 821.984917] env[61923]: DEBUG oslo_concurrency.lockutils [None req-28282894-8344-49f7-8696-a5bcafc57c69 tempest-ServerRescueTestJSONUnderV235-1784422495 tempest-ServerRescueTestJSONUnderV235-1784422495-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.869s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 821.985584] env[61923]: ERROR nova.compute.manager [None req-28282894-8344-49f7-8696-a5bcafc57c69 tempest-ServerRescueTestJSONUnderV235-1784422495 tempest-ServerRescueTestJSONUnderV235-1784422495-project-member] [instance: b2d2b34d-827f-4f96-af8e-936135ba8700] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 85462a2f-eba7-481a-8f13-c572e92d6634, please check neutron logs for more information. [ 821.985584] env[61923]: ERROR nova.compute.manager [instance: b2d2b34d-827f-4f96-af8e-936135ba8700] Traceback (most recent call last): [ 821.985584] env[61923]: ERROR nova.compute.manager [instance: b2d2b34d-827f-4f96-af8e-936135ba8700] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 821.985584] env[61923]: ERROR nova.compute.manager [instance: b2d2b34d-827f-4f96-af8e-936135ba8700] self.driver.spawn(context, instance, image_meta, [ 821.985584] env[61923]: ERROR nova.compute.manager [instance: b2d2b34d-827f-4f96-af8e-936135ba8700] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 821.985584] env[61923]: ERROR nova.compute.manager [instance: b2d2b34d-827f-4f96-af8e-936135ba8700] self._vmops.spawn(context, instance, image_meta, injected_files, [ 821.985584] env[61923]: ERROR nova.compute.manager [instance: b2d2b34d-827f-4f96-af8e-936135ba8700] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 821.985584] env[61923]: ERROR nova.compute.manager [instance: b2d2b34d-827f-4f96-af8e-936135ba8700] vm_ref = self.build_virtual_machine(instance, [ 821.985584] env[61923]: ERROR nova.compute.manager [instance: b2d2b34d-827f-4f96-af8e-936135ba8700] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 821.985584] env[61923]: ERROR nova.compute.manager [instance: b2d2b34d-827f-4f96-af8e-936135ba8700] vif_infos = vmwarevif.get_vif_info(self._session, [ 821.985584] env[61923]: ERROR nova.compute.manager [instance: b2d2b34d-827f-4f96-af8e-936135ba8700] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 821.985966] env[61923]: ERROR nova.compute.manager [instance: b2d2b34d-827f-4f96-af8e-936135ba8700] for vif in network_info: [ 821.985966] env[61923]: ERROR nova.compute.manager [instance: b2d2b34d-827f-4f96-af8e-936135ba8700] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 821.985966] env[61923]: ERROR nova.compute.manager [instance: b2d2b34d-827f-4f96-af8e-936135ba8700] return self._sync_wrapper(fn, *args, **kwargs) [ 821.985966] env[61923]: ERROR nova.compute.manager [instance: b2d2b34d-827f-4f96-af8e-936135ba8700] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 821.985966] env[61923]: ERROR nova.compute.manager [instance: b2d2b34d-827f-4f96-af8e-936135ba8700] self.wait() [ 821.985966] env[61923]: ERROR nova.compute.manager [instance: b2d2b34d-827f-4f96-af8e-936135ba8700] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 821.985966] env[61923]: ERROR nova.compute.manager [instance: b2d2b34d-827f-4f96-af8e-936135ba8700] self[:] = self._gt.wait() [ 821.985966] env[61923]: ERROR nova.compute.manager [instance: b2d2b34d-827f-4f96-af8e-936135ba8700] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 821.985966] env[61923]: ERROR nova.compute.manager [instance: b2d2b34d-827f-4f96-af8e-936135ba8700] return self._exit_event.wait() [ 821.985966] env[61923]: ERROR nova.compute.manager [instance: b2d2b34d-827f-4f96-af8e-936135ba8700] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 821.985966] env[61923]: ERROR nova.compute.manager [instance: b2d2b34d-827f-4f96-af8e-936135ba8700] current.throw(*self._exc) [ 821.985966] env[61923]: ERROR nova.compute.manager [instance: b2d2b34d-827f-4f96-af8e-936135ba8700] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 821.985966] env[61923]: ERROR nova.compute.manager [instance: b2d2b34d-827f-4f96-af8e-936135ba8700] result = function(*args, **kwargs) [ 821.986342] env[61923]: ERROR nova.compute.manager [instance: b2d2b34d-827f-4f96-af8e-936135ba8700] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 821.986342] env[61923]: ERROR nova.compute.manager [instance: b2d2b34d-827f-4f96-af8e-936135ba8700] return func(*args, **kwargs) [ 821.986342] env[61923]: ERROR nova.compute.manager [instance: b2d2b34d-827f-4f96-af8e-936135ba8700] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 821.986342] env[61923]: ERROR nova.compute.manager [instance: b2d2b34d-827f-4f96-af8e-936135ba8700] raise e [ 821.986342] env[61923]: ERROR nova.compute.manager [instance: b2d2b34d-827f-4f96-af8e-936135ba8700] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 821.986342] env[61923]: ERROR nova.compute.manager [instance: b2d2b34d-827f-4f96-af8e-936135ba8700] nwinfo = self.network_api.allocate_for_instance( [ 821.986342] env[61923]: ERROR nova.compute.manager [instance: b2d2b34d-827f-4f96-af8e-936135ba8700] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 821.986342] env[61923]: ERROR nova.compute.manager [instance: b2d2b34d-827f-4f96-af8e-936135ba8700] created_port_ids = self._update_ports_for_instance( [ 821.986342] env[61923]: ERROR nova.compute.manager [instance: b2d2b34d-827f-4f96-af8e-936135ba8700] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 821.986342] env[61923]: ERROR nova.compute.manager [instance: b2d2b34d-827f-4f96-af8e-936135ba8700] with excutils.save_and_reraise_exception(): [ 821.986342] env[61923]: ERROR nova.compute.manager [instance: b2d2b34d-827f-4f96-af8e-936135ba8700] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 821.986342] env[61923]: ERROR nova.compute.manager [instance: b2d2b34d-827f-4f96-af8e-936135ba8700] self.force_reraise() [ 821.986342] env[61923]: ERROR nova.compute.manager [instance: b2d2b34d-827f-4f96-af8e-936135ba8700] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 821.986694] env[61923]: ERROR nova.compute.manager [instance: b2d2b34d-827f-4f96-af8e-936135ba8700] raise self.value [ 821.986694] env[61923]: ERROR nova.compute.manager [instance: b2d2b34d-827f-4f96-af8e-936135ba8700] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 821.986694] env[61923]: ERROR nova.compute.manager [instance: b2d2b34d-827f-4f96-af8e-936135ba8700] updated_port = self._update_port( [ 821.986694] env[61923]: ERROR nova.compute.manager [instance: b2d2b34d-827f-4f96-af8e-936135ba8700] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 821.986694] env[61923]: ERROR nova.compute.manager [instance: b2d2b34d-827f-4f96-af8e-936135ba8700] _ensure_no_port_binding_failure(port) [ 821.986694] env[61923]: ERROR nova.compute.manager [instance: b2d2b34d-827f-4f96-af8e-936135ba8700] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 821.986694] env[61923]: ERROR nova.compute.manager [instance: b2d2b34d-827f-4f96-af8e-936135ba8700] raise exception.PortBindingFailed(port_id=port['id']) [ 821.986694] env[61923]: ERROR nova.compute.manager [instance: b2d2b34d-827f-4f96-af8e-936135ba8700] nova.exception.PortBindingFailed: Binding failed for port 85462a2f-eba7-481a-8f13-c572e92d6634, please check neutron logs for more information. [ 821.986694] env[61923]: ERROR nova.compute.manager [instance: b2d2b34d-827f-4f96-af8e-936135ba8700] [ 821.986694] env[61923]: DEBUG nova.compute.utils [None req-28282894-8344-49f7-8696-a5bcafc57c69 tempest-ServerRescueTestJSONUnderV235-1784422495 tempest-ServerRescueTestJSONUnderV235-1784422495-project-member] [instance: b2d2b34d-827f-4f96-af8e-936135ba8700] Binding failed for port 85462a2f-eba7-481a-8f13-c572e92d6634, please check neutron logs for more information. {{(pid=61923) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 821.987967] env[61923]: DEBUG oslo_concurrency.lockutils [None req-a9ea97fa-71e5-4463-a1f3-ab69743d14a4 tempest-ServerActionsTestJSON-1295690556 tempest-ServerActionsTestJSON-1295690556-project-member] Lock "ad3b3933-5bd9-444a-af7b-c81e868e0622" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 125.150s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 821.988195] env[61923]: DEBUG oslo_concurrency.lockutils [None req-13baf863-98e6-41c7-ae75-48fd1a9cef32 tempest-ServerAddressesTestJSON-1846886542 tempest-ServerAddressesTestJSON-1846886542-project-member] Acquiring lock "5b73e07a-d6ef-4dcf-bdae-eea91d2aeb6f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 821.988432] env[61923]: DEBUG oslo_concurrency.lockutils [None req-13baf863-98e6-41c7-ae75-48fd1a9cef32 tempest-ServerAddressesTestJSON-1846886542 tempest-ServerAddressesTestJSON-1846886542-project-member] Lock "5b73e07a-d6ef-4dcf-bdae-eea91d2aeb6f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 821.988877] env[61923]: DEBUG nova.compute.manager [None req-28282894-8344-49f7-8696-a5bcafc57c69 tempest-ServerRescueTestJSONUnderV235-1784422495 tempest-ServerRescueTestJSONUnderV235-1784422495-project-member] [instance: b2d2b34d-827f-4f96-af8e-936135ba8700] Build of instance b2d2b34d-827f-4f96-af8e-936135ba8700 was re-scheduled: Binding failed for port 85462a2f-eba7-481a-8f13-c572e92d6634, please check neutron logs for more information. {{(pid=61923) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 821.989364] env[61923]: DEBUG nova.compute.manager [None req-28282894-8344-49f7-8696-a5bcafc57c69 tempest-ServerRescueTestJSONUnderV235-1784422495 tempest-ServerRescueTestJSONUnderV235-1784422495-project-member] [instance: b2d2b34d-827f-4f96-af8e-936135ba8700] Unplugging VIFs for instance {{(pid=61923) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 821.989610] env[61923]: DEBUG oslo_concurrency.lockutils [None req-28282894-8344-49f7-8696-a5bcafc57c69 tempest-ServerRescueTestJSONUnderV235-1784422495 tempest-ServerRescueTestJSONUnderV235-1784422495-project-member] Acquiring lock "refresh_cache-b2d2b34d-827f-4f96-af8e-936135ba8700" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 821.989778] env[61923]: DEBUG oslo_concurrency.lockutils [None req-28282894-8344-49f7-8696-a5bcafc57c69 tempest-ServerRescueTestJSONUnderV235-1784422495 tempest-ServerRescueTestJSONUnderV235-1784422495-project-member] Acquired lock "refresh_cache-b2d2b34d-827f-4f96-af8e-936135ba8700" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 821.989952] env[61923]: DEBUG nova.network.neutron [None req-28282894-8344-49f7-8696-a5bcafc57c69 tempest-ServerRescueTestJSONUnderV235-1784422495 tempest-ServerRescueTestJSONUnderV235-1784422495-project-member] [instance: b2d2b34d-827f-4f96-af8e-936135ba8700] Building network info cache for instance {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 821.991026] env[61923]: DEBUG oslo_concurrency.lockutils [None req-998fcb75-b821-4586-ae48-2fe4d02a9864 tempest-AttachVolumeShelveTestJSON-1252849694 tempest-AttachVolumeShelveTestJSON-1252849694-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.727s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 822.343420] env[61923]: INFO nova.compute.manager [None req-0adbd41a-bb08-4b5a-bc10-ab9778329dd8 tempest-VolumesAdminNegativeTest-1422120048 tempest-VolumesAdminNegativeTest-1422120048-project-member] [instance: 7bd3204b-5754-4ff2-b0ff-f30556dd72a1] Took 1.03 seconds to deallocate network for instance. [ 822.500463] env[61923]: DEBUG nova.compute.manager [None req-513daa64-2b05-4c0d-94e9-55b95de63eec tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] [instance: 59198f4d-4dde-4eaf-9f6c-a962cbe53c6e] Starting instance... {{(pid=61923) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 822.525745] env[61923]: DEBUG nova.network.neutron [None req-28282894-8344-49f7-8696-a5bcafc57c69 tempest-ServerRescueTestJSONUnderV235-1784422495 tempest-ServerRescueTestJSONUnderV235-1784422495-project-member] [instance: b2d2b34d-827f-4f96-af8e-936135ba8700] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 822.639764] env[61923]: DEBUG nova.network.neutron [None req-28282894-8344-49f7-8696-a5bcafc57c69 tempest-ServerRescueTestJSONUnderV235-1784422495 tempest-ServerRescueTestJSONUnderV235-1784422495-project-member] [instance: b2d2b34d-827f-4f96-af8e-936135ba8700] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 822.776080] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3cc9b86-79e1-4410-80f4-1b25ea2ee4bb {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.783751] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e18e2a1-59bd-48aa-bf97-c44e9add0811 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.821572] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6861a374-9a98-4057-a00f-3b7566221e20 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.830356] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d5939e2-77e9-418b-aa7a-d3eb6653f3eb {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.848154] env[61923]: DEBUG nova.compute.provider_tree [None req-998fcb75-b821-4586-ae48-2fe4d02a9864 tempest-AttachVolumeShelveTestJSON-1252849694 tempest-AttachVolumeShelveTestJSON-1252849694-project-member] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 823.025776] env[61923]: DEBUG oslo_concurrency.lockutils [None req-513daa64-2b05-4c0d-94e9-55b95de63eec tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 823.147922] env[61923]: DEBUG oslo_concurrency.lockutils [None req-28282894-8344-49f7-8696-a5bcafc57c69 tempest-ServerRescueTestJSONUnderV235-1784422495 tempest-ServerRescueTestJSONUnderV235-1784422495-project-member] Releasing lock "refresh_cache-b2d2b34d-827f-4f96-af8e-936135ba8700" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 823.147922] env[61923]: DEBUG nova.compute.manager [None req-28282894-8344-49f7-8696-a5bcafc57c69 tempest-ServerRescueTestJSONUnderV235-1784422495 tempest-ServerRescueTestJSONUnderV235-1784422495-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61923) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 823.148048] env[61923]: DEBUG nova.compute.manager [None req-28282894-8344-49f7-8696-a5bcafc57c69 tempest-ServerRescueTestJSONUnderV235-1784422495 tempest-ServerRescueTestJSONUnderV235-1784422495-project-member] [instance: b2d2b34d-827f-4f96-af8e-936135ba8700] Deallocating network for instance {{(pid=61923) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 823.148203] env[61923]: DEBUG nova.network.neutron [None req-28282894-8344-49f7-8696-a5bcafc57c69 tempest-ServerRescueTestJSONUnderV235-1784422495 tempest-ServerRescueTestJSONUnderV235-1784422495-project-member] [instance: b2d2b34d-827f-4f96-af8e-936135ba8700] deallocate_for_instance() {{(pid=61923) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 823.163803] env[61923]: DEBUG nova.network.neutron [None req-28282894-8344-49f7-8696-a5bcafc57c69 tempest-ServerRescueTestJSONUnderV235-1784422495 tempest-ServerRescueTestJSONUnderV235-1784422495-project-member] [instance: b2d2b34d-827f-4f96-af8e-936135ba8700] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 823.352836] env[61923]: DEBUG nova.scheduler.client.report [None req-998fcb75-b821-4586-ae48-2fe4d02a9864 tempest-AttachVolumeShelveTestJSON-1252849694 tempest-AttachVolumeShelveTestJSON-1252849694-project-member] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 823.386427] env[61923]: INFO nova.scheduler.client.report [None req-0adbd41a-bb08-4b5a-bc10-ab9778329dd8 tempest-VolumesAdminNegativeTest-1422120048 tempest-VolumesAdminNegativeTest-1422120048-project-member] Deleted allocations for instance 7bd3204b-5754-4ff2-b0ff-f30556dd72a1 [ 823.667380] env[61923]: DEBUG nova.network.neutron [None req-28282894-8344-49f7-8696-a5bcafc57c69 tempest-ServerRescueTestJSONUnderV235-1784422495 tempest-ServerRescueTestJSONUnderV235-1784422495-project-member] [instance: b2d2b34d-827f-4f96-af8e-936135ba8700] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 823.861021] env[61923]: DEBUG oslo_concurrency.lockutils [None req-998fcb75-b821-4586-ae48-2fe4d02a9864 tempest-AttachVolumeShelveTestJSON-1252849694 tempest-AttachVolumeShelveTestJSON-1252849694-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.870s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 823.861923] env[61923]: ERROR nova.compute.manager [None req-998fcb75-b821-4586-ae48-2fe4d02a9864 tempest-AttachVolumeShelveTestJSON-1252849694 tempest-AttachVolumeShelveTestJSON-1252849694-project-member] [instance: 870602dd-1b85-4c97-9506-f08e9f1bade6] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port eb1ea65c-43f8-4f04-ae2d-6429b469985d, please check neutron logs for more information. [ 823.861923] env[61923]: ERROR nova.compute.manager [instance: 870602dd-1b85-4c97-9506-f08e9f1bade6] Traceback (most recent call last): [ 823.861923] env[61923]: ERROR nova.compute.manager [instance: 870602dd-1b85-4c97-9506-f08e9f1bade6] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 823.861923] env[61923]: ERROR nova.compute.manager [instance: 870602dd-1b85-4c97-9506-f08e9f1bade6] self.driver.spawn(context, instance, image_meta, [ 823.861923] env[61923]: ERROR nova.compute.manager [instance: 870602dd-1b85-4c97-9506-f08e9f1bade6] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 823.861923] env[61923]: ERROR nova.compute.manager [instance: 870602dd-1b85-4c97-9506-f08e9f1bade6] self._vmops.spawn(context, instance, image_meta, injected_files, [ 823.861923] env[61923]: ERROR nova.compute.manager [instance: 870602dd-1b85-4c97-9506-f08e9f1bade6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 823.861923] env[61923]: ERROR nova.compute.manager [instance: 870602dd-1b85-4c97-9506-f08e9f1bade6] vm_ref = self.build_virtual_machine(instance, [ 823.861923] env[61923]: ERROR nova.compute.manager [instance: 870602dd-1b85-4c97-9506-f08e9f1bade6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 823.861923] env[61923]: ERROR nova.compute.manager [instance: 870602dd-1b85-4c97-9506-f08e9f1bade6] vif_infos = vmwarevif.get_vif_info(self._session, [ 823.861923] env[61923]: ERROR nova.compute.manager [instance: 870602dd-1b85-4c97-9506-f08e9f1bade6] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 823.862338] env[61923]: ERROR nova.compute.manager [instance: 870602dd-1b85-4c97-9506-f08e9f1bade6] for vif in network_info: [ 823.862338] env[61923]: ERROR nova.compute.manager [instance: 870602dd-1b85-4c97-9506-f08e9f1bade6] File "/opt/stack/nova/nova/network/model.py", line 614, in __iter__ [ 823.862338] env[61923]: ERROR nova.compute.manager [instance: 870602dd-1b85-4c97-9506-f08e9f1bade6] return self._sync_wrapper(fn, *args, **kwargs) [ 823.862338] env[61923]: ERROR nova.compute.manager [instance: 870602dd-1b85-4c97-9506-f08e9f1bade6] File "/opt/stack/nova/nova/network/model.py", line 605, in _sync_wrapper [ 823.862338] env[61923]: ERROR nova.compute.manager [instance: 870602dd-1b85-4c97-9506-f08e9f1bade6] self.wait() [ 823.862338] env[61923]: ERROR nova.compute.manager [instance: 870602dd-1b85-4c97-9506-f08e9f1bade6] File "/opt/stack/nova/nova/network/model.py", line 637, in wait [ 823.862338] env[61923]: ERROR nova.compute.manager [instance: 870602dd-1b85-4c97-9506-f08e9f1bade6] self[:] = self._gt.wait() [ 823.862338] env[61923]: ERROR nova.compute.manager [instance: 870602dd-1b85-4c97-9506-f08e9f1bade6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 823.862338] env[61923]: ERROR nova.compute.manager [instance: 870602dd-1b85-4c97-9506-f08e9f1bade6] return self._exit_event.wait() [ 823.862338] env[61923]: ERROR nova.compute.manager [instance: 870602dd-1b85-4c97-9506-f08e9f1bade6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 823.862338] env[61923]: ERROR nova.compute.manager [instance: 870602dd-1b85-4c97-9506-f08e9f1bade6] result = hub.switch() [ 823.862338] env[61923]: ERROR nova.compute.manager [instance: 870602dd-1b85-4c97-9506-f08e9f1bade6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 823.862338] env[61923]: ERROR nova.compute.manager [instance: 870602dd-1b85-4c97-9506-f08e9f1bade6] return self.greenlet.switch() [ 823.862717] env[61923]: ERROR nova.compute.manager [instance: 870602dd-1b85-4c97-9506-f08e9f1bade6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 823.862717] env[61923]: ERROR nova.compute.manager [instance: 870602dd-1b85-4c97-9506-f08e9f1bade6] result = function(*args, **kwargs) [ 823.862717] env[61923]: ERROR nova.compute.manager [instance: 870602dd-1b85-4c97-9506-f08e9f1bade6] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 823.862717] env[61923]: ERROR nova.compute.manager [instance: 870602dd-1b85-4c97-9506-f08e9f1bade6] return func(*args, **kwargs) [ 823.862717] env[61923]: ERROR nova.compute.manager [instance: 870602dd-1b85-4c97-9506-f08e9f1bade6] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 823.862717] env[61923]: ERROR nova.compute.manager [instance: 870602dd-1b85-4c97-9506-f08e9f1bade6] raise e [ 823.862717] env[61923]: ERROR nova.compute.manager [instance: 870602dd-1b85-4c97-9506-f08e9f1bade6] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 823.862717] env[61923]: ERROR nova.compute.manager [instance: 870602dd-1b85-4c97-9506-f08e9f1bade6] nwinfo = self.network_api.allocate_for_instance( [ 823.862717] env[61923]: ERROR nova.compute.manager [instance: 870602dd-1b85-4c97-9506-f08e9f1bade6] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 823.862717] env[61923]: ERROR nova.compute.manager [instance: 870602dd-1b85-4c97-9506-f08e9f1bade6] created_port_ids = self._update_ports_for_instance( [ 823.862717] env[61923]: ERROR nova.compute.manager [instance: 870602dd-1b85-4c97-9506-f08e9f1bade6] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 823.862717] env[61923]: ERROR nova.compute.manager [instance: 870602dd-1b85-4c97-9506-f08e9f1bade6] with excutils.save_and_reraise_exception(): [ 823.862717] env[61923]: ERROR nova.compute.manager [instance: 870602dd-1b85-4c97-9506-f08e9f1bade6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 823.863065] env[61923]: ERROR nova.compute.manager [instance: 870602dd-1b85-4c97-9506-f08e9f1bade6] self.force_reraise() [ 823.863065] env[61923]: ERROR nova.compute.manager [instance: 870602dd-1b85-4c97-9506-f08e9f1bade6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 823.863065] env[61923]: ERROR nova.compute.manager [instance: 870602dd-1b85-4c97-9506-f08e9f1bade6] raise self.value [ 823.863065] env[61923]: ERROR nova.compute.manager [instance: 870602dd-1b85-4c97-9506-f08e9f1bade6] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 823.863065] env[61923]: ERROR nova.compute.manager [instance: 870602dd-1b85-4c97-9506-f08e9f1bade6] updated_port = self._update_port( [ 823.863065] env[61923]: ERROR nova.compute.manager [instance: 870602dd-1b85-4c97-9506-f08e9f1bade6] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 823.863065] env[61923]: ERROR nova.compute.manager [instance: 870602dd-1b85-4c97-9506-f08e9f1bade6] _ensure_no_port_binding_failure(port) [ 823.863065] env[61923]: ERROR nova.compute.manager [instance: 870602dd-1b85-4c97-9506-f08e9f1bade6] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 823.863065] env[61923]: ERROR nova.compute.manager [instance: 870602dd-1b85-4c97-9506-f08e9f1bade6] raise exception.PortBindingFailed(port_id=port['id']) [ 823.863065] env[61923]: ERROR nova.compute.manager [instance: 870602dd-1b85-4c97-9506-f08e9f1bade6] nova.exception.PortBindingFailed: Binding failed for port eb1ea65c-43f8-4f04-ae2d-6429b469985d, please check neutron logs for more information. [ 823.863065] env[61923]: ERROR nova.compute.manager [instance: 870602dd-1b85-4c97-9506-f08e9f1bade6] [ 823.863337] env[61923]: DEBUG nova.compute.utils [None req-998fcb75-b821-4586-ae48-2fe4d02a9864 tempest-AttachVolumeShelveTestJSON-1252849694 tempest-AttachVolumeShelveTestJSON-1252849694-project-member] [instance: 870602dd-1b85-4c97-9506-f08e9f1bade6] Binding failed for port eb1ea65c-43f8-4f04-ae2d-6429b469985d, please check neutron logs for more information. {{(pid=61923) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 823.863837] env[61923]: DEBUG oslo_concurrency.lockutils [None req-c06a566a-3828-47dc-91c4-25d15d717c22 tempest-ServerShowV257Test-986496257 tempest-ServerShowV257Test-986496257-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 11.673s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 823.865296] env[61923]: INFO nova.compute.claims [None req-c06a566a-3828-47dc-91c4-25d15d717c22 tempest-ServerShowV257Test-986496257 tempest-ServerShowV257Test-986496257-project-member] [instance: b8c60af7-3561-468c-a9a5-1c645080fc69] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 823.868086] env[61923]: DEBUG nova.compute.manager [None req-998fcb75-b821-4586-ae48-2fe4d02a9864 tempest-AttachVolumeShelveTestJSON-1252849694 tempest-AttachVolumeShelveTestJSON-1252849694-project-member] [instance: 870602dd-1b85-4c97-9506-f08e9f1bade6] Build of instance 870602dd-1b85-4c97-9506-f08e9f1bade6 was re-scheduled: Binding failed for port eb1ea65c-43f8-4f04-ae2d-6429b469985d, please check neutron logs for more information. {{(pid=61923) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 823.869276] env[61923]: DEBUG nova.compute.manager [None req-998fcb75-b821-4586-ae48-2fe4d02a9864 tempest-AttachVolumeShelveTestJSON-1252849694 tempest-AttachVolumeShelveTestJSON-1252849694-project-member] [instance: 870602dd-1b85-4c97-9506-f08e9f1bade6] Unplugging VIFs for instance {{(pid=61923) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 823.869276] env[61923]: DEBUG oslo_concurrency.lockutils [None req-998fcb75-b821-4586-ae48-2fe4d02a9864 tempest-AttachVolumeShelveTestJSON-1252849694 tempest-AttachVolumeShelveTestJSON-1252849694-project-member] Acquiring lock "refresh_cache-870602dd-1b85-4c97-9506-f08e9f1bade6" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 823.869276] env[61923]: DEBUG oslo_concurrency.lockutils [None req-998fcb75-b821-4586-ae48-2fe4d02a9864 tempest-AttachVolumeShelveTestJSON-1252849694 tempest-AttachVolumeShelveTestJSON-1252849694-project-member] Acquired lock "refresh_cache-870602dd-1b85-4c97-9506-f08e9f1bade6" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 823.869276] env[61923]: DEBUG nova.network.neutron [None req-998fcb75-b821-4586-ae48-2fe4d02a9864 tempest-AttachVolumeShelveTestJSON-1252849694 tempest-AttachVolumeShelveTestJSON-1252849694-project-member] [instance: 870602dd-1b85-4c97-9506-f08e9f1bade6] Building network info cache for instance {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 823.893835] env[61923]: DEBUG oslo_concurrency.lockutils [None req-0adbd41a-bb08-4b5a-bc10-ab9778329dd8 tempest-VolumesAdminNegativeTest-1422120048 tempest-VolumesAdminNegativeTest-1422120048-project-member] Lock "7bd3204b-5754-4ff2-b0ff-f30556dd72a1" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 109.422s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 824.170986] env[61923]: INFO nova.compute.manager [None req-28282894-8344-49f7-8696-a5bcafc57c69 tempest-ServerRescueTestJSONUnderV235-1784422495 tempest-ServerRescueTestJSONUnderV235-1784422495-project-member] [instance: b2d2b34d-827f-4f96-af8e-936135ba8700] Took 1.02 seconds to deallocate network for instance. [ 824.392809] env[61923]: DEBUG nova.network.neutron [None req-998fcb75-b821-4586-ae48-2fe4d02a9864 tempest-AttachVolumeShelveTestJSON-1252849694 tempest-AttachVolumeShelveTestJSON-1252849694-project-member] [instance: 870602dd-1b85-4c97-9506-f08e9f1bade6] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 824.396363] env[61923]: DEBUG nova.compute.manager [None req-80740a2c-9779-4afc-be30-c4469cd3724d tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] [instance: fc4fdf89-4615-4586-9dcd-ea90dc258361] Starting instance... {{(pid=61923) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 824.509750] env[61923]: DEBUG nova.network.neutron [None req-998fcb75-b821-4586-ae48-2fe4d02a9864 tempest-AttachVolumeShelveTestJSON-1252849694 tempest-AttachVolumeShelveTestJSON-1252849694-project-member] [instance: 870602dd-1b85-4c97-9506-f08e9f1bade6] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 824.915125] env[61923]: DEBUG oslo_concurrency.lockutils [None req-80740a2c-9779-4afc-be30-c4469cd3724d tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 825.011810] env[61923]: DEBUG oslo_concurrency.lockutils [None req-998fcb75-b821-4586-ae48-2fe4d02a9864 tempest-AttachVolumeShelveTestJSON-1252849694 tempest-AttachVolumeShelveTestJSON-1252849694-project-member] Releasing lock "refresh_cache-870602dd-1b85-4c97-9506-f08e9f1bade6" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 825.012065] env[61923]: DEBUG nova.compute.manager [None req-998fcb75-b821-4586-ae48-2fe4d02a9864 tempest-AttachVolumeShelveTestJSON-1252849694 tempest-AttachVolumeShelveTestJSON-1252849694-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=61923) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 825.012256] env[61923]: DEBUG nova.compute.manager [None req-998fcb75-b821-4586-ae48-2fe4d02a9864 tempest-AttachVolumeShelveTestJSON-1252849694 tempest-AttachVolumeShelveTestJSON-1252849694-project-member] [instance: 870602dd-1b85-4c97-9506-f08e9f1bade6] Deallocating network for instance {{(pid=61923) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 825.012422] env[61923]: DEBUG nova.network.neutron [None req-998fcb75-b821-4586-ae48-2fe4d02a9864 tempest-AttachVolumeShelveTestJSON-1252849694 tempest-AttachVolumeShelveTestJSON-1252849694-project-member] [instance: 870602dd-1b85-4c97-9506-f08e9f1bade6] deallocate_for_instance() {{(pid=61923) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 825.027638] env[61923]: DEBUG nova.network.neutron [None req-998fcb75-b821-4586-ae48-2fe4d02a9864 tempest-AttachVolumeShelveTestJSON-1252849694 tempest-AttachVolumeShelveTestJSON-1252849694-project-member] [instance: 870602dd-1b85-4c97-9506-f08e9f1bade6] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 825.064328] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f82d5e10-2870-4d32-9bb8-5b7716463ca4 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.075042] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-240c323a-f9c4-469d-8145-0f3a917dbe75 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.106252] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a3930ad-eb52-41cb-b363-ec3cfcd4fe67 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.113806] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a26f3e8-4abc-42c7-8325-0fd7ed20286f {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.131332] env[61923]: DEBUG nova.compute.provider_tree [None req-c06a566a-3828-47dc-91c4-25d15d717c22 tempest-ServerShowV257Test-986496257 tempest-ServerShowV257Test-986496257-project-member] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 825.198976] env[61923]: INFO nova.scheduler.client.report [None req-28282894-8344-49f7-8696-a5bcafc57c69 tempest-ServerRescueTestJSONUnderV235-1784422495 tempest-ServerRescueTestJSONUnderV235-1784422495-project-member] Deleted allocations for instance b2d2b34d-827f-4f96-af8e-936135ba8700 [ 825.533600] env[61923]: DEBUG nova.network.neutron [None req-998fcb75-b821-4586-ae48-2fe4d02a9864 tempest-AttachVolumeShelveTestJSON-1252849694 tempest-AttachVolumeShelveTestJSON-1252849694-project-member] [instance: 870602dd-1b85-4c97-9506-f08e9f1bade6] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 825.635462] env[61923]: DEBUG nova.scheduler.client.report [None req-c06a566a-3828-47dc-91c4-25d15d717c22 tempest-ServerShowV257Test-986496257 tempest-ServerShowV257Test-986496257-project-member] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 825.710937] env[61923]: DEBUG oslo_concurrency.lockutils [None req-28282894-8344-49f7-8696-a5bcafc57c69 tempest-ServerRescueTestJSONUnderV235-1784422495 tempest-ServerRescueTestJSONUnderV235-1784422495-project-member] Lock "b2d2b34d-827f-4f96-af8e-936135ba8700" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 108.939s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 826.038819] env[61923]: INFO nova.compute.manager [None req-998fcb75-b821-4586-ae48-2fe4d02a9864 tempest-AttachVolumeShelveTestJSON-1252849694 tempest-AttachVolumeShelveTestJSON-1252849694-project-member] [instance: 870602dd-1b85-4c97-9506-f08e9f1bade6] Took 1.03 seconds to deallocate network for instance. [ 826.140509] env[61923]: DEBUG oslo_concurrency.lockutils [None req-c06a566a-3828-47dc-91c4-25d15d717c22 tempest-ServerShowV257Test-986496257 tempest-ServerShowV257Test-986496257-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.277s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 826.141049] env[61923]: DEBUG nova.compute.manager [None req-c06a566a-3828-47dc-91c4-25d15d717c22 tempest-ServerShowV257Test-986496257 tempest-ServerShowV257Test-986496257-project-member] [instance: b8c60af7-3561-468c-a9a5-1c645080fc69] Start building networks asynchronously for instance. {{(pid=61923) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 826.144311] env[61923]: DEBUG oslo_concurrency.lockutils [None req-8e8efa20-a2ed-4da9-89f1-5ff799c0382c tempest-ServersNegativeTestMultiTenantJSON-1497120864 tempest-ServersNegativeTestMultiTenantJSON-1497120864-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 12.246s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 826.149595] env[61923]: INFO nova.compute.claims [None req-8e8efa20-a2ed-4da9-89f1-5ff799c0382c tempest-ServersNegativeTestMultiTenantJSON-1497120864 tempest-ServersNegativeTestMultiTenantJSON-1497120864-project-member] [instance: 1d1b1dc4-c701-42b2-beba-344d622aef41] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 826.212908] env[61923]: DEBUG nova.compute.manager [None req-1214e428-bf00-4cbe-9e1e-7c8be1612c36 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] [instance: af3c317a-4007-4cea-a060-1e7dde5ce49e] Starting instance... {{(pid=61923) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 826.659971] env[61923]: DEBUG nova.compute.utils [None req-c06a566a-3828-47dc-91c4-25d15d717c22 tempest-ServerShowV257Test-986496257 tempest-ServerShowV257Test-986496257-project-member] Using /dev/sd instead of None {{(pid=61923) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 826.661023] env[61923]: DEBUG nova.compute.manager [None req-c06a566a-3828-47dc-91c4-25d15d717c22 tempest-ServerShowV257Test-986496257 tempest-ServerShowV257Test-986496257-project-member] [instance: b8c60af7-3561-468c-a9a5-1c645080fc69] Not allocating networking since 'none' was specified. {{(pid=61923) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 826.740170] env[61923]: DEBUG oslo_concurrency.lockutils [None req-1214e428-bf00-4cbe-9e1e-7c8be1612c36 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 827.082840] env[61923]: INFO nova.scheduler.client.report [None req-998fcb75-b821-4586-ae48-2fe4d02a9864 tempest-AttachVolumeShelveTestJSON-1252849694 tempest-AttachVolumeShelveTestJSON-1252849694-project-member] Deleted allocations for instance 870602dd-1b85-4c97-9506-f08e9f1bade6 [ 827.168172] env[61923]: DEBUG nova.compute.manager [None req-c06a566a-3828-47dc-91c4-25d15d717c22 tempest-ServerShowV257Test-986496257 tempest-ServerShowV257Test-986496257-project-member] [instance: b8c60af7-3561-468c-a9a5-1c645080fc69] Start building block device mappings for instance. {{(pid=61923) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 827.405351] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73d5491f-2ffb-400c-9d8d-f7f06caf53d6 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.416488] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96faea05-728f-4974-ab16-3d5a31b0169a {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.447816] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f42b7f2e-20ab-448e-b983-8d3474c32a97 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.456090] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b563b99-5501-43ea-ba3e-5f6ea04c2a7d {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.472800] env[61923]: DEBUG nova.compute.provider_tree [None req-8e8efa20-a2ed-4da9-89f1-5ff799c0382c tempest-ServersNegativeTestMultiTenantJSON-1497120864 tempest-ServersNegativeTestMultiTenantJSON-1497120864-project-member] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 827.597307] env[61923]: DEBUG oslo_concurrency.lockutils [None req-998fcb75-b821-4586-ae48-2fe4d02a9864 tempest-AttachVolumeShelveTestJSON-1252849694 tempest-AttachVolumeShelveTestJSON-1252849694-project-member] Lock "870602dd-1b85-4c97-9506-f08e9f1bade6" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 109.398s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 827.976435] env[61923]: DEBUG nova.scheduler.client.report [None req-8e8efa20-a2ed-4da9-89f1-5ff799c0382c tempest-ServersNegativeTestMultiTenantJSON-1497120864 tempest-ServersNegativeTestMultiTenantJSON-1497120864-project-member] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 828.100216] env[61923]: DEBUG nova.compute.manager [None req-52eadf86-cc45-4cae-af2a-645ac3cb8658 tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] [instance: 87ae37dc-d0d0-4b76-afdd-0ba3e8f6ce0b] Starting instance... {{(pid=61923) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 828.181105] env[61923]: DEBUG nova.compute.manager [None req-c06a566a-3828-47dc-91c4-25d15d717c22 tempest-ServerShowV257Test-986496257 tempest-ServerShowV257Test-986496257-project-member] [instance: b8c60af7-3561-468c-a9a5-1c645080fc69] Start spawning the instance on the hypervisor. {{(pid=61923) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 828.210509] env[61923]: DEBUG nova.virt.hardware [None req-c06a566a-3828-47dc-91c4-25d15d717c22 tempest-ServerShowV257Test-986496257 tempest-ServerShowV257Test-986496257-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-29T20:07:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-29T20:07:35Z,direct_url=,disk_format='vmdk',id=0f153f63-ae0a-45b1-b7f5-7f9b673c947f,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='4e7b5e3b3c3443c8bd5c70f8a15739f5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-29T20:07:36Z,virtual_size=,visibility=), allow threads: False {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 828.212461] env[61923]: DEBUG nova.virt.hardware [None req-c06a566a-3828-47dc-91c4-25d15d717c22 tempest-ServerShowV257Test-986496257 tempest-ServerShowV257Test-986496257-project-member] Flavor limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 828.212461] env[61923]: DEBUG nova.virt.hardware [None req-c06a566a-3828-47dc-91c4-25d15d717c22 tempest-ServerShowV257Test-986496257 tempest-ServerShowV257Test-986496257-project-member] Image limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 828.212461] env[61923]: DEBUG nova.virt.hardware [None req-c06a566a-3828-47dc-91c4-25d15d717c22 tempest-ServerShowV257Test-986496257 tempest-ServerShowV257Test-986496257-project-member] Flavor pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 828.212461] env[61923]: DEBUG nova.virt.hardware [None req-c06a566a-3828-47dc-91c4-25d15d717c22 tempest-ServerShowV257Test-986496257 tempest-ServerShowV257Test-986496257-project-member] Image pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 828.212461] env[61923]: DEBUG nova.virt.hardware [None req-c06a566a-3828-47dc-91c4-25d15d717c22 tempest-ServerShowV257Test-986496257 tempest-ServerShowV257Test-986496257-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 828.212461] env[61923]: DEBUG nova.virt.hardware [None req-c06a566a-3828-47dc-91c4-25d15d717c22 tempest-ServerShowV257Test-986496257 tempest-ServerShowV257Test-986496257-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 828.212713] env[61923]: DEBUG nova.virt.hardware [None req-c06a566a-3828-47dc-91c4-25d15d717c22 tempest-ServerShowV257Test-986496257 tempest-ServerShowV257Test-986496257-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 828.212713] env[61923]: DEBUG nova.virt.hardware [None req-c06a566a-3828-47dc-91c4-25d15d717c22 tempest-ServerShowV257Test-986496257 tempest-ServerShowV257Test-986496257-project-member] Got 1 possible topologies {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 828.212713] env[61923]: DEBUG nova.virt.hardware [None req-c06a566a-3828-47dc-91c4-25d15d717c22 tempest-ServerShowV257Test-986496257 tempest-ServerShowV257Test-986496257-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 828.213359] env[61923]: DEBUG nova.virt.hardware [None req-c06a566a-3828-47dc-91c4-25d15d717c22 tempest-ServerShowV257Test-986496257 tempest-ServerShowV257Test-986496257-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 828.214436] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd8ce9e2-ab65-4d2d-aae1-e11a9734c4a0 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.223188] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f212b20-b094-4956-a456-e86a74ef9e0b {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.238223] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-c06a566a-3828-47dc-91c4-25d15d717c22 tempest-ServerShowV257Test-986496257 tempest-ServerShowV257Test-986496257-project-member] [instance: b8c60af7-3561-468c-a9a5-1c645080fc69] Instance VIF info [] {{(pid=61923) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 828.244428] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-c06a566a-3828-47dc-91c4-25d15d717c22 tempest-ServerShowV257Test-986496257 tempest-ServerShowV257Test-986496257-project-member] Creating folder: Project (bc986c0f2d4a406cb47220f2b093145b). Parent ref: group-v292629. {{(pid=61923) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 828.244798] env[61923]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-eaff5522-17f8-49a5-b642-2a99e75f369b {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.255409] env[61923]: INFO nova.virt.vmwareapi.vm_util [None req-c06a566a-3828-47dc-91c4-25d15d717c22 tempest-ServerShowV257Test-986496257 tempest-ServerShowV257Test-986496257-project-member] Created folder: Project (bc986c0f2d4a406cb47220f2b093145b) in parent group-v292629. [ 828.257229] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-c06a566a-3828-47dc-91c4-25d15d717c22 tempest-ServerShowV257Test-986496257 tempest-ServerShowV257Test-986496257-project-member] Creating folder: Instances. Parent ref: group-v292650. {{(pid=61923) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 828.257229] env[61923]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-1cdd5174-6e13-4917-9e6e-46a913439452 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.265087] env[61923]: INFO nova.virt.vmwareapi.vm_util [None req-c06a566a-3828-47dc-91c4-25d15d717c22 tempest-ServerShowV257Test-986496257 tempest-ServerShowV257Test-986496257-project-member] Created folder: Instances in parent group-v292650. [ 828.265337] env[61923]: DEBUG oslo.service.loopingcall [None req-c06a566a-3828-47dc-91c4-25d15d717c22 tempest-ServerShowV257Test-986496257 tempest-ServerShowV257Test-986496257-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61923) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 828.265918] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b8c60af7-3561-468c-a9a5-1c645080fc69] Creating VM on the ESX host {{(pid=61923) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 828.265918] env[61923]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-4e049846-e664-4464-979c-2c263bfb5275 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.282857] env[61923]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 828.282857] env[61923]: value = "task-1377516" [ 828.282857] env[61923]: _type = "Task" [ 828.282857] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 828.290332] env[61923]: DEBUG oslo_vmware.api [-] Task: {'id': task-1377516, 'name': CreateVM_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 828.482377] env[61923]: DEBUG oslo_concurrency.lockutils [None req-8e8efa20-a2ed-4da9-89f1-5ff799c0382c tempest-ServersNegativeTestMultiTenantJSON-1497120864 tempest-ServersNegativeTestMultiTenantJSON-1497120864-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.338s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 828.483047] env[61923]: DEBUG nova.compute.manager [None req-8e8efa20-a2ed-4da9-89f1-5ff799c0382c tempest-ServersNegativeTestMultiTenantJSON-1497120864 tempest-ServersNegativeTestMultiTenantJSON-1497120864-project-member] [instance: 1d1b1dc4-c701-42b2-beba-344d622aef41] Start building networks asynchronously for instance. {{(pid=61923) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 828.486093] env[61923]: DEBUG oslo_concurrency.lockutils [None req-458eed44-c3ca-4049-bcef-fbcaf60c50ed tempest-ServersTestManualDisk-2012652654 tempest-ServersTestManualDisk-2012652654-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 12.804s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 828.487747] env[61923]: INFO nova.compute.claims [None req-458eed44-c3ca-4049-bcef-fbcaf60c50ed tempest-ServersTestManualDisk-2012652654 tempest-ServersTestManualDisk-2012652654-project-member] [instance: f1839f61-4314-48fe-8ab6-14b5e49d569d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 828.625356] env[61923]: DEBUG oslo_concurrency.lockutils [None req-52eadf86-cc45-4cae-af2a-645ac3cb8658 tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 828.796024] env[61923]: DEBUG oslo_vmware.api [-] Task: {'id': task-1377516, 'name': CreateVM_Task, 'duration_secs': 0.313282} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 828.796024] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b8c60af7-3561-468c-a9a5-1c645080fc69] Created VM on the ESX host {{(pid=61923) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 828.796024] env[61923]: DEBUG oslo_concurrency.lockutils [None req-c06a566a-3828-47dc-91c4-25d15d717c22 tempest-ServerShowV257Test-986496257 tempest-ServerShowV257Test-986496257-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 828.796024] env[61923]: DEBUG oslo_concurrency.lockutils [None req-c06a566a-3828-47dc-91c4-25d15d717c22 tempest-ServerShowV257Test-986496257 tempest-ServerShowV257Test-986496257-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 828.796024] env[61923]: DEBUG oslo_concurrency.lockutils [None req-c06a566a-3828-47dc-91c4-25d15d717c22 tempest-ServerShowV257Test-986496257 tempest-ServerShowV257Test-986496257-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 828.796024] env[61923]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-073b521f-1133-4687-adb3-2a7fcb627e20 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.799386] env[61923]: DEBUG oslo_vmware.api [None req-c06a566a-3828-47dc-91c4-25d15d717c22 tempest-ServerShowV257Test-986496257 tempest-ServerShowV257Test-986496257-project-member] Waiting for the task: (returnval){ [ 828.799386] env[61923]: value = "session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52454868-0705-cd9b-d4b6-d0c403dd0011" [ 828.799386] env[61923]: _type = "Task" [ 828.799386] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 828.807163] env[61923]: DEBUG oslo_vmware.api [None req-c06a566a-3828-47dc-91c4-25d15d717c22 tempest-ServerShowV257Test-986496257 tempest-ServerShowV257Test-986496257-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52454868-0705-cd9b-d4b6-d0c403dd0011, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 828.997027] env[61923]: DEBUG nova.compute.utils [None req-8e8efa20-a2ed-4da9-89f1-5ff799c0382c tempest-ServersNegativeTestMultiTenantJSON-1497120864 tempest-ServersNegativeTestMultiTenantJSON-1497120864-project-member] Using /dev/sd instead of None {{(pid=61923) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 828.998414] env[61923]: DEBUG nova.compute.manager [None req-8e8efa20-a2ed-4da9-89f1-5ff799c0382c tempest-ServersNegativeTestMultiTenantJSON-1497120864 tempest-ServersNegativeTestMultiTenantJSON-1497120864-project-member] [instance: 1d1b1dc4-c701-42b2-beba-344d622aef41] Allocating IP information in the background. {{(pid=61923) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 828.998812] env[61923]: DEBUG nova.network.neutron [None req-8e8efa20-a2ed-4da9-89f1-5ff799c0382c tempest-ServersNegativeTestMultiTenantJSON-1497120864 tempest-ServersNegativeTestMultiTenantJSON-1497120864-project-member] [instance: 1d1b1dc4-c701-42b2-beba-344d622aef41] allocate_for_instance() {{(pid=61923) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 829.166110] env[61923]: DEBUG nova.policy [None req-8e8efa20-a2ed-4da9-89f1-5ff799c0382c tempest-ServersNegativeTestMultiTenantJSON-1497120864 tempest-ServersNegativeTestMultiTenantJSON-1497120864-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a3d5c9d80015406789fcb8ae93c7e7ac', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '93b2ac1779c1498d9cabeb015353024a', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61923) authorize /opt/stack/nova/nova/policy.py:201}} [ 829.315026] env[61923]: DEBUG oslo_vmware.api [None req-c06a566a-3828-47dc-91c4-25d15d717c22 tempest-ServerShowV257Test-986496257 tempest-ServerShowV257Test-986496257-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52454868-0705-cd9b-d4b6-d0c403dd0011, 'name': SearchDatastore_Task, 'duration_secs': 0.01101} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 829.315026] env[61923]: DEBUG oslo_concurrency.lockutils [None req-c06a566a-3828-47dc-91c4-25d15d717c22 tempest-ServerShowV257Test-986496257 tempest-ServerShowV257Test-986496257-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 829.315026] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-c06a566a-3828-47dc-91c4-25d15d717c22 tempest-ServerShowV257Test-986496257 tempest-ServerShowV257Test-986496257-project-member] [instance: b8c60af7-3561-468c-a9a5-1c645080fc69] Processing image 0f153f63-ae0a-45b1-b7f5-7f9b673c947f {{(pid=61923) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 829.315026] env[61923]: DEBUG oslo_concurrency.lockutils [None req-c06a566a-3828-47dc-91c4-25d15d717c22 tempest-ServerShowV257Test-986496257 tempest-ServerShowV257Test-986496257-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 829.315211] env[61923]: DEBUG oslo_concurrency.lockutils [None req-c06a566a-3828-47dc-91c4-25d15d717c22 tempest-ServerShowV257Test-986496257 tempest-ServerShowV257Test-986496257-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 829.315211] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-c06a566a-3828-47dc-91c4-25d15d717c22 tempest-ServerShowV257Test-986496257 tempest-ServerShowV257Test-986496257-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61923) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 829.315211] env[61923]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-0ab4e628-32a8-44e5-94c9-1a1a07d4bb6b {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.325164] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-c06a566a-3828-47dc-91c4-25d15d717c22 tempest-ServerShowV257Test-986496257 tempest-ServerShowV257Test-986496257-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61923) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 829.325164] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-c06a566a-3828-47dc-91c4-25d15d717c22 tempest-ServerShowV257Test-986496257 tempest-ServerShowV257Test-986496257-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61923) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 829.325164] env[61923]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7a4766d0-b757-4a16-835a-4b0ab62fd095 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.330260] env[61923]: DEBUG oslo_vmware.api [None req-c06a566a-3828-47dc-91c4-25d15d717c22 tempest-ServerShowV257Test-986496257 tempest-ServerShowV257Test-986496257-project-member] Waiting for the task: (returnval){ [ 829.330260] env[61923]: value = "session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52c0687c-d992-61dd-8c02-09b793138816" [ 829.330260] env[61923]: _type = "Task" [ 829.330260] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 829.339446] env[61923]: DEBUG oslo_vmware.api [None req-c06a566a-3828-47dc-91c4-25d15d717c22 tempest-ServerShowV257Test-986496257 tempest-ServerShowV257Test-986496257-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52c0687c-d992-61dd-8c02-09b793138816, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 829.501073] env[61923]: DEBUG nova.compute.manager [None req-8e8efa20-a2ed-4da9-89f1-5ff799c0382c tempest-ServersNegativeTestMultiTenantJSON-1497120864 tempest-ServersNegativeTestMultiTenantJSON-1497120864-project-member] [instance: 1d1b1dc4-c701-42b2-beba-344d622aef41] Start building block device mappings for instance. {{(pid=61923) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 829.669052] env[61923]: DEBUG nova.network.neutron [None req-8e8efa20-a2ed-4da9-89f1-5ff799c0382c tempest-ServersNegativeTestMultiTenantJSON-1497120864 tempest-ServersNegativeTestMultiTenantJSON-1497120864-project-member] [instance: 1d1b1dc4-c701-42b2-beba-344d622aef41] Successfully created port: 8118febd-f949-4bed-9959-775f44cfbd07 {{(pid=61923) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 829.732843] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-459324ea-afc4-443e-bfc0-9d6b06f20e6d {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.743264] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7753dd20-4fef-4e99-b7e0-f97dde771377 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.777423] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4167b5b-acce-4611-8ca0-cf2eaa199ff6 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.785538] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c7187bc-d4c4-486a-b85c-66ed17c1072f {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.801693] env[61923]: DEBUG nova.compute.provider_tree [None req-458eed44-c3ca-4049-bcef-fbcaf60c50ed tempest-ServersTestManualDisk-2012652654 tempest-ServersTestManualDisk-2012652654-project-member] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 829.840996] env[61923]: DEBUG oslo_vmware.api [None req-c06a566a-3828-47dc-91c4-25d15d717c22 tempest-ServerShowV257Test-986496257 tempest-ServerShowV257Test-986496257-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52c0687c-d992-61dd-8c02-09b793138816, 'name': SearchDatastore_Task, 'duration_secs': 0.008743} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 829.844922] env[61923]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d275e369-3635-4414-a9bb-3f7016eda824 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.852904] env[61923]: DEBUG oslo_vmware.api [None req-c06a566a-3828-47dc-91c4-25d15d717c22 tempest-ServerShowV257Test-986496257 tempest-ServerShowV257Test-986496257-project-member] Waiting for the task: (returnval){ [ 829.852904] env[61923]: value = "session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]520a298b-7b30-4b14-a3a7-170e70ca76fa" [ 829.852904] env[61923]: _type = "Task" [ 829.852904] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 829.866639] env[61923]: DEBUG oslo_vmware.api [None req-c06a566a-3828-47dc-91c4-25d15d717c22 tempest-ServerShowV257Test-986496257 tempest-ServerShowV257Test-986496257-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]520a298b-7b30-4b14-a3a7-170e70ca76fa, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 830.305759] env[61923]: DEBUG nova.scheduler.client.report [None req-458eed44-c3ca-4049-bcef-fbcaf60c50ed tempest-ServersTestManualDisk-2012652654 tempest-ServersTestManualDisk-2012652654-project-member] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 830.365042] env[61923]: DEBUG oslo_vmware.api [None req-c06a566a-3828-47dc-91c4-25d15d717c22 tempest-ServerShowV257Test-986496257 tempest-ServerShowV257Test-986496257-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]520a298b-7b30-4b14-a3a7-170e70ca76fa, 'name': SearchDatastore_Task, 'duration_secs': 0.010462} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 830.365403] env[61923]: DEBUG oslo_concurrency.lockutils [None req-c06a566a-3828-47dc-91c4-25d15d717c22 tempest-ServerShowV257Test-986496257 tempest-ServerShowV257Test-986496257-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 830.366037] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-c06a566a-3828-47dc-91c4-25d15d717c22 tempest-ServerShowV257Test-986496257 tempest-ServerShowV257Test-986496257-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk to [datastore1] b8c60af7-3561-468c-a9a5-1c645080fc69/b8c60af7-3561-468c-a9a5-1c645080fc69.vmdk {{(pid=61923) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 830.366279] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-bab1df4c-c3b0-4d3f-95ce-25456fc5c738 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.374498] env[61923]: DEBUG oslo_vmware.api [None req-c06a566a-3828-47dc-91c4-25d15d717c22 tempest-ServerShowV257Test-986496257 tempest-ServerShowV257Test-986496257-project-member] Waiting for the task: (returnval){ [ 830.374498] env[61923]: value = "task-1377517" [ 830.374498] env[61923]: _type = "Task" [ 830.374498] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 830.382396] env[61923]: DEBUG oslo_vmware.api [None req-c06a566a-3828-47dc-91c4-25d15d717c22 tempest-ServerShowV257Test-986496257 tempest-ServerShowV257Test-986496257-project-member] Task: {'id': task-1377517, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 830.512905] env[61923]: DEBUG nova.compute.manager [None req-8e8efa20-a2ed-4da9-89f1-5ff799c0382c tempest-ServersNegativeTestMultiTenantJSON-1497120864 tempest-ServersNegativeTestMultiTenantJSON-1497120864-project-member] [instance: 1d1b1dc4-c701-42b2-beba-344d622aef41] Start spawning the instance on the hypervisor. {{(pid=61923) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 830.553030] env[61923]: DEBUG nova.virt.hardware [None req-8e8efa20-a2ed-4da9-89f1-5ff799c0382c tempest-ServersNegativeTestMultiTenantJSON-1497120864 tempest-ServersNegativeTestMultiTenantJSON-1497120864-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-29T20:07:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-29T20:07:35Z,direct_url=,disk_format='vmdk',id=0f153f63-ae0a-45b1-b7f5-7f9b673c947f,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='4e7b5e3b3c3443c8bd5c70f8a15739f5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-29T20:07:36Z,virtual_size=,visibility=), allow threads: False {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 830.553397] env[61923]: DEBUG nova.virt.hardware [None req-8e8efa20-a2ed-4da9-89f1-5ff799c0382c tempest-ServersNegativeTestMultiTenantJSON-1497120864 tempest-ServersNegativeTestMultiTenantJSON-1497120864-project-member] Flavor limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 830.553666] env[61923]: DEBUG nova.virt.hardware [None req-8e8efa20-a2ed-4da9-89f1-5ff799c0382c tempest-ServersNegativeTestMultiTenantJSON-1497120864 tempest-ServersNegativeTestMultiTenantJSON-1497120864-project-member] Image limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 830.553870] env[61923]: DEBUG nova.virt.hardware [None req-8e8efa20-a2ed-4da9-89f1-5ff799c0382c tempest-ServersNegativeTestMultiTenantJSON-1497120864 tempest-ServersNegativeTestMultiTenantJSON-1497120864-project-member] Flavor pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 830.554031] env[61923]: DEBUG nova.virt.hardware [None req-8e8efa20-a2ed-4da9-89f1-5ff799c0382c tempest-ServersNegativeTestMultiTenantJSON-1497120864 tempest-ServersNegativeTestMultiTenantJSON-1497120864-project-member] Image pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 830.554200] env[61923]: DEBUG nova.virt.hardware [None req-8e8efa20-a2ed-4da9-89f1-5ff799c0382c tempest-ServersNegativeTestMultiTenantJSON-1497120864 tempest-ServersNegativeTestMultiTenantJSON-1497120864-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 830.554510] env[61923]: DEBUG nova.virt.hardware [None req-8e8efa20-a2ed-4da9-89f1-5ff799c0382c tempest-ServersNegativeTestMultiTenantJSON-1497120864 tempest-ServersNegativeTestMultiTenantJSON-1497120864-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 830.554612] env[61923]: DEBUG nova.virt.hardware [None req-8e8efa20-a2ed-4da9-89f1-5ff799c0382c tempest-ServersNegativeTestMultiTenantJSON-1497120864 tempest-ServersNegativeTestMultiTenantJSON-1497120864-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 830.554763] env[61923]: DEBUG nova.virt.hardware [None req-8e8efa20-a2ed-4da9-89f1-5ff799c0382c tempest-ServersNegativeTestMultiTenantJSON-1497120864 tempest-ServersNegativeTestMultiTenantJSON-1497120864-project-member] Got 1 possible topologies {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 830.554918] env[61923]: DEBUG nova.virt.hardware [None req-8e8efa20-a2ed-4da9-89f1-5ff799c0382c tempest-ServersNegativeTestMultiTenantJSON-1497120864 tempest-ServersNegativeTestMultiTenantJSON-1497120864-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 830.557347] env[61923]: DEBUG nova.virt.hardware [None req-8e8efa20-a2ed-4da9-89f1-5ff799c0382c tempest-ServersNegativeTestMultiTenantJSON-1497120864 tempest-ServersNegativeTestMultiTenantJSON-1497120864-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 830.558271] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c9c8f4c-79d3-4f4a-a74e-7d222c0285b2 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.572404] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8108a6e-4a8a-4567-8fb6-7371c6d6022a {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 830.812286] env[61923]: DEBUG oslo_concurrency.lockutils [None req-458eed44-c3ca-4049-bcef-fbcaf60c50ed tempest-ServersTestManualDisk-2012652654 tempest-ServersTestManualDisk-2012652654-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.325s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 830.813081] env[61923]: DEBUG nova.compute.manager [None req-458eed44-c3ca-4049-bcef-fbcaf60c50ed tempest-ServersTestManualDisk-2012652654 tempest-ServersTestManualDisk-2012652654-project-member] [instance: f1839f61-4314-48fe-8ab6-14b5e49d569d] Start building networks asynchronously for instance. {{(pid=61923) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 830.816518] env[61923]: DEBUG oslo_concurrency.lockutils [None req-9ff8b121-368d-4886-9882-962e7ef81bcf tempest-ServerTagsTestJSON-19215696 tempest-ServerTagsTestJSON-19215696-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.271s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 830.818035] env[61923]: INFO nova.compute.claims [None req-9ff8b121-368d-4886-9882-962e7ef81bcf tempest-ServerTagsTestJSON-19215696 tempest-ServerTagsTestJSON-19215696-project-member] [instance: 48825592-dcdc-41d5-9fbf-500d1f31017b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 830.886793] env[61923]: DEBUG oslo_vmware.api [None req-c06a566a-3828-47dc-91c4-25d15d717c22 tempest-ServerShowV257Test-986496257 tempest-ServerShowV257Test-986496257-project-member] Task: {'id': task-1377517, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.325479] env[61923]: DEBUG nova.compute.utils [None req-458eed44-c3ca-4049-bcef-fbcaf60c50ed tempest-ServersTestManualDisk-2012652654 tempest-ServersTestManualDisk-2012652654-project-member] Using /dev/sd instead of None {{(pid=61923) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 831.328703] env[61923]: DEBUG nova.compute.manager [None req-458eed44-c3ca-4049-bcef-fbcaf60c50ed tempest-ServersTestManualDisk-2012652654 tempest-ServersTestManualDisk-2012652654-project-member] [instance: f1839f61-4314-48fe-8ab6-14b5e49d569d] Allocating IP information in the background. {{(pid=61923) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 831.328703] env[61923]: DEBUG nova.network.neutron [None req-458eed44-c3ca-4049-bcef-fbcaf60c50ed tempest-ServersTestManualDisk-2012652654 tempest-ServersTestManualDisk-2012652654-project-member] [instance: f1839f61-4314-48fe-8ab6-14b5e49d569d] allocate_for_instance() {{(pid=61923) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 831.387931] env[61923]: DEBUG oslo_vmware.api [None req-c06a566a-3828-47dc-91c4-25d15d717c22 tempest-ServerShowV257Test-986496257 tempest-ServerShowV257Test-986496257-project-member] Task: {'id': task-1377517, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.787335} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 831.388279] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-c06a566a-3828-47dc-91c4-25d15d717c22 tempest-ServerShowV257Test-986496257 tempest-ServerShowV257Test-986496257-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk to [datastore1] b8c60af7-3561-468c-a9a5-1c645080fc69/b8c60af7-3561-468c-a9a5-1c645080fc69.vmdk {{(pid=61923) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 831.388501] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-c06a566a-3828-47dc-91c4-25d15d717c22 tempest-ServerShowV257Test-986496257 tempest-ServerShowV257Test-986496257-project-member] [instance: b8c60af7-3561-468c-a9a5-1c645080fc69] Extending root virtual disk to 1048576 {{(pid=61923) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 831.388796] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-38092c74-ca30-47d6-b6ae-f45635bda5f7 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.395131] env[61923]: DEBUG oslo_vmware.api [None req-c06a566a-3828-47dc-91c4-25d15d717c22 tempest-ServerShowV257Test-986496257 tempest-ServerShowV257Test-986496257-project-member] Waiting for the task: (returnval){ [ 831.395131] env[61923]: value = "task-1377518" [ 831.395131] env[61923]: _type = "Task" [ 831.395131] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 831.404845] env[61923]: DEBUG oslo_vmware.api [None req-c06a566a-3828-47dc-91c4-25d15d717c22 tempest-ServerShowV257Test-986496257 tempest-ServerShowV257Test-986496257-project-member] Task: {'id': task-1377518, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 831.409766] env[61923]: DEBUG nova.policy [None req-458eed44-c3ca-4049-bcef-fbcaf60c50ed tempest-ServersTestManualDisk-2012652654 tempest-ServersTestManualDisk-2012652654-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '03df51b48f5c4d02962b14bbe6eef55f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5357434049fb4acd910988fc5ff0d2c4', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61923) authorize /opt/stack/nova/nova/policy.py:201}} [ 831.513851] env[61923]: DEBUG nova.compute.manager [req-94dff1bc-f572-4fca-ab17-0305b5fb14dc req-ccd86ac8-0e4f-4e86-b7aa-afdd98c16b1d service nova] [instance: 1d1b1dc4-c701-42b2-beba-344d622aef41] Received event network-vif-plugged-8118febd-f949-4bed-9959-775f44cfbd07 {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 831.514176] env[61923]: DEBUG oslo_concurrency.lockutils [req-94dff1bc-f572-4fca-ab17-0305b5fb14dc req-ccd86ac8-0e4f-4e86-b7aa-afdd98c16b1d service nova] Acquiring lock "1d1b1dc4-c701-42b2-beba-344d622aef41-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 831.514404] env[61923]: DEBUG oslo_concurrency.lockutils [req-94dff1bc-f572-4fca-ab17-0305b5fb14dc req-ccd86ac8-0e4f-4e86-b7aa-afdd98c16b1d service nova] Lock "1d1b1dc4-c701-42b2-beba-344d622aef41-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 831.515393] env[61923]: DEBUG oslo_concurrency.lockutils [req-94dff1bc-f572-4fca-ab17-0305b5fb14dc req-ccd86ac8-0e4f-4e86-b7aa-afdd98c16b1d service nova] Lock "1d1b1dc4-c701-42b2-beba-344d622aef41-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 831.515393] env[61923]: DEBUG nova.compute.manager [req-94dff1bc-f572-4fca-ab17-0305b5fb14dc req-ccd86ac8-0e4f-4e86-b7aa-afdd98c16b1d service nova] [instance: 1d1b1dc4-c701-42b2-beba-344d622aef41] No waiting events found dispatching network-vif-plugged-8118febd-f949-4bed-9959-775f44cfbd07 {{(pid=61923) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 831.515518] env[61923]: WARNING nova.compute.manager [req-94dff1bc-f572-4fca-ab17-0305b5fb14dc req-ccd86ac8-0e4f-4e86-b7aa-afdd98c16b1d service nova] [instance: 1d1b1dc4-c701-42b2-beba-344d622aef41] Received unexpected event network-vif-plugged-8118febd-f949-4bed-9959-775f44cfbd07 for instance with vm_state building and task_state spawning. [ 831.829645] env[61923]: DEBUG nova.compute.manager [None req-458eed44-c3ca-4049-bcef-fbcaf60c50ed tempest-ServersTestManualDisk-2012652654 tempest-ServersTestManualDisk-2012652654-project-member] [instance: f1839f61-4314-48fe-8ab6-14b5e49d569d] Start building block device mappings for instance. {{(pid=61923) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 831.911736] env[61923]: DEBUG oslo_vmware.api [None req-c06a566a-3828-47dc-91c4-25d15d717c22 tempest-ServerShowV257Test-986496257 tempest-ServerShowV257Test-986496257-project-member] Task: {'id': task-1377518, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.066512} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 831.915146] env[61923]: DEBUG nova.network.neutron [None req-8e8efa20-a2ed-4da9-89f1-5ff799c0382c tempest-ServersNegativeTestMultiTenantJSON-1497120864 tempest-ServersNegativeTestMultiTenantJSON-1497120864-project-member] [instance: 1d1b1dc4-c701-42b2-beba-344d622aef41] Successfully updated port: 8118febd-f949-4bed-9959-775f44cfbd07 {{(pid=61923) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 831.917749] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-c06a566a-3828-47dc-91c4-25d15d717c22 tempest-ServerShowV257Test-986496257 tempest-ServerShowV257Test-986496257-project-member] [instance: b8c60af7-3561-468c-a9a5-1c645080fc69] Extended root virtual disk {{(pid=61923) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 831.918673] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9acf654-6050-4fcc-a743-9efd74c80ffd {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.947698] env[61923]: DEBUG nova.virt.vmwareapi.volumeops [None req-c06a566a-3828-47dc-91c4-25d15d717c22 tempest-ServerShowV257Test-986496257 tempest-ServerShowV257Test-986496257-project-member] [instance: b8c60af7-3561-468c-a9a5-1c645080fc69] Reconfiguring VM instance instance-0000003f to attach disk [datastore1] b8c60af7-3561-468c-a9a5-1c645080fc69/b8c60af7-3561-468c-a9a5-1c645080fc69.vmdk or device None with type sparse {{(pid=61923) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 831.951247] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-123bd04e-09d4-425d-9b7b-62c6027bb25f {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.972374] env[61923]: DEBUG oslo_vmware.api [None req-c06a566a-3828-47dc-91c4-25d15d717c22 tempest-ServerShowV257Test-986496257 tempest-ServerShowV257Test-986496257-project-member] Waiting for the task: (returnval){ [ 831.972374] env[61923]: value = "task-1377519" [ 831.972374] env[61923]: _type = "Task" [ 831.972374] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 831.982237] env[61923]: DEBUG oslo_vmware.api [None req-c06a566a-3828-47dc-91c4-25d15d717c22 tempest-ServerShowV257Test-986496257 tempest-ServerShowV257Test-986496257-project-member] Task: {'id': task-1377519, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.126823] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe6b9812-c4fa-4b11-851f-ff510db2b59c {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.135202] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6fa4d9bf-c3a9-44d2-95ae-198effe14bc0 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.165351] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c331688c-53b7-40e3-aebd-02efbf48497c {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.172922] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64d2a03f-5757-4ff6-9c91-0122416cdc6f {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.185811] env[61923]: DEBUG nova.compute.provider_tree [None req-9ff8b121-368d-4886-9882-962e7ef81bcf tempest-ServerTagsTestJSON-19215696 tempest-ServerTagsTestJSON-19215696-project-member] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 832.187623] env[61923]: DEBUG nova.network.neutron [None req-458eed44-c3ca-4049-bcef-fbcaf60c50ed tempest-ServersTestManualDisk-2012652654 tempest-ServersTestManualDisk-2012652654-project-member] [instance: f1839f61-4314-48fe-8ab6-14b5e49d569d] Successfully created port: f36077f4-1c49-48e7-9c41-0f45a58883e7 {{(pid=61923) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 832.419299] env[61923]: DEBUG oslo_concurrency.lockutils [None req-8e8efa20-a2ed-4da9-89f1-5ff799c0382c tempest-ServersNegativeTestMultiTenantJSON-1497120864 tempest-ServersNegativeTestMultiTenantJSON-1497120864-project-member] Acquiring lock "refresh_cache-1d1b1dc4-c701-42b2-beba-344d622aef41" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 832.419538] env[61923]: DEBUG oslo_concurrency.lockutils [None req-8e8efa20-a2ed-4da9-89f1-5ff799c0382c tempest-ServersNegativeTestMultiTenantJSON-1497120864 tempest-ServersNegativeTestMultiTenantJSON-1497120864-project-member] Acquired lock "refresh_cache-1d1b1dc4-c701-42b2-beba-344d622aef41" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 832.419685] env[61923]: DEBUG nova.network.neutron [None req-8e8efa20-a2ed-4da9-89f1-5ff799c0382c tempest-ServersNegativeTestMultiTenantJSON-1497120864 tempest-ServersNegativeTestMultiTenantJSON-1497120864-project-member] [instance: 1d1b1dc4-c701-42b2-beba-344d622aef41] Building network info cache for instance {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 832.482653] env[61923]: DEBUG oslo_vmware.api [None req-c06a566a-3828-47dc-91c4-25d15d717c22 tempest-ServerShowV257Test-986496257 tempest-ServerShowV257Test-986496257-project-member] Task: {'id': task-1377519, 'name': ReconfigVM_Task, 'duration_secs': 0.321512} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 832.482653] env[61923]: DEBUG nova.virt.vmwareapi.volumeops [None req-c06a566a-3828-47dc-91c4-25d15d717c22 tempest-ServerShowV257Test-986496257 tempest-ServerShowV257Test-986496257-project-member] [instance: b8c60af7-3561-468c-a9a5-1c645080fc69] Reconfigured VM instance instance-0000003f to attach disk [datastore1] b8c60af7-3561-468c-a9a5-1c645080fc69/b8c60af7-3561-468c-a9a5-1c645080fc69.vmdk or device None with type sparse {{(pid=61923) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 832.483747] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e485a607-5862-4644-8dec-aa128a2440ec {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.491958] env[61923]: DEBUG oslo_vmware.api [None req-c06a566a-3828-47dc-91c4-25d15d717c22 tempest-ServerShowV257Test-986496257 tempest-ServerShowV257Test-986496257-project-member] Waiting for the task: (returnval){ [ 832.491958] env[61923]: value = "task-1377520" [ 832.491958] env[61923]: _type = "Task" [ 832.491958] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 832.500020] env[61923]: DEBUG oslo_vmware.api [None req-c06a566a-3828-47dc-91c4-25d15d717c22 tempest-ServerShowV257Test-986496257 tempest-ServerShowV257Test-986496257-project-member] Task: {'id': task-1377520, 'name': Rename_Task} progress is 5%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 832.691551] env[61923]: DEBUG nova.scheduler.client.report [None req-9ff8b121-368d-4886-9882-962e7ef81bcf tempest-ServerTagsTestJSON-19215696 tempest-ServerTagsTestJSON-19215696-project-member] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 832.848015] env[61923]: DEBUG nova.compute.manager [None req-458eed44-c3ca-4049-bcef-fbcaf60c50ed tempest-ServersTestManualDisk-2012652654 tempest-ServersTestManualDisk-2012652654-project-member] [instance: f1839f61-4314-48fe-8ab6-14b5e49d569d] Start spawning the instance on the hypervisor. {{(pid=61923) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 832.872100] env[61923]: DEBUG nova.virt.hardware [None req-458eed44-c3ca-4049-bcef-fbcaf60c50ed tempest-ServersTestManualDisk-2012652654 tempest-ServersTestManualDisk-2012652654-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-29T20:07:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-29T20:07:35Z,direct_url=,disk_format='vmdk',id=0f153f63-ae0a-45b1-b7f5-7f9b673c947f,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='4e7b5e3b3c3443c8bd5c70f8a15739f5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-29T20:07:36Z,virtual_size=,visibility=), allow threads: False {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 832.872356] env[61923]: DEBUG nova.virt.hardware [None req-458eed44-c3ca-4049-bcef-fbcaf60c50ed tempest-ServersTestManualDisk-2012652654 tempest-ServersTestManualDisk-2012652654-project-member] Flavor limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 832.872510] env[61923]: DEBUG nova.virt.hardware [None req-458eed44-c3ca-4049-bcef-fbcaf60c50ed tempest-ServersTestManualDisk-2012652654 tempest-ServersTestManualDisk-2012652654-project-member] Image limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 832.872685] env[61923]: DEBUG nova.virt.hardware [None req-458eed44-c3ca-4049-bcef-fbcaf60c50ed tempest-ServersTestManualDisk-2012652654 tempest-ServersTestManualDisk-2012652654-project-member] Flavor pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 832.872825] env[61923]: DEBUG nova.virt.hardware [None req-458eed44-c3ca-4049-bcef-fbcaf60c50ed tempest-ServersTestManualDisk-2012652654 tempest-ServersTestManualDisk-2012652654-project-member] Image pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 832.872967] env[61923]: DEBUG nova.virt.hardware [None req-458eed44-c3ca-4049-bcef-fbcaf60c50ed tempest-ServersTestManualDisk-2012652654 tempest-ServersTestManualDisk-2012652654-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 832.873190] env[61923]: DEBUG nova.virt.hardware [None req-458eed44-c3ca-4049-bcef-fbcaf60c50ed tempest-ServersTestManualDisk-2012652654 tempest-ServersTestManualDisk-2012652654-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 832.873344] env[61923]: DEBUG nova.virt.hardware [None req-458eed44-c3ca-4049-bcef-fbcaf60c50ed tempest-ServersTestManualDisk-2012652654 tempest-ServersTestManualDisk-2012652654-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 832.873503] env[61923]: DEBUG nova.virt.hardware [None req-458eed44-c3ca-4049-bcef-fbcaf60c50ed tempest-ServersTestManualDisk-2012652654 tempest-ServersTestManualDisk-2012652654-project-member] Got 1 possible topologies {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 832.873657] env[61923]: DEBUG nova.virt.hardware [None req-458eed44-c3ca-4049-bcef-fbcaf60c50ed tempest-ServersTestManualDisk-2012652654 tempest-ServersTestManualDisk-2012652654-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 832.873824] env[61923]: DEBUG nova.virt.hardware [None req-458eed44-c3ca-4049-bcef-fbcaf60c50ed tempest-ServersTestManualDisk-2012652654 tempest-ServersTestManualDisk-2012652654-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 832.874686] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98c57881-1d85-49c3-941b-9c80b48717d9 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.882294] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c11b5e6e-f55e-4afa-aff2-2a488193bedd {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.963701] env[61923]: DEBUG nova.network.neutron [None req-8e8efa20-a2ed-4da9-89f1-5ff799c0382c tempest-ServersNegativeTestMultiTenantJSON-1497120864 tempest-ServersNegativeTestMultiTenantJSON-1497120864-project-member] [instance: 1d1b1dc4-c701-42b2-beba-344d622aef41] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 833.003020] env[61923]: DEBUG oslo_vmware.api [None req-c06a566a-3828-47dc-91c4-25d15d717c22 tempest-ServerShowV257Test-986496257 tempest-ServerShowV257Test-986496257-project-member] Task: {'id': task-1377520, 'name': Rename_Task, 'duration_secs': 0.134996} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 833.003368] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-c06a566a-3828-47dc-91c4-25d15d717c22 tempest-ServerShowV257Test-986496257 tempest-ServerShowV257Test-986496257-project-member] [instance: b8c60af7-3561-468c-a9a5-1c645080fc69] Powering on the VM {{(pid=61923) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 833.003641] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-94594e0a-93d4-4130-b253-00e6328eb9ca {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.009805] env[61923]: DEBUG oslo_vmware.api [None req-c06a566a-3828-47dc-91c4-25d15d717c22 tempest-ServerShowV257Test-986496257 tempest-ServerShowV257Test-986496257-project-member] Waiting for the task: (returnval){ [ 833.009805] env[61923]: value = "task-1377521" [ 833.009805] env[61923]: _type = "Task" [ 833.009805] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 833.017676] env[61923]: DEBUG oslo_vmware.api [None req-c06a566a-3828-47dc-91c4-25d15d717c22 tempest-ServerShowV257Test-986496257 tempest-ServerShowV257Test-986496257-project-member] Task: {'id': task-1377521, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 833.148436] env[61923]: DEBUG oslo_concurrency.lockutils [None req-00703b42-f309-4672-8dc7-10f4a69426aa tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Acquiring lock "b779d183-89ae-4e4d-ae99-e514e145ed43" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 833.148436] env[61923]: DEBUG oslo_concurrency.lockutils [None req-00703b42-f309-4672-8dc7-10f4a69426aa tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Lock "b779d183-89ae-4e4d-ae99-e514e145ed43" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 833.199021] env[61923]: DEBUG oslo_concurrency.lockutils [None req-9ff8b121-368d-4886-9882-962e7ef81bcf tempest-ServerTagsTestJSON-19215696 tempest-ServerTagsTestJSON-19215696-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.379s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 833.199021] env[61923]: DEBUG nova.compute.manager [None req-9ff8b121-368d-4886-9882-962e7ef81bcf tempest-ServerTagsTestJSON-19215696 tempest-ServerTagsTestJSON-19215696-project-member] [instance: 48825592-dcdc-41d5-9fbf-500d1f31017b] Start building networks asynchronously for instance. {{(pid=61923) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 833.199854] env[61923]: DEBUG oslo_concurrency.lockutils [None req-950e07c2-502c-43fa-8598-da5f189b07e3 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.770s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 833.201704] env[61923]: INFO nova.compute.claims [None req-950e07c2-502c-43fa-8598-da5f189b07e3 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] [instance: 422d06d6-6932-46c3-bb25-841e6f65c028] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 833.205892] env[61923]: DEBUG nova.network.neutron [None req-8e8efa20-a2ed-4da9-89f1-5ff799c0382c tempest-ServersNegativeTestMultiTenantJSON-1497120864 tempest-ServersNegativeTestMultiTenantJSON-1497120864-project-member] [instance: 1d1b1dc4-c701-42b2-beba-344d622aef41] Updating instance_info_cache with network_info: [{"id": "8118febd-f949-4bed-9959-775f44cfbd07", "address": "fa:16:3e:51:b5:e2", "network": {"id": "f967adc7-af3d-438f-a4eb-e98e9dbb969d", "bridge": "br-int", "label": "tempest-ServersNegativeTestMultiTenantJSON-1189375605-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "93b2ac1779c1498d9cabeb015353024a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "604c9724-b4ef-4393-a76e-eb4a2b510796", "external-id": "nsx-vlan-transportzone-909", "segmentation_id": 909, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8118febd-f9", "ovs_interfaceid": "8118febd-f949-4bed-9959-775f44cfbd07", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 833.518679] env[61923]: DEBUG oslo_vmware.api [None req-c06a566a-3828-47dc-91c4-25d15d717c22 tempest-ServerShowV257Test-986496257 tempest-ServerShowV257Test-986496257-project-member] Task: {'id': task-1377521, 'name': PowerOnVM_Task, 'duration_secs': 0.413515} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 833.519032] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-c06a566a-3828-47dc-91c4-25d15d717c22 tempest-ServerShowV257Test-986496257 tempest-ServerShowV257Test-986496257-project-member] [instance: b8c60af7-3561-468c-a9a5-1c645080fc69] Powered on the VM {{(pid=61923) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 833.519241] env[61923]: INFO nova.compute.manager [None req-c06a566a-3828-47dc-91c4-25d15d717c22 tempest-ServerShowV257Test-986496257 tempest-ServerShowV257Test-986496257-project-member] [instance: b8c60af7-3561-468c-a9a5-1c645080fc69] Took 5.34 seconds to spawn the instance on the hypervisor. [ 833.519419] env[61923]: DEBUG nova.compute.manager [None req-c06a566a-3828-47dc-91c4-25d15d717c22 tempest-ServerShowV257Test-986496257 tempest-ServerShowV257Test-986496257-project-member] [instance: b8c60af7-3561-468c-a9a5-1c645080fc69] Checking state {{(pid=61923) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 833.520311] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de395d5d-826e-468e-802d-cec3b5ee237e {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.535918] env[61923]: DEBUG nova.compute.manager [req-7d9f72df-38f7-4082-98ba-9f2055fcf77d req-6fc284ce-01ed-4557-b6b3-5301d35309d8 service nova] [instance: 1d1b1dc4-c701-42b2-beba-344d622aef41] Received event network-changed-8118febd-f949-4bed-9959-775f44cfbd07 {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 833.535918] env[61923]: DEBUG nova.compute.manager [req-7d9f72df-38f7-4082-98ba-9f2055fcf77d req-6fc284ce-01ed-4557-b6b3-5301d35309d8 service nova] [instance: 1d1b1dc4-c701-42b2-beba-344d622aef41] Refreshing instance network info cache due to event network-changed-8118febd-f949-4bed-9959-775f44cfbd07. {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 833.535918] env[61923]: DEBUG oslo_concurrency.lockutils [req-7d9f72df-38f7-4082-98ba-9f2055fcf77d req-6fc284ce-01ed-4557-b6b3-5301d35309d8 service nova] Acquiring lock "refresh_cache-1d1b1dc4-c701-42b2-beba-344d622aef41" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 833.701534] env[61923]: DEBUG nova.compute.utils [None req-9ff8b121-368d-4886-9882-962e7ef81bcf tempest-ServerTagsTestJSON-19215696 tempest-ServerTagsTestJSON-19215696-project-member] Using /dev/sd instead of None {{(pid=61923) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 833.703444] env[61923]: DEBUG nova.compute.manager [None req-9ff8b121-368d-4886-9882-962e7ef81bcf tempest-ServerTagsTestJSON-19215696 tempest-ServerTagsTestJSON-19215696-project-member] [instance: 48825592-dcdc-41d5-9fbf-500d1f31017b] Allocating IP information in the background. {{(pid=61923) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 833.703444] env[61923]: DEBUG nova.network.neutron [None req-9ff8b121-368d-4886-9882-962e7ef81bcf tempest-ServerTagsTestJSON-19215696 tempest-ServerTagsTestJSON-19215696-project-member] [instance: 48825592-dcdc-41d5-9fbf-500d1f31017b] allocate_for_instance() {{(pid=61923) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 833.708460] env[61923]: DEBUG oslo_concurrency.lockutils [None req-8e8efa20-a2ed-4da9-89f1-5ff799c0382c tempest-ServersNegativeTestMultiTenantJSON-1497120864 tempest-ServersNegativeTestMultiTenantJSON-1497120864-project-member] Releasing lock "refresh_cache-1d1b1dc4-c701-42b2-beba-344d622aef41" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 833.708843] env[61923]: DEBUG nova.compute.manager [None req-8e8efa20-a2ed-4da9-89f1-5ff799c0382c tempest-ServersNegativeTestMultiTenantJSON-1497120864 tempest-ServersNegativeTestMultiTenantJSON-1497120864-project-member] [instance: 1d1b1dc4-c701-42b2-beba-344d622aef41] Instance network_info: |[{"id": "8118febd-f949-4bed-9959-775f44cfbd07", "address": "fa:16:3e:51:b5:e2", "network": {"id": "f967adc7-af3d-438f-a4eb-e98e9dbb969d", "bridge": "br-int", "label": "tempest-ServersNegativeTestMultiTenantJSON-1189375605-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "93b2ac1779c1498d9cabeb015353024a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "604c9724-b4ef-4393-a76e-eb4a2b510796", "external-id": "nsx-vlan-transportzone-909", "segmentation_id": 909, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8118febd-f9", "ovs_interfaceid": "8118febd-f949-4bed-9959-775f44cfbd07", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61923) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 833.709760] env[61923]: DEBUG oslo_concurrency.lockutils [req-7d9f72df-38f7-4082-98ba-9f2055fcf77d req-6fc284ce-01ed-4557-b6b3-5301d35309d8 service nova] Acquired lock "refresh_cache-1d1b1dc4-c701-42b2-beba-344d622aef41" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 833.709962] env[61923]: DEBUG nova.network.neutron [req-7d9f72df-38f7-4082-98ba-9f2055fcf77d req-6fc284ce-01ed-4557-b6b3-5301d35309d8 service nova] [instance: 1d1b1dc4-c701-42b2-beba-344d622aef41] Refreshing network info cache for port 8118febd-f949-4bed-9959-775f44cfbd07 {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 833.711897] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-8e8efa20-a2ed-4da9-89f1-5ff799c0382c tempest-ServersNegativeTestMultiTenantJSON-1497120864 tempest-ServersNegativeTestMultiTenantJSON-1497120864-project-member] [instance: 1d1b1dc4-c701-42b2-beba-344d622aef41] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:51:b5:e2', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '604c9724-b4ef-4393-a76e-eb4a2b510796', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '8118febd-f949-4bed-9959-775f44cfbd07', 'vif_model': 'vmxnet3'}] {{(pid=61923) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 833.719895] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-8e8efa20-a2ed-4da9-89f1-5ff799c0382c tempest-ServersNegativeTestMultiTenantJSON-1497120864 tempest-ServersNegativeTestMultiTenantJSON-1497120864-project-member] Creating folder: Project (93b2ac1779c1498d9cabeb015353024a). Parent ref: group-v292629. {{(pid=61923) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 833.724268] env[61923]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-531260a8-fde1-492e-880c-55b5050e10ca {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.739235] env[61923]: INFO nova.virt.vmwareapi.vm_util [None req-8e8efa20-a2ed-4da9-89f1-5ff799c0382c tempest-ServersNegativeTestMultiTenantJSON-1497120864 tempest-ServersNegativeTestMultiTenantJSON-1497120864-project-member] Created folder: Project (93b2ac1779c1498d9cabeb015353024a) in parent group-v292629. [ 833.739235] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-8e8efa20-a2ed-4da9-89f1-5ff799c0382c tempest-ServersNegativeTestMultiTenantJSON-1497120864 tempest-ServersNegativeTestMultiTenantJSON-1497120864-project-member] Creating folder: Instances. Parent ref: group-v292653. {{(pid=61923) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 833.739235] env[61923]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-ea019979-27ef-4a21-95c8-5250ffd2d9c2 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.750942] env[61923]: INFO nova.virt.vmwareapi.vm_util [None req-8e8efa20-a2ed-4da9-89f1-5ff799c0382c tempest-ServersNegativeTestMultiTenantJSON-1497120864 tempest-ServersNegativeTestMultiTenantJSON-1497120864-project-member] Created folder: Instances in parent group-v292653. [ 833.751449] env[61923]: DEBUG oslo.service.loopingcall [None req-8e8efa20-a2ed-4da9-89f1-5ff799c0382c tempest-ServersNegativeTestMultiTenantJSON-1497120864 tempest-ServersNegativeTestMultiTenantJSON-1497120864-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61923) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 833.751449] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1d1b1dc4-c701-42b2-beba-344d622aef41] Creating VM on the ESX host {{(pid=61923) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 833.751687] env[61923]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ca612366-6487-4c30-a655-9d06f4af13c2 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.781274] env[61923]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 833.781274] env[61923]: value = "task-1377524" [ 833.781274] env[61923]: _type = "Task" [ 833.781274] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 833.787383] env[61923]: DEBUG nova.policy [None req-9ff8b121-368d-4886-9882-962e7ef81bcf tempest-ServerTagsTestJSON-19215696 tempest-ServerTagsTestJSON-19215696-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'bce7c7aec2434aacaacdd5d52098599d', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8e955c461948478b9a5e5d6841cc64c6', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61923) authorize /opt/stack/nova/nova/policy.py:201}} [ 833.797412] env[61923]: DEBUG oslo_vmware.api [-] Task: {'id': task-1377524, 'name': CreateVM_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.016381] env[61923]: DEBUG nova.network.neutron [req-7d9f72df-38f7-4082-98ba-9f2055fcf77d req-6fc284ce-01ed-4557-b6b3-5301d35309d8 service nova] [instance: 1d1b1dc4-c701-42b2-beba-344d622aef41] Updated VIF entry in instance network info cache for port 8118febd-f949-4bed-9959-775f44cfbd07. {{(pid=61923) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 834.016808] env[61923]: DEBUG nova.network.neutron [req-7d9f72df-38f7-4082-98ba-9f2055fcf77d req-6fc284ce-01ed-4557-b6b3-5301d35309d8 service nova] [instance: 1d1b1dc4-c701-42b2-beba-344d622aef41] Updating instance_info_cache with network_info: [{"id": "8118febd-f949-4bed-9959-775f44cfbd07", "address": "fa:16:3e:51:b5:e2", "network": {"id": "f967adc7-af3d-438f-a4eb-e98e9dbb969d", "bridge": "br-int", "label": "tempest-ServersNegativeTestMultiTenantJSON-1189375605-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "93b2ac1779c1498d9cabeb015353024a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "604c9724-b4ef-4393-a76e-eb4a2b510796", "external-id": "nsx-vlan-transportzone-909", "segmentation_id": 909, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8118febd-f9", "ovs_interfaceid": "8118febd-f949-4bed-9959-775f44cfbd07", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 834.038693] env[61923]: INFO nova.compute.manager [None req-c06a566a-3828-47dc-91c4-25d15d717c22 tempest-ServerShowV257Test-986496257 tempest-ServerShowV257Test-986496257-project-member] [instance: b8c60af7-3561-468c-a9a5-1c645080fc69] Took 21.87 seconds to build instance. [ 834.097667] env[61923]: DEBUG nova.network.neutron [None req-458eed44-c3ca-4049-bcef-fbcaf60c50ed tempest-ServersTestManualDisk-2012652654 tempest-ServersTestManualDisk-2012652654-project-member] [instance: f1839f61-4314-48fe-8ab6-14b5e49d569d] Successfully updated port: f36077f4-1c49-48e7-9c41-0f45a58883e7 {{(pid=61923) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 834.209962] env[61923]: DEBUG nova.compute.manager [None req-9ff8b121-368d-4886-9882-962e7ef81bcf tempest-ServerTagsTestJSON-19215696 tempest-ServerTagsTestJSON-19215696-project-member] [instance: 48825592-dcdc-41d5-9fbf-500d1f31017b] Start building block device mappings for instance. {{(pid=61923) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 834.248120] env[61923]: DEBUG nova.network.neutron [None req-9ff8b121-368d-4886-9882-962e7ef81bcf tempest-ServerTagsTestJSON-19215696 tempest-ServerTagsTestJSON-19215696-project-member] [instance: 48825592-dcdc-41d5-9fbf-500d1f31017b] Successfully created port: 04aea882-6981-45ab-a2d9-685c86089980 {{(pid=61923) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 834.291544] env[61923]: DEBUG oslo_vmware.api [-] Task: {'id': task-1377524, 'name': CreateVM_Task, 'duration_secs': 0.313098} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 834.291747] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 1d1b1dc4-c701-42b2-beba-344d622aef41] Created VM on the ESX host {{(pid=61923) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 834.299018] env[61923]: INFO nova.compute.manager [None req-fc756c76-ebec-403f-b8ba-39ff93058631 tempest-ServerShowV257Test-986496257 tempest-ServerShowV257Test-986496257-project-member] [instance: b8c60af7-3561-468c-a9a5-1c645080fc69] Rebuilding instance [ 834.307346] env[61923]: DEBUG oslo_concurrency.lockutils [None req-8e8efa20-a2ed-4da9-89f1-5ff799c0382c tempest-ServersNegativeTestMultiTenantJSON-1497120864 tempest-ServersNegativeTestMultiTenantJSON-1497120864-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 834.307683] env[61923]: DEBUG oslo_concurrency.lockutils [None req-8e8efa20-a2ed-4da9-89f1-5ff799c0382c tempest-ServersNegativeTestMultiTenantJSON-1497120864 tempest-ServersNegativeTestMultiTenantJSON-1497120864-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 834.307837] env[61923]: DEBUG oslo_concurrency.lockutils [None req-8e8efa20-a2ed-4da9-89f1-5ff799c0382c tempest-ServersNegativeTestMultiTenantJSON-1497120864 tempest-ServersNegativeTestMultiTenantJSON-1497120864-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 834.308112] env[61923]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-add1fc87-9fcd-40aa-8ae6-af97a0b017ff {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.312904] env[61923]: DEBUG oslo_vmware.api [None req-8e8efa20-a2ed-4da9-89f1-5ff799c0382c tempest-ServersNegativeTestMultiTenantJSON-1497120864 tempest-ServersNegativeTestMultiTenantJSON-1497120864-project-member] Waiting for the task: (returnval){ [ 834.312904] env[61923]: value = "session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52f6f56d-d982-c360-8963-89e02feac57a" [ 834.312904] env[61923]: _type = "Task" [ 834.312904] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 834.326515] env[61923]: DEBUG oslo_vmware.api [None req-8e8efa20-a2ed-4da9-89f1-5ff799c0382c tempest-ServersNegativeTestMultiTenantJSON-1497120864 tempest-ServersNegativeTestMultiTenantJSON-1497120864-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52f6f56d-d982-c360-8963-89e02feac57a, 'name': SearchDatastore_Task, 'duration_secs': 0.008988} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 834.327035] env[61923]: DEBUG oslo_concurrency.lockutils [None req-8e8efa20-a2ed-4da9-89f1-5ff799c0382c tempest-ServersNegativeTestMultiTenantJSON-1497120864 tempest-ServersNegativeTestMultiTenantJSON-1497120864-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 834.327303] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-8e8efa20-a2ed-4da9-89f1-5ff799c0382c tempest-ServersNegativeTestMultiTenantJSON-1497120864 tempest-ServersNegativeTestMultiTenantJSON-1497120864-project-member] [instance: 1d1b1dc4-c701-42b2-beba-344d622aef41] Processing image 0f153f63-ae0a-45b1-b7f5-7f9b673c947f {{(pid=61923) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 834.327537] env[61923]: DEBUG oslo_concurrency.lockutils [None req-8e8efa20-a2ed-4da9-89f1-5ff799c0382c tempest-ServersNegativeTestMultiTenantJSON-1497120864 tempest-ServersNegativeTestMultiTenantJSON-1497120864-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 834.327683] env[61923]: DEBUG oslo_concurrency.lockutils [None req-8e8efa20-a2ed-4da9-89f1-5ff799c0382c tempest-ServersNegativeTestMultiTenantJSON-1497120864 tempest-ServersNegativeTestMultiTenantJSON-1497120864-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 834.327862] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-8e8efa20-a2ed-4da9-89f1-5ff799c0382c tempest-ServersNegativeTestMultiTenantJSON-1497120864 tempest-ServersNegativeTestMultiTenantJSON-1497120864-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61923) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 834.331514] env[61923]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-995472a0-554e-4cf5-aafa-a69ad285f463 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.341448] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-8e8efa20-a2ed-4da9-89f1-5ff799c0382c tempest-ServersNegativeTestMultiTenantJSON-1497120864 tempest-ServersNegativeTestMultiTenantJSON-1497120864-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61923) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 834.341636] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-8e8efa20-a2ed-4da9-89f1-5ff799c0382c tempest-ServersNegativeTestMultiTenantJSON-1497120864 tempest-ServersNegativeTestMultiTenantJSON-1497120864-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61923) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 834.342404] env[61923]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-24e17fcd-4729-4541-b3e7-dea1eaadc374 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.351675] env[61923]: DEBUG nova.compute.manager [None req-fc756c76-ebec-403f-b8ba-39ff93058631 tempest-ServerShowV257Test-986496257 tempest-ServerShowV257Test-986496257-project-member] [instance: b8c60af7-3561-468c-a9a5-1c645080fc69] Checking state {{(pid=61923) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 834.352408] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20e6e5d4-67aa-40a6-8055-f95cf5462e5f {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.355952] env[61923]: DEBUG oslo_vmware.api [None req-8e8efa20-a2ed-4da9-89f1-5ff799c0382c tempest-ServersNegativeTestMultiTenantJSON-1497120864 tempest-ServersNegativeTestMultiTenantJSON-1497120864-project-member] Waiting for the task: (returnval){ [ 834.355952] env[61923]: value = "session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52b83cfe-0158-3408-9f41-aff9f709c445" [ 834.355952] env[61923]: _type = "Task" [ 834.355952] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 834.372720] env[61923]: DEBUG oslo_vmware.api [None req-8e8efa20-a2ed-4da9-89f1-5ff799c0382c tempest-ServersNegativeTestMultiTenantJSON-1497120864 tempest-ServersNegativeTestMultiTenantJSON-1497120864-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52b83cfe-0158-3408-9f41-aff9f709c445, 'name': SearchDatastore_Task, 'duration_secs': 0.009152} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 834.373504] env[61923]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2d8607da-4174-4893-a6d4-10a4d643e651 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.379330] env[61923]: DEBUG oslo_vmware.api [None req-8e8efa20-a2ed-4da9-89f1-5ff799c0382c tempest-ServersNegativeTestMultiTenantJSON-1497120864 tempest-ServersNegativeTestMultiTenantJSON-1497120864-project-member] Waiting for the task: (returnval){ [ 834.379330] env[61923]: value = "session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52cfd059-e4d5-239a-f4c8-a507451520b5" [ 834.379330] env[61923]: _type = "Task" [ 834.379330] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 834.388939] env[61923]: DEBUG oslo_vmware.api [None req-8e8efa20-a2ed-4da9-89f1-5ff799c0382c tempest-ServersNegativeTestMultiTenantJSON-1497120864 tempest-ServersNegativeTestMultiTenantJSON-1497120864-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52cfd059-e4d5-239a-f4c8-a507451520b5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.444867] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6881d454-7ea4-4dad-a37b-9781ddb27625 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.452682] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ad0367a-d5cc-4f49-aee8-00eb192bceb1 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.482840] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28e0941e-4c5a-4ea4-8dd0-e5c0a6bf2801 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.490351] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3e1bbec-976d-45c2-b270-1efdcba24b56 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.503747] env[61923]: DEBUG nova.compute.provider_tree [None req-950e07c2-502c-43fa-8598-da5f189b07e3 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 834.522463] env[61923]: DEBUG oslo_concurrency.lockutils [req-7d9f72df-38f7-4082-98ba-9f2055fcf77d req-6fc284ce-01ed-4557-b6b3-5301d35309d8 service nova] Releasing lock "refresh_cache-1d1b1dc4-c701-42b2-beba-344d622aef41" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 834.541249] env[61923]: DEBUG oslo_concurrency.lockutils [None req-c06a566a-3828-47dc-91c4-25d15d717c22 tempest-ServerShowV257Test-986496257 tempest-ServerShowV257Test-986496257-project-member] Lock "b8c60af7-3561-468c-a9a5-1c645080fc69" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 113.294s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 834.599947] env[61923]: DEBUG oslo_concurrency.lockutils [None req-458eed44-c3ca-4049-bcef-fbcaf60c50ed tempest-ServersTestManualDisk-2012652654 tempest-ServersTestManualDisk-2012652654-project-member] Acquiring lock "refresh_cache-f1839f61-4314-48fe-8ab6-14b5e49d569d" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 834.600165] env[61923]: DEBUG oslo_concurrency.lockutils [None req-458eed44-c3ca-4049-bcef-fbcaf60c50ed tempest-ServersTestManualDisk-2012652654 tempest-ServersTestManualDisk-2012652654-project-member] Acquired lock "refresh_cache-f1839f61-4314-48fe-8ab6-14b5e49d569d" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 834.600235] env[61923]: DEBUG nova.network.neutron [None req-458eed44-c3ca-4049-bcef-fbcaf60c50ed tempest-ServersTestManualDisk-2012652654 tempest-ServersTestManualDisk-2012652654-project-member] [instance: f1839f61-4314-48fe-8ab6-14b5e49d569d] Building network info cache for instance {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 834.870219] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-fc756c76-ebec-403f-b8ba-39ff93058631 tempest-ServerShowV257Test-986496257 tempest-ServerShowV257Test-986496257-project-member] [instance: b8c60af7-3561-468c-a9a5-1c645080fc69] Powering off the VM {{(pid=61923) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 834.870516] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6458ee32-b7b2-49ad-b1f1-eb57c7e37e56 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.877737] env[61923]: DEBUG oslo_vmware.api [None req-fc756c76-ebec-403f-b8ba-39ff93058631 tempest-ServerShowV257Test-986496257 tempest-ServerShowV257Test-986496257-project-member] Waiting for the task: (returnval){ [ 834.877737] env[61923]: value = "task-1377525" [ 834.877737] env[61923]: _type = "Task" [ 834.877737] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 834.888628] env[61923]: DEBUG oslo_vmware.api [None req-fc756c76-ebec-403f-b8ba-39ff93058631 tempest-ServerShowV257Test-986496257 tempest-ServerShowV257Test-986496257-project-member] Task: {'id': task-1377525, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 834.893955] env[61923]: DEBUG oslo_vmware.api [None req-8e8efa20-a2ed-4da9-89f1-5ff799c0382c tempest-ServersNegativeTestMultiTenantJSON-1497120864 tempest-ServersNegativeTestMultiTenantJSON-1497120864-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52cfd059-e4d5-239a-f4c8-a507451520b5, 'name': SearchDatastore_Task, 'duration_secs': 0.008816} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 834.894300] env[61923]: DEBUG oslo_concurrency.lockutils [None req-8e8efa20-a2ed-4da9-89f1-5ff799c0382c tempest-ServersNegativeTestMultiTenantJSON-1497120864 tempest-ServersNegativeTestMultiTenantJSON-1497120864-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 834.894569] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-8e8efa20-a2ed-4da9-89f1-5ff799c0382c tempest-ServersNegativeTestMultiTenantJSON-1497120864 tempest-ServersNegativeTestMultiTenantJSON-1497120864-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk to [datastore1] 1d1b1dc4-c701-42b2-beba-344d622aef41/1d1b1dc4-c701-42b2-beba-344d622aef41.vmdk {{(pid=61923) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 834.894814] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-841d9ad6-359a-4b5e-9982-ae33bedbdd75 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.900657] env[61923]: DEBUG oslo_vmware.api [None req-8e8efa20-a2ed-4da9-89f1-5ff799c0382c tempest-ServersNegativeTestMultiTenantJSON-1497120864 tempest-ServersNegativeTestMultiTenantJSON-1497120864-project-member] Waiting for the task: (returnval){ [ 834.900657] env[61923]: value = "task-1377526" [ 834.900657] env[61923]: _type = "Task" [ 834.900657] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 834.908354] env[61923]: DEBUG oslo_vmware.api [None req-8e8efa20-a2ed-4da9-89f1-5ff799c0382c tempest-ServersNegativeTestMultiTenantJSON-1497120864 tempest-ServersNegativeTestMultiTenantJSON-1497120864-project-member] Task: {'id': task-1377526, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.007152] env[61923]: DEBUG nova.scheduler.client.report [None req-950e07c2-502c-43fa-8598-da5f189b07e3 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 835.043666] env[61923]: DEBUG nova.compute.manager [None req-5fbf0568-2a85-4e3c-9b63-7d91ecc77593 tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] [instance: 2719569b-8572-4199-8158-7bb367d17dc5] Starting instance... {{(pid=61923) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 835.161724] env[61923]: DEBUG nova.network.neutron [None req-458eed44-c3ca-4049-bcef-fbcaf60c50ed tempest-ServersTestManualDisk-2012652654 tempest-ServersTestManualDisk-2012652654-project-member] [instance: f1839f61-4314-48fe-8ab6-14b5e49d569d] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 835.226779] env[61923]: DEBUG nova.compute.manager [None req-9ff8b121-368d-4886-9882-962e7ef81bcf tempest-ServerTagsTestJSON-19215696 tempest-ServerTagsTestJSON-19215696-project-member] [instance: 48825592-dcdc-41d5-9fbf-500d1f31017b] Start spawning the instance on the hypervisor. {{(pid=61923) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 835.252368] env[61923]: DEBUG nova.virt.hardware [None req-9ff8b121-368d-4886-9882-962e7ef81bcf tempest-ServerTagsTestJSON-19215696 tempest-ServerTagsTestJSON-19215696-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-29T20:07:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-29T20:07:35Z,direct_url=,disk_format='vmdk',id=0f153f63-ae0a-45b1-b7f5-7f9b673c947f,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='4e7b5e3b3c3443c8bd5c70f8a15739f5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-29T20:07:36Z,virtual_size=,visibility=), allow threads: False {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 835.252621] env[61923]: DEBUG nova.virt.hardware [None req-9ff8b121-368d-4886-9882-962e7ef81bcf tempest-ServerTagsTestJSON-19215696 tempest-ServerTagsTestJSON-19215696-project-member] Flavor limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 835.252768] env[61923]: DEBUG nova.virt.hardware [None req-9ff8b121-368d-4886-9882-962e7ef81bcf tempest-ServerTagsTestJSON-19215696 tempest-ServerTagsTestJSON-19215696-project-member] Image limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 835.252947] env[61923]: DEBUG nova.virt.hardware [None req-9ff8b121-368d-4886-9882-962e7ef81bcf tempest-ServerTagsTestJSON-19215696 tempest-ServerTagsTestJSON-19215696-project-member] Flavor pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 835.253107] env[61923]: DEBUG nova.virt.hardware [None req-9ff8b121-368d-4886-9882-962e7ef81bcf tempest-ServerTagsTestJSON-19215696 tempest-ServerTagsTestJSON-19215696-project-member] Image pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 835.253254] env[61923]: DEBUG nova.virt.hardware [None req-9ff8b121-368d-4886-9882-962e7ef81bcf tempest-ServerTagsTestJSON-19215696 tempest-ServerTagsTestJSON-19215696-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 835.253461] env[61923]: DEBUG nova.virt.hardware [None req-9ff8b121-368d-4886-9882-962e7ef81bcf tempest-ServerTagsTestJSON-19215696 tempest-ServerTagsTestJSON-19215696-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 835.253613] env[61923]: DEBUG nova.virt.hardware [None req-9ff8b121-368d-4886-9882-962e7ef81bcf tempest-ServerTagsTestJSON-19215696 tempest-ServerTagsTestJSON-19215696-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 835.253775] env[61923]: DEBUG nova.virt.hardware [None req-9ff8b121-368d-4886-9882-962e7ef81bcf tempest-ServerTagsTestJSON-19215696 tempest-ServerTagsTestJSON-19215696-project-member] Got 1 possible topologies {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 835.253934] env[61923]: DEBUG nova.virt.hardware [None req-9ff8b121-368d-4886-9882-962e7ef81bcf tempest-ServerTagsTestJSON-19215696 tempest-ServerTagsTestJSON-19215696-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 835.254148] env[61923]: DEBUG nova.virt.hardware [None req-9ff8b121-368d-4886-9882-962e7ef81bcf tempest-ServerTagsTestJSON-19215696 tempest-ServerTagsTestJSON-19215696-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 835.255047] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ad49c6a-5a0d-44ea-ae8a-3ac653e1e0ca {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.265459] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4793693-6d83-429d-b5ed-79be0a779283 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.374542] env[61923]: DEBUG nova.network.neutron [None req-458eed44-c3ca-4049-bcef-fbcaf60c50ed tempest-ServersTestManualDisk-2012652654 tempest-ServersTestManualDisk-2012652654-project-member] [instance: f1839f61-4314-48fe-8ab6-14b5e49d569d] Updating instance_info_cache with network_info: [{"id": "f36077f4-1c49-48e7-9c41-0f45a58883e7", "address": "fa:16:3e:f2:ba:61", "network": {"id": "df531efc-77d0-447e-b85e-6323fbc26dfb", "bridge": "br-int", "label": "tempest-ServersTestManualDisk-1619780476-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5357434049fb4acd910988fc5ff0d2c4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "20fbc86c-0935-406c-9659-3ffc33fc0d08", "external-id": "nsx-vlan-transportzone-191", "segmentation_id": 191, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf36077f4-1c", "ovs_interfaceid": "f36077f4-1c49-48e7-9c41-0f45a58883e7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 835.388404] env[61923]: DEBUG oslo_vmware.api [None req-fc756c76-ebec-403f-b8ba-39ff93058631 tempest-ServerShowV257Test-986496257 tempest-ServerShowV257Test-986496257-project-member] Task: {'id': task-1377525, 'name': PowerOffVM_Task, 'duration_secs': 0.176323} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 835.389466] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-fc756c76-ebec-403f-b8ba-39ff93058631 tempest-ServerShowV257Test-986496257 tempest-ServerShowV257Test-986496257-project-member] [instance: b8c60af7-3561-468c-a9a5-1c645080fc69] Powered off the VM {{(pid=61923) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 835.389793] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-fc756c76-ebec-403f-b8ba-39ff93058631 tempest-ServerShowV257Test-986496257 tempest-ServerShowV257Test-986496257-project-member] [instance: b8c60af7-3561-468c-a9a5-1c645080fc69] Destroying instance {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 835.390533] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62e7b67d-c69c-403c-bb3b-7a4e76f761cb {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.398580] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-fc756c76-ebec-403f-b8ba-39ff93058631 tempest-ServerShowV257Test-986496257 tempest-ServerShowV257Test-986496257-project-member] [instance: b8c60af7-3561-468c-a9a5-1c645080fc69] Unregistering the VM {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 835.398884] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-7a439096-17c9-42fc-b87d-2429ecdf449f {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.410511] env[61923]: DEBUG oslo_vmware.api [None req-8e8efa20-a2ed-4da9-89f1-5ff799c0382c tempest-ServersNegativeTestMultiTenantJSON-1497120864 tempest-ServersNegativeTestMultiTenantJSON-1497120864-project-member] Task: {'id': task-1377526, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.421699] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-fc756c76-ebec-403f-b8ba-39ff93058631 tempest-ServerShowV257Test-986496257 tempest-ServerShowV257Test-986496257-project-member] [instance: b8c60af7-3561-468c-a9a5-1c645080fc69] Unregistered the VM {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 835.421924] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-fc756c76-ebec-403f-b8ba-39ff93058631 tempest-ServerShowV257Test-986496257 tempest-ServerShowV257Test-986496257-project-member] [instance: b8c60af7-3561-468c-a9a5-1c645080fc69] Deleting contents of the VM from datastore datastore1 {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 835.422118] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-fc756c76-ebec-403f-b8ba-39ff93058631 tempest-ServerShowV257Test-986496257 tempest-ServerShowV257Test-986496257-project-member] Deleting the datastore file [datastore1] b8c60af7-3561-468c-a9a5-1c645080fc69 {{(pid=61923) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 835.422380] env[61923]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-fc8d6224-f467-49a0-ac62-d3fe8f82fd06 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.429086] env[61923]: DEBUG oslo_vmware.api [None req-fc756c76-ebec-403f-b8ba-39ff93058631 tempest-ServerShowV257Test-986496257 tempest-ServerShowV257Test-986496257-project-member] Waiting for the task: (returnval){ [ 835.429086] env[61923]: value = "task-1377528" [ 835.429086] env[61923]: _type = "Task" [ 835.429086] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 835.437623] env[61923]: DEBUG oslo_vmware.api [None req-fc756c76-ebec-403f-b8ba-39ff93058631 tempest-ServerShowV257Test-986496257 tempest-ServerShowV257Test-986496257-project-member] Task: {'id': task-1377528, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.512509] env[61923]: DEBUG oslo_concurrency.lockutils [None req-950e07c2-502c-43fa-8598-da5f189b07e3 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.313s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 835.513232] env[61923]: DEBUG nova.compute.manager [None req-950e07c2-502c-43fa-8598-da5f189b07e3 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] [instance: 422d06d6-6932-46c3-bb25-841e6f65c028] Start building networks asynchronously for instance. {{(pid=61923) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 835.516720] env[61923]: DEBUG oslo_concurrency.lockutils [None req-e62df7da-42b5-4263-bcf8-001a2d06427a tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.271s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 835.518631] env[61923]: INFO nova.compute.claims [None req-e62df7da-42b5-4263-bcf8-001a2d06427a tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: 73692517-1816-4e9b-ab2d-8265c683e83d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 835.562854] env[61923]: DEBUG nova.compute.manager [req-fd4d306a-a49c-4eaa-8df2-b7aa88144fa7 req-91b05c70-645d-4cf5-9979-821afeb173a0 service nova] [instance: f1839f61-4314-48fe-8ab6-14b5e49d569d] Received event network-vif-plugged-f36077f4-1c49-48e7-9c41-0f45a58883e7 {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 835.563143] env[61923]: DEBUG oslo_concurrency.lockutils [req-fd4d306a-a49c-4eaa-8df2-b7aa88144fa7 req-91b05c70-645d-4cf5-9979-821afeb173a0 service nova] Acquiring lock "f1839f61-4314-48fe-8ab6-14b5e49d569d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 835.563319] env[61923]: DEBUG oslo_concurrency.lockutils [req-fd4d306a-a49c-4eaa-8df2-b7aa88144fa7 req-91b05c70-645d-4cf5-9979-821afeb173a0 service nova] Lock "f1839f61-4314-48fe-8ab6-14b5e49d569d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 835.563486] env[61923]: DEBUG oslo_concurrency.lockutils [req-fd4d306a-a49c-4eaa-8df2-b7aa88144fa7 req-91b05c70-645d-4cf5-9979-821afeb173a0 service nova] Lock "f1839f61-4314-48fe-8ab6-14b5e49d569d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 835.563640] env[61923]: DEBUG nova.compute.manager [req-fd4d306a-a49c-4eaa-8df2-b7aa88144fa7 req-91b05c70-645d-4cf5-9979-821afeb173a0 service nova] [instance: f1839f61-4314-48fe-8ab6-14b5e49d569d] No waiting events found dispatching network-vif-plugged-f36077f4-1c49-48e7-9c41-0f45a58883e7 {{(pid=61923) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 835.563817] env[61923]: WARNING nova.compute.manager [req-fd4d306a-a49c-4eaa-8df2-b7aa88144fa7 req-91b05c70-645d-4cf5-9979-821afeb173a0 service nova] [instance: f1839f61-4314-48fe-8ab6-14b5e49d569d] Received unexpected event network-vif-plugged-f36077f4-1c49-48e7-9c41-0f45a58883e7 for instance with vm_state building and task_state spawning. [ 835.563992] env[61923]: DEBUG nova.compute.manager [req-fd4d306a-a49c-4eaa-8df2-b7aa88144fa7 req-91b05c70-645d-4cf5-9979-821afeb173a0 service nova] [instance: f1839f61-4314-48fe-8ab6-14b5e49d569d] Received event network-changed-f36077f4-1c49-48e7-9c41-0f45a58883e7 {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 835.564158] env[61923]: DEBUG nova.compute.manager [req-fd4d306a-a49c-4eaa-8df2-b7aa88144fa7 req-91b05c70-645d-4cf5-9979-821afeb173a0 service nova] [instance: f1839f61-4314-48fe-8ab6-14b5e49d569d] Refreshing instance network info cache due to event network-changed-f36077f4-1c49-48e7-9c41-0f45a58883e7. {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 835.564376] env[61923]: DEBUG oslo_concurrency.lockutils [req-fd4d306a-a49c-4eaa-8df2-b7aa88144fa7 req-91b05c70-645d-4cf5-9979-821afeb173a0 service nova] Acquiring lock "refresh_cache-f1839f61-4314-48fe-8ab6-14b5e49d569d" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 835.566480] env[61923]: DEBUG oslo_concurrency.lockutils [None req-5fbf0568-2a85-4e3c-9b63-7d91ecc77593 tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 835.877840] env[61923]: DEBUG oslo_concurrency.lockutils [None req-458eed44-c3ca-4049-bcef-fbcaf60c50ed tempest-ServersTestManualDisk-2012652654 tempest-ServersTestManualDisk-2012652654-project-member] Releasing lock "refresh_cache-f1839f61-4314-48fe-8ab6-14b5e49d569d" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 835.878211] env[61923]: DEBUG nova.compute.manager [None req-458eed44-c3ca-4049-bcef-fbcaf60c50ed tempest-ServersTestManualDisk-2012652654 tempest-ServersTestManualDisk-2012652654-project-member] [instance: f1839f61-4314-48fe-8ab6-14b5e49d569d] Instance network_info: |[{"id": "f36077f4-1c49-48e7-9c41-0f45a58883e7", "address": "fa:16:3e:f2:ba:61", "network": {"id": "df531efc-77d0-447e-b85e-6323fbc26dfb", "bridge": "br-int", "label": "tempest-ServersTestManualDisk-1619780476-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5357434049fb4acd910988fc5ff0d2c4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "20fbc86c-0935-406c-9659-3ffc33fc0d08", "external-id": "nsx-vlan-transportzone-191", "segmentation_id": 191, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf36077f4-1c", "ovs_interfaceid": "f36077f4-1c49-48e7-9c41-0f45a58883e7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61923) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 835.878666] env[61923]: DEBUG oslo_concurrency.lockutils [req-fd4d306a-a49c-4eaa-8df2-b7aa88144fa7 req-91b05c70-645d-4cf5-9979-821afeb173a0 service nova] Acquired lock "refresh_cache-f1839f61-4314-48fe-8ab6-14b5e49d569d" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 835.878946] env[61923]: DEBUG nova.network.neutron [req-fd4d306a-a49c-4eaa-8df2-b7aa88144fa7 req-91b05c70-645d-4cf5-9979-821afeb173a0 service nova] [instance: f1839f61-4314-48fe-8ab6-14b5e49d569d] Refreshing network info cache for port f36077f4-1c49-48e7-9c41-0f45a58883e7 {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 835.880291] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-458eed44-c3ca-4049-bcef-fbcaf60c50ed tempest-ServersTestManualDisk-2012652654 tempest-ServersTestManualDisk-2012652654-project-member] [instance: f1839f61-4314-48fe-8ab6-14b5e49d569d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:f2:ba:61', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '20fbc86c-0935-406c-9659-3ffc33fc0d08', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'f36077f4-1c49-48e7-9c41-0f45a58883e7', 'vif_model': 'vmxnet3'}] {{(pid=61923) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 835.893096] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-458eed44-c3ca-4049-bcef-fbcaf60c50ed tempest-ServersTestManualDisk-2012652654 tempest-ServersTestManualDisk-2012652654-project-member] Creating folder: Project (5357434049fb4acd910988fc5ff0d2c4). Parent ref: group-v292629. {{(pid=61923) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 835.898227] env[61923]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-fa066cf9-9727-4b2d-b123-0f2ed6575ff3 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.916019] env[61923]: DEBUG oslo_vmware.api [None req-8e8efa20-a2ed-4da9-89f1-5ff799c0382c tempest-ServersNegativeTestMultiTenantJSON-1497120864 tempest-ServersNegativeTestMultiTenantJSON-1497120864-project-member] Task: {'id': task-1377526, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.636126} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 835.916019] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-8e8efa20-a2ed-4da9-89f1-5ff799c0382c tempest-ServersNegativeTestMultiTenantJSON-1497120864 tempest-ServersNegativeTestMultiTenantJSON-1497120864-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk to [datastore1] 1d1b1dc4-c701-42b2-beba-344d622aef41/1d1b1dc4-c701-42b2-beba-344d622aef41.vmdk {{(pid=61923) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 835.916019] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-8e8efa20-a2ed-4da9-89f1-5ff799c0382c tempest-ServersNegativeTestMultiTenantJSON-1497120864 tempest-ServersNegativeTestMultiTenantJSON-1497120864-project-member] [instance: 1d1b1dc4-c701-42b2-beba-344d622aef41] Extending root virtual disk to 1048576 {{(pid=61923) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 835.916019] env[61923]: INFO nova.virt.vmwareapi.vm_util [None req-458eed44-c3ca-4049-bcef-fbcaf60c50ed tempest-ServersTestManualDisk-2012652654 tempest-ServersTestManualDisk-2012652654-project-member] Created folder: Project (5357434049fb4acd910988fc5ff0d2c4) in parent group-v292629. [ 835.916262] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-458eed44-c3ca-4049-bcef-fbcaf60c50ed tempest-ServersTestManualDisk-2012652654 tempest-ServersTestManualDisk-2012652654-project-member] Creating folder: Instances. Parent ref: group-v292656. {{(pid=61923) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 835.916262] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-d6a1a9c2-c924-4e05-be69-e7ad53c41faf {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.917412] env[61923]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-69011a24-c8f8-42ea-844c-7a042f377eb5 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.923955] env[61923]: DEBUG oslo_vmware.api [None req-8e8efa20-a2ed-4da9-89f1-5ff799c0382c tempest-ServersNegativeTestMultiTenantJSON-1497120864 tempest-ServersNegativeTestMultiTenantJSON-1497120864-project-member] Waiting for the task: (returnval){ [ 835.923955] env[61923]: value = "task-1377530" [ 835.923955] env[61923]: _type = "Task" [ 835.923955] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 835.927951] env[61923]: INFO nova.virt.vmwareapi.vm_util [None req-458eed44-c3ca-4049-bcef-fbcaf60c50ed tempest-ServersTestManualDisk-2012652654 tempest-ServersTestManualDisk-2012652654-project-member] Created folder: Instances in parent group-v292656. [ 835.928263] env[61923]: DEBUG oslo.service.loopingcall [None req-458eed44-c3ca-4049-bcef-fbcaf60c50ed tempest-ServersTestManualDisk-2012652654 tempest-ServersTestManualDisk-2012652654-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61923) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 835.928838] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f1839f61-4314-48fe-8ab6-14b5e49d569d] Creating VM on the ESX host {{(pid=61923) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 835.929083] env[61923]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b13703c3-41b0-4e55-90b7-60b59e189b2a {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 835.946281] env[61923]: DEBUG oslo_vmware.api [None req-8e8efa20-a2ed-4da9-89f1-5ff799c0382c tempest-ServersNegativeTestMultiTenantJSON-1497120864 tempest-ServersNegativeTestMultiTenantJSON-1497120864-project-member] Task: {'id': task-1377530, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 835.954051] env[61923]: DEBUG oslo_vmware.api [None req-fc756c76-ebec-403f-b8ba-39ff93058631 tempest-ServerShowV257Test-986496257 tempest-ServerShowV257Test-986496257-project-member] Task: {'id': task-1377528, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.185846} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 835.955219] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-fc756c76-ebec-403f-b8ba-39ff93058631 tempest-ServerShowV257Test-986496257 tempest-ServerShowV257Test-986496257-project-member] Deleted the datastore file {{(pid=61923) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 835.955415] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-fc756c76-ebec-403f-b8ba-39ff93058631 tempest-ServerShowV257Test-986496257 tempest-ServerShowV257Test-986496257-project-member] [instance: b8c60af7-3561-468c-a9a5-1c645080fc69] Deleted contents of the VM from datastore datastore1 {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 835.955583] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-fc756c76-ebec-403f-b8ba-39ff93058631 tempest-ServerShowV257Test-986496257 tempest-ServerShowV257Test-986496257-project-member] [instance: b8c60af7-3561-468c-a9a5-1c645080fc69] Instance destroyed {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 835.957925] env[61923]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 835.957925] env[61923]: value = "task-1377532" [ 835.957925] env[61923]: _type = "Task" [ 835.957925] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 835.968138] env[61923]: DEBUG oslo_vmware.api [-] Task: {'id': task-1377532, 'name': CreateVM_Task} progress is 6%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.024747] env[61923]: DEBUG nova.compute.utils [None req-950e07c2-502c-43fa-8598-da5f189b07e3 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Using /dev/sd instead of None {{(pid=61923) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 836.030390] env[61923]: DEBUG nova.compute.manager [None req-950e07c2-502c-43fa-8598-da5f189b07e3 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] [instance: 422d06d6-6932-46c3-bb25-841e6f65c028] Allocating IP information in the background. {{(pid=61923) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 836.030769] env[61923]: DEBUG nova.network.neutron [None req-950e07c2-502c-43fa-8598-da5f189b07e3 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] [instance: 422d06d6-6932-46c3-bb25-841e6f65c028] allocate_for_instance() {{(pid=61923) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 836.079367] env[61923]: DEBUG nova.policy [None req-950e07c2-502c-43fa-8598-da5f189b07e3 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'cd5039bf8f4344e490df828e643d2779', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '912300a09b80452e85e3cd13c4a644cc', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61923) authorize /opt/stack/nova/nova/policy.py:201}} [ 836.150241] env[61923]: DEBUG nova.network.neutron [None req-9ff8b121-368d-4886-9882-962e7ef81bcf tempest-ServerTagsTestJSON-19215696 tempest-ServerTagsTestJSON-19215696-project-member] [instance: 48825592-dcdc-41d5-9fbf-500d1f31017b] Successfully updated port: 04aea882-6981-45ab-a2d9-685c86089980 {{(pid=61923) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 836.232278] env[61923]: DEBUG nova.network.neutron [req-fd4d306a-a49c-4eaa-8df2-b7aa88144fa7 req-91b05c70-645d-4cf5-9979-821afeb173a0 service nova] [instance: f1839f61-4314-48fe-8ab6-14b5e49d569d] Updated VIF entry in instance network info cache for port f36077f4-1c49-48e7-9c41-0f45a58883e7. {{(pid=61923) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 836.232668] env[61923]: DEBUG nova.network.neutron [req-fd4d306a-a49c-4eaa-8df2-b7aa88144fa7 req-91b05c70-645d-4cf5-9979-821afeb173a0 service nova] [instance: f1839f61-4314-48fe-8ab6-14b5e49d569d] Updating instance_info_cache with network_info: [{"id": "f36077f4-1c49-48e7-9c41-0f45a58883e7", "address": "fa:16:3e:f2:ba:61", "network": {"id": "df531efc-77d0-447e-b85e-6323fbc26dfb", "bridge": "br-int", "label": "tempest-ServersTestManualDisk-1619780476-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5357434049fb4acd910988fc5ff0d2c4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "20fbc86c-0935-406c-9659-3ffc33fc0d08", "external-id": "nsx-vlan-transportzone-191", "segmentation_id": 191, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf36077f4-1c", "ovs_interfaceid": "f36077f4-1c49-48e7-9c41-0f45a58883e7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 836.386936] env[61923]: DEBUG nova.network.neutron [None req-950e07c2-502c-43fa-8598-da5f189b07e3 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] [instance: 422d06d6-6932-46c3-bb25-841e6f65c028] Successfully created port: 97caa1b8-427c-46b0-be7a-d59b932900c3 {{(pid=61923) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 836.435033] env[61923]: DEBUG oslo_vmware.api [None req-8e8efa20-a2ed-4da9-89f1-5ff799c0382c tempest-ServersNegativeTestMultiTenantJSON-1497120864 tempest-ServersNegativeTestMultiTenantJSON-1497120864-project-member] Task: {'id': task-1377530, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.0575} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 836.435868] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-8e8efa20-a2ed-4da9-89f1-5ff799c0382c tempest-ServersNegativeTestMultiTenantJSON-1497120864 tempest-ServersNegativeTestMultiTenantJSON-1497120864-project-member] [instance: 1d1b1dc4-c701-42b2-beba-344d622aef41] Extended root virtual disk {{(pid=61923) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 836.436488] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce043fa3-a615-495e-bdad-1a1c94bce315 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.458854] env[61923]: DEBUG nova.virt.vmwareapi.volumeops [None req-8e8efa20-a2ed-4da9-89f1-5ff799c0382c tempest-ServersNegativeTestMultiTenantJSON-1497120864 tempest-ServersNegativeTestMultiTenantJSON-1497120864-project-member] [instance: 1d1b1dc4-c701-42b2-beba-344d622aef41] Reconfiguring VM instance instance-00000040 to attach disk [datastore1] 1d1b1dc4-c701-42b2-beba-344d622aef41/1d1b1dc4-c701-42b2-beba-344d622aef41.vmdk or device None with type sparse {{(pid=61923) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 836.459161] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5cb800e6-12a4-4e73-9485-33fbbeb11c05 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.488664] env[61923]: DEBUG oslo_vmware.api [-] Task: {'id': task-1377532, 'name': CreateVM_Task, 'duration_secs': 0.363875} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 836.489871] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f1839f61-4314-48fe-8ab6-14b5e49d569d] Created VM on the ESX host {{(pid=61923) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 836.490195] env[61923]: DEBUG oslo_vmware.api [None req-8e8efa20-a2ed-4da9-89f1-5ff799c0382c tempest-ServersNegativeTestMultiTenantJSON-1497120864 tempest-ServersNegativeTestMultiTenantJSON-1497120864-project-member] Waiting for the task: (returnval){ [ 836.490195] env[61923]: value = "task-1377533" [ 836.490195] env[61923]: _type = "Task" [ 836.490195] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 836.490797] env[61923]: DEBUG oslo_concurrency.lockutils [None req-458eed44-c3ca-4049-bcef-fbcaf60c50ed tempest-ServersTestManualDisk-2012652654 tempest-ServersTestManualDisk-2012652654-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 836.490952] env[61923]: DEBUG oslo_concurrency.lockutils [None req-458eed44-c3ca-4049-bcef-fbcaf60c50ed tempest-ServersTestManualDisk-2012652654 tempest-ServersTestManualDisk-2012652654-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 836.491302] env[61923]: DEBUG oslo_concurrency.lockutils [None req-458eed44-c3ca-4049-bcef-fbcaf60c50ed tempest-ServersTestManualDisk-2012652654 tempest-ServersTestManualDisk-2012652654-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 836.491593] env[61923]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-51f7bf60-06d3-4856-ae66-239ca956626d {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.498845] env[61923]: DEBUG oslo_vmware.api [None req-458eed44-c3ca-4049-bcef-fbcaf60c50ed tempest-ServersTestManualDisk-2012652654 tempest-ServersTestManualDisk-2012652654-project-member] Waiting for the task: (returnval){ [ 836.498845] env[61923]: value = "session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52389812-5caa-fd0c-66aa-3590f0688f24" [ 836.498845] env[61923]: _type = "Task" [ 836.498845] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 836.504282] env[61923]: DEBUG oslo_vmware.api [None req-8e8efa20-a2ed-4da9-89f1-5ff799c0382c tempest-ServersNegativeTestMultiTenantJSON-1497120864 tempest-ServersNegativeTestMultiTenantJSON-1497120864-project-member] Task: {'id': task-1377533, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.513467] env[61923]: DEBUG oslo_vmware.api [None req-458eed44-c3ca-4049-bcef-fbcaf60c50ed tempest-ServersTestManualDisk-2012652654 tempest-ServersTestManualDisk-2012652654-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52389812-5caa-fd0c-66aa-3590f0688f24, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 836.531550] env[61923]: DEBUG nova.compute.manager [None req-950e07c2-502c-43fa-8598-da5f189b07e3 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] [instance: 422d06d6-6932-46c3-bb25-841e6f65c028] Start building block device mappings for instance. {{(pid=61923) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 836.653838] env[61923]: DEBUG oslo_concurrency.lockutils [None req-9ff8b121-368d-4886-9882-962e7ef81bcf tempest-ServerTagsTestJSON-19215696 tempest-ServerTagsTestJSON-19215696-project-member] Acquiring lock "refresh_cache-48825592-dcdc-41d5-9fbf-500d1f31017b" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 836.653838] env[61923]: DEBUG oslo_concurrency.lockutils [None req-9ff8b121-368d-4886-9882-962e7ef81bcf tempest-ServerTagsTestJSON-19215696 tempest-ServerTagsTestJSON-19215696-project-member] Acquired lock "refresh_cache-48825592-dcdc-41d5-9fbf-500d1f31017b" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 836.653838] env[61923]: DEBUG nova.network.neutron [None req-9ff8b121-368d-4886-9882-962e7ef81bcf tempest-ServerTagsTestJSON-19215696 tempest-ServerTagsTestJSON-19215696-project-member] [instance: 48825592-dcdc-41d5-9fbf-500d1f31017b] Building network info cache for instance {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 836.736146] env[61923]: DEBUG oslo_concurrency.lockutils [req-fd4d306a-a49c-4eaa-8df2-b7aa88144fa7 req-91b05c70-645d-4cf5-9979-821afeb173a0 service nova] Releasing lock "refresh_cache-f1839f61-4314-48fe-8ab6-14b5e49d569d" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 836.781682] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60a9f93e-5506-4661-921a-fa69ee6d691b {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.789605] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b15c662f-a10c-444b-8c04-fb46fd8a4248 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.821844] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6fa0492e-ab93-4cc1-9541-f5cb0f1552a3 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.829331] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c94d682-d82b-4548-937b-3b9e310a9563 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.843556] env[61923]: DEBUG nova.compute.provider_tree [None req-e62df7da-42b5-4263-bcf8-001a2d06427a tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 837.001661] env[61923]: DEBUG oslo_vmware.api [None req-8e8efa20-a2ed-4da9-89f1-5ff799c0382c tempest-ServersNegativeTestMultiTenantJSON-1497120864 tempest-ServersNegativeTestMultiTenantJSON-1497120864-project-member] Task: {'id': task-1377533, 'name': ReconfigVM_Task, 'duration_secs': 0.269922} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 837.001661] env[61923]: DEBUG nova.virt.vmwareapi.volumeops [None req-8e8efa20-a2ed-4da9-89f1-5ff799c0382c tempest-ServersNegativeTestMultiTenantJSON-1497120864 tempest-ServersNegativeTestMultiTenantJSON-1497120864-project-member] [instance: 1d1b1dc4-c701-42b2-beba-344d622aef41] Reconfigured VM instance instance-00000040 to attach disk [datastore1] 1d1b1dc4-c701-42b2-beba-344d622aef41/1d1b1dc4-c701-42b2-beba-344d622aef41.vmdk or device None with type sparse {{(pid=61923) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 837.002046] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c117a2a6-2831-4427-a4cc-2449812e304b {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.010555] env[61923]: DEBUG oslo_vmware.api [None req-8e8efa20-a2ed-4da9-89f1-5ff799c0382c tempest-ServersNegativeTestMultiTenantJSON-1497120864 tempest-ServersNegativeTestMultiTenantJSON-1497120864-project-member] Waiting for the task: (returnval){ [ 837.010555] env[61923]: value = "task-1377534" [ 837.010555] env[61923]: _type = "Task" [ 837.010555] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 837.017033] env[61923]: DEBUG nova.virt.hardware [None req-fc756c76-ebec-403f-b8ba-39ff93058631 tempest-ServerShowV257Test-986496257 tempest-ServerShowV257Test-986496257-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-29T20:07:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-29T20:07:35Z,direct_url=,disk_format='vmdk',id=0f153f63-ae0a-45b1-b7f5-7f9b673c947f,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='4e7b5e3b3c3443c8bd5c70f8a15739f5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-29T20:07:36Z,virtual_size=,visibility=), allow threads: False {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 837.017033] env[61923]: DEBUG nova.virt.hardware [None req-fc756c76-ebec-403f-b8ba-39ff93058631 tempest-ServerShowV257Test-986496257 tempest-ServerShowV257Test-986496257-project-member] Flavor limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 837.017033] env[61923]: DEBUG nova.virt.hardware [None req-fc756c76-ebec-403f-b8ba-39ff93058631 tempest-ServerShowV257Test-986496257 tempest-ServerShowV257Test-986496257-project-member] Image limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 837.017033] env[61923]: DEBUG nova.virt.hardware [None req-fc756c76-ebec-403f-b8ba-39ff93058631 tempest-ServerShowV257Test-986496257 tempest-ServerShowV257Test-986496257-project-member] Flavor pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 837.017284] env[61923]: DEBUG nova.virt.hardware [None req-fc756c76-ebec-403f-b8ba-39ff93058631 tempest-ServerShowV257Test-986496257 tempest-ServerShowV257Test-986496257-project-member] Image pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 837.017284] env[61923]: DEBUG nova.virt.hardware [None req-fc756c76-ebec-403f-b8ba-39ff93058631 tempest-ServerShowV257Test-986496257 tempest-ServerShowV257Test-986496257-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 837.017284] env[61923]: DEBUG nova.virt.hardware [None req-fc756c76-ebec-403f-b8ba-39ff93058631 tempest-ServerShowV257Test-986496257 tempest-ServerShowV257Test-986496257-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 837.017284] env[61923]: DEBUG nova.virt.hardware [None req-fc756c76-ebec-403f-b8ba-39ff93058631 tempest-ServerShowV257Test-986496257 tempest-ServerShowV257Test-986496257-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 837.017284] env[61923]: DEBUG nova.virt.hardware [None req-fc756c76-ebec-403f-b8ba-39ff93058631 tempest-ServerShowV257Test-986496257 tempest-ServerShowV257Test-986496257-project-member] Got 1 possible topologies {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 837.017441] env[61923]: DEBUG nova.virt.hardware [None req-fc756c76-ebec-403f-b8ba-39ff93058631 tempest-ServerShowV257Test-986496257 tempest-ServerShowV257Test-986496257-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 837.017441] env[61923]: DEBUG nova.virt.hardware [None req-fc756c76-ebec-403f-b8ba-39ff93058631 tempest-ServerShowV257Test-986496257 tempest-ServerShowV257Test-986496257-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 837.019964] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ec12518-14d4-4884-8f88-1a8ae854d213 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.022597] env[61923]: DEBUG oslo_vmware.api [None req-458eed44-c3ca-4049-bcef-fbcaf60c50ed tempest-ServersTestManualDisk-2012652654 tempest-ServersTestManualDisk-2012652654-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52389812-5caa-fd0c-66aa-3590f0688f24, 'name': SearchDatastore_Task, 'duration_secs': 0.012992} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 837.025500] env[61923]: DEBUG oslo_concurrency.lockutils [None req-458eed44-c3ca-4049-bcef-fbcaf60c50ed tempest-ServersTestManualDisk-2012652654 tempest-ServersTestManualDisk-2012652654-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 837.025832] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-458eed44-c3ca-4049-bcef-fbcaf60c50ed tempest-ServersTestManualDisk-2012652654 tempest-ServersTestManualDisk-2012652654-project-member] [instance: f1839f61-4314-48fe-8ab6-14b5e49d569d] Processing image 0f153f63-ae0a-45b1-b7f5-7f9b673c947f {{(pid=61923) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 837.026233] env[61923]: DEBUG oslo_concurrency.lockutils [None req-458eed44-c3ca-4049-bcef-fbcaf60c50ed tempest-ServersTestManualDisk-2012652654 tempest-ServersTestManualDisk-2012652654-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 837.026510] env[61923]: DEBUG oslo_concurrency.lockutils [None req-458eed44-c3ca-4049-bcef-fbcaf60c50ed tempest-ServersTestManualDisk-2012652654 tempest-ServersTestManualDisk-2012652654-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 837.026793] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-458eed44-c3ca-4049-bcef-fbcaf60c50ed tempest-ServersTestManualDisk-2012652654 tempest-ServersTestManualDisk-2012652654-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61923) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 837.027540] env[61923]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-08292b6f-882f-4934-8834-69b9009b4226 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.035476] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a33e386-d51b-4172-b38c-219c1a03d123 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.047965] env[61923]: DEBUG oslo_vmware.api [None req-8e8efa20-a2ed-4da9-89f1-5ff799c0382c tempest-ServersNegativeTestMultiTenantJSON-1497120864 tempest-ServersNegativeTestMultiTenantJSON-1497120864-project-member] Task: {'id': task-1377534, 'name': Rename_Task} progress is 14%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.058288] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-fc756c76-ebec-403f-b8ba-39ff93058631 tempest-ServerShowV257Test-986496257 tempest-ServerShowV257Test-986496257-project-member] [instance: b8c60af7-3561-468c-a9a5-1c645080fc69] Instance VIF info [] {{(pid=61923) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 837.063757] env[61923]: DEBUG oslo.service.loopingcall [None req-fc756c76-ebec-403f-b8ba-39ff93058631 tempest-ServerShowV257Test-986496257 tempest-ServerShowV257Test-986496257-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61923) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 837.064797] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b8c60af7-3561-468c-a9a5-1c645080fc69] Creating VM on the ESX host {{(pid=61923) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 837.065081] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-458eed44-c3ca-4049-bcef-fbcaf60c50ed tempest-ServersTestManualDisk-2012652654 tempest-ServersTestManualDisk-2012652654-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61923) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 837.065247] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-458eed44-c3ca-4049-bcef-fbcaf60c50ed tempest-ServersTestManualDisk-2012652654 tempest-ServersTestManualDisk-2012652654-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61923) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 837.065898] env[61923]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d63655dd-0ff4-49c8-9c4c-729ff7f2c144 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.077120] env[61923]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0c0ebc9b-de9f-47ac-9060-532db3f31f14 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.082611] env[61923]: DEBUG oslo_vmware.api [None req-458eed44-c3ca-4049-bcef-fbcaf60c50ed tempest-ServersTestManualDisk-2012652654 tempest-ServersTestManualDisk-2012652654-project-member] Waiting for the task: (returnval){ [ 837.082611] env[61923]: value = "session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]524a7faf-1417-8b32-2db1-ee66f392c1b3" [ 837.082611] env[61923]: _type = "Task" [ 837.082611] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 837.086839] env[61923]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 837.086839] env[61923]: value = "task-1377535" [ 837.086839] env[61923]: _type = "Task" [ 837.086839] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 837.093215] env[61923]: DEBUG oslo_vmware.api [None req-458eed44-c3ca-4049-bcef-fbcaf60c50ed tempest-ServersTestManualDisk-2012652654 tempest-ServersTestManualDisk-2012652654-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]524a7faf-1417-8b32-2db1-ee66f392c1b3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.097905] env[61923]: DEBUG oslo_vmware.api [-] Task: {'id': task-1377535, 'name': CreateVM_Task} progress is 5%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.203240] env[61923]: DEBUG nova.network.neutron [None req-9ff8b121-368d-4886-9882-962e7ef81bcf tempest-ServerTagsTestJSON-19215696 tempest-ServerTagsTestJSON-19215696-project-member] [instance: 48825592-dcdc-41d5-9fbf-500d1f31017b] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 837.346953] env[61923]: DEBUG nova.scheduler.client.report [None req-e62df7da-42b5-4263-bcf8-001a2d06427a tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 837.370937] env[61923]: DEBUG nova.network.neutron [None req-9ff8b121-368d-4886-9882-962e7ef81bcf tempest-ServerTagsTestJSON-19215696 tempest-ServerTagsTestJSON-19215696-project-member] [instance: 48825592-dcdc-41d5-9fbf-500d1f31017b] Updating instance_info_cache with network_info: [{"id": "04aea882-6981-45ab-a2d9-685c86089980", "address": "fa:16:3e:90:83:d0", "network": {"id": "739e14c3-4cfb-4507-8880-b608b8528ad5", "bridge": "br-int", "label": "tempest-ServerTagsTestJSON-1880597985-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8e955c461948478b9a5e5d6841cc64c6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "31e77685-b4dd-4810-80ef-24115ea9ea62", "external-id": "nsx-vlan-transportzone-56", "segmentation_id": 56, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap04aea882-69", "ovs_interfaceid": "04aea882-6981-45ab-a2d9-685c86089980", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 837.528700] env[61923]: DEBUG oslo_vmware.api [None req-8e8efa20-a2ed-4da9-89f1-5ff799c0382c tempest-ServersNegativeTestMultiTenantJSON-1497120864 tempest-ServersNegativeTestMultiTenantJSON-1497120864-project-member] Task: {'id': task-1377534, 'name': Rename_Task, 'duration_secs': 0.141631} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 837.528987] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-8e8efa20-a2ed-4da9-89f1-5ff799c0382c tempest-ServersNegativeTestMultiTenantJSON-1497120864 tempest-ServersNegativeTestMultiTenantJSON-1497120864-project-member] [instance: 1d1b1dc4-c701-42b2-beba-344d622aef41] Powering on the VM {{(pid=61923) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 837.529205] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b13d80f4-2658-4466-a74e-b7438883263e {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.535444] env[61923]: DEBUG oslo_vmware.api [None req-8e8efa20-a2ed-4da9-89f1-5ff799c0382c tempest-ServersNegativeTestMultiTenantJSON-1497120864 tempest-ServersNegativeTestMultiTenantJSON-1497120864-project-member] Waiting for the task: (returnval){ [ 837.535444] env[61923]: value = "task-1377536" [ 837.535444] env[61923]: _type = "Task" [ 837.535444] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 837.544536] env[61923]: DEBUG oslo_vmware.api [None req-8e8efa20-a2ed-4da9-89f1-5ff799c0382c tempest-ServersNegativeTestMultiTenantJSON-1497120864 tempest-ServersNegativeTestMultiTenantJSON-1497120864-project-member] Task: {'id': task-1377536, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.550831] env[61923]: DEBUG nova.compute.manager [None req-950e07c2-502c-43fa-8598-da5f189b07e3 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] [instance: 422d06d6-6932-46c3-bb25-841e6f65c028] Start spawning the instance on the hypervisor. {{(pid=61923) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 837.573089] env[61923]: DEBUG nova.virt.hardware [None req-950e07c2-502c-43fa-8598-da5f189b07e3 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-29T20:07:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-29T20:07:35Z,direct_url=,disk_format='vmdk',id=0f153f63-ae0a-45b1-b7f5-7f9b673c947f,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='4e7b5e3b3c3443c8bd5c70f8a15739f5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-29T20:07:36Z,virtual_size=,visibility=), allow threads: False {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 837.573359] env[61923]: DEBUG nova.virt.hardware [None req-950e07c2-502c-43fa-8598-da5f189b07e3 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Flavor limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 837.573515] env[61923]: DEBUG nova.virt.hardware [None req-950e07c2-502c-43fa-8598-da5f189b07e3 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Image limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 837.573690] env[61923]: DEBUG nova.virt.hardware [None req-950e07c2-502c-43fa-8598-da5f189b07e3 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Flavor pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 837.573832] env[61923]: DEBUG nova.virt.hardware [None req-950e07c2-502c-43fa-8598-da5f189b07e3 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Image pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 837.573980] env[61923]: DEBUG nova.virt.hardware [None req-950e07c2-502c-43fa-8598-da5f189b07e3 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 837.574231] env[61923]: DEBUG nova.virt.hardware [None req-950e07c2-502c-43fa-8598-da5f189b07e3 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 837.574405] env[61923]: DEBUG nova.virt.hardware [None req-950e07c2-502c-43fa-8598-da5f189b07e3 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 837.574575] env[61923]: DEBUG nova.virt.hardware [None req-950e07c2-502c-43fa-8598-da5f189b07e3 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Got 1 possible topologies {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 837.574736] env[61923]: DEBUG nova.virt.hardware [None req-950e07c2-502c-43fa-8598-da5f189b07e3 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 837.574953] env[61923]: DEBUG nova.virt.hardware [None req-950e07c2-502c-43fa-8598-da5f189b07e3 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 837.576064] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bdf9220d-14cb-470b-959b-f9912bc3df3f {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.583571] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37ede960-efd3-4e4d-8ae9-215bf399fe6b {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.595471] env[61923]: DEBUG nova.compute.manager [req-6a9d4023-8331-4236-bbf6-f010fc1229bc req-3dc6dfe8-86c7-4fb1-8b62-ffdf51cee4a6 service nova] [instance: 48825592-dcdc-41d5-9fbf-500d1f31017b] Received event network-vif-plugged-04aea882-6981-45ab-a2d9-685c86089980 {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 837.595688] env[61923]: DEBUG oslo_concurrency.lockutils [req-6a9d4023-8331-4236-bbf6-f010fc1229bc req-3dc6dfe8-86c7-4fb1-8b62-ffdf51cee4a6 service nova] Acquiring lock "48825592-dcdc-41d5-9fbf-500d1f31017b-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 837.595899] env[61923]: DEBUG oslo_concurrency.lockutils [req-6a9d4023-8331-4236-bbf6-f010fc1229bc req-3dc6dfe8-86c7-4fb1-8b62-ffdf51cee4a6 service nova] Lock "48825592-dcdc-41d5-9fbf-500d1f31017b-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 837.596076] env[61923]: DEBUG oslo_concurrency.lockutils [req-6a9d4023-8331-4236-bbf6-f010fc1229bc req-3dc6dfe8-86c7-4fb1-8b62-ffdf51cee4a6 service nova] Lock "48825592-dcdc-41d5-9fbf-500d1f31017b-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 837.596246] env[61923]: DEBUG nova.compute.manager [req-6a9d4023-8331-4236-bbf6-f010fc1229bc req-3dc6dfe8-86c7-4fb1-8b62-ffdf51cee4a6 service nova] [instance: 48825592-dcdc-41d5-9fbf-500d1f31017b] No waiting events found dispatching network-vif-plugged-04aea882-6981-45ab-a2d9-685c86089980 {{(pid=61923) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 837.596406] env[61923]: WARNING nova.compute.manager [req-6a9d4023-8331-4236-bbf6-f010fc1229bc req-3dc6dfe8-86c7-4fb1-8b62-ffdf51cee4a6 service nova] [instance: 48825592-dcdc-41d5-9fbf-500d1f31017b] Received unexpected event network-vif-plugged-04aea882-6981-45ab-a2d9-685c86089980 for instance with vm_state building and task_state spawning. [ 837.596559] env[61923]: DEBUG nova.compute.manager [req-6a9d4023-8331-4236-bbf6-f010fc1229bc req-3dc6dfe8-86c7-4fb1-8b62-ffdf51cee4a6 service nova] [instance: 48825592-dcdc-41d5-9fbf-500d1f31017b] Received event network-changed-04aea882-6981-45ab-a2d9-685c86089980 {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 837.596708] env[61923]: DEBUG nova.compute.manager [req-6a9d4023-8331-4236-bbf6-f010fc1229bc req-3dc6dfe8-86c7-4fb1-8b62-ffdf51cee4a6 service nova] [instance: 48825592-dcdc-41d5-9fbf-500d1f31017b] Refreshing instance network info cache due to event network-changed-04aea882-6981-45ab-a2d9-685c86089980. {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 837.596894] env[61923]: DEBUG oslo_concurrency.lockutils [req-6a9d4023-8331-4236-bbf6-f010fc1229bc req-3dc6dfe8-86c7-4fb1-8b62-ffdf51cee4a6 service nova] Acquiring lock "refresh_cache-48825592-dcdc-41d5-9fbf-500d1f31017b" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 837.609679] env[61923]: DEBUG oslo_vmware.api [None req-458eed44-c3ca-4049-bcef-fbcaf60c50ed tempest-ServersTestManualDisk-2012652654 tempest-ServersTestManualDisk-2012652654-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]524a7faf-1417-8b32-2db1-ee66f392c1b3, 'name': SearchDatastore_Task, 'duration_secs': 0.012882} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 837.613135] env[61923]: DEBUG oslo_vmware.api [-] Task: {'id': task-1377535, 'name': CreateVM_Task, 'duration_secs': 0.257915} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 837.613338] env[61923]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-68302258-d3b5-44b5-92f8-a456264dcd51 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.615374] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b8c60af7-3561-468c-a9a5-1c645080fc69] Created VM on the ESX host {{(pid=61923) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 837.615982] env[61923]: DEBUG oslo_concurrency.lockutils [None req-fc756c76-ebec-403f-b8ba-39ff93058631 tempest-ServerShowV257Test-986496257 tempest-ServerShowV257Test-986496257-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 837.616160] env[61923]: DEBUG oslo_concurrency.lockutils [None req-fc756c76-ebec-403f-b8ba-39ff93058631 tempest-ServerShowV257Test-986496257 tempest-ServerShowV257Test-986496257-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 837.616465] env[61923]: DEBUG oslo_concurrency.lockutils [None req-fc756c76-ebec-403f-b8ba-39ff93058631 tempest-ServerShowV257Test-986496257 tempest-ServerShowV257Test-986496257-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 837.616981] env[61923]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-dcb1eb9a-007d-45bd-ba4b-7d0ddd5e1b7e {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.619357] env[61923]: DEBUG oslo_vmware.api [None req-458eed44-c3ca-4049-bcef-fbcaf60c50ed tempest-ServersTestManualDisk-2012652654 tempest-ServersTestManualDisk-2012652654-project-member] Waiting for the task: (returnval){ [ 837.619357] env[61923]: value = "session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52fba138-9d1e-9807-b825-f1229b02b865" [ 837.619357] env[61923]: _type = "Task" [ 837.619357] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 837.623581] env[61923]: DEBUG oslo_vmware.api [None req-fc756c76-ebec-403f-b8ba-39ff93058631 tempest-ServerShowV257Test-986496257 tempest-ServerShowV257Test-986496257-project-member] Waiting for the task: (returnval){ [ 837.623581] env[61923]: value = "session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]526daeb1-56c5-8fcb-b81a-d99580fe8e30" [ 837.623581] env[61923]: _type = "Task" [ 837.623581] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 837.630822] env[61923]: DEBUG oslo_vmware.api [None req-458eed44-c3ca-4049-bcef-fbcaf60c50ed tempest-ServersTestManualDisk-2012652654 tempest-ServersTestManualDisk-2012652654-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52fba138-9d1e-9807-b825-f1229b02b865, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.635175] env[61923]: DEBUG oslo_vmware.api [None req-fc756c76-ebec-403f-b8ba-39ff93058631 tempest-ServerShowV257Test-986496257 tempest-ServerShowV257Test-986496257-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]526daeb1-56c5-8fcb-b81a-d99580fe8e30, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 837.858419] env[61923]: DEBUG oslo_concurrency.lockutils [None req-e62df7da-42b5-4263-bcf8-001a2d06427a tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.338s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 837.858419] env[61923]: DEBUG nova.compute.manager [None req-e62df7da-42b5-4263-bcf8-001a2d06427a tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: 73692517-1816-4e9b-ab2d-8265c683e83d] Start building networks asynchronously for instance. {{(pid=61923) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 837.861313] env[61923]: DEBUG oslo_concurrency.lockutils [None req-513daa64-2b05-4c0d-94e9-55b95de63eec tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.836s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 837.864326] env[61923]: INFO nova.compute.claims [None req-513daa64-2b05-4c0d-94e9-55b95de63eec tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] [instance: 59198f4d-4dde-4eaf-9f6c-a962cbe53c6e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 837.881031] env[61923]: DEBUG oslo_concurrency.lockutils [None req-9ff8b121-368d-4886-9882-962e7ef81bcf tempest-ServerTagsTestJSON-19215696 tempest-ServerTagsTestJSON-19215696-project-member] Releasing lock "refresh_cache-48825592-dcdc-41d5-9fbf-500d1f31017b" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 837.881031] env[61923]: DEBUG nova.compute.manager [None req-9ff8b121-368d-4886-9882-962e7ef81bcf tempest-ServerTagsTestJSON-19215696 tempest-ServerTagsTestJSON-19215696-project-member] [instance: 48825592-dcdc-41d5-9fbf-500d1f31017b] Instance network_info: |[{"id": "04aea882-6981-45ab-a2d9-685c86089980", "address": "fa:16:3e:90:83:d0", "network": {"id": "739e14c3-4cfb-4507-8880-b608b8528ad5", "bridge": "br-int", "label": "tempest-ServerTagsTestJSON-1880597985-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8e955c461948478b9a5e5d6841cc64c6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "31e77685-b4dd-4810-80ef-24115ea9ea62", "external-id": "nsx-vlan-transportzone-56", "segmentation_id": 56, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap04aea882-69", "ovs_interfaceid": "04aea882-6981-45ab-a2d9-685c86089980", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61923) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 837.881449] env[61923]: DEBUG oslo_concurrency.lockutils [req-6a9d4023-8331-4236-bbf6-f010fc1229bc req-3dc6dfe8-86c7-4fb1-8b62-ffdf51cee4a6 service nova] Acquired lock "refresh_cache-48825592-dcdc-41d5-9fbf-500d1f31017b" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 837.881449] env[61923]: DEBUG nova.network.neutron [req-6a9d4023-8331-4236-bbf6-f010fc1229bc req-3dc6dfe8-86c7-4fb1-8b62-ffdf51cee4a6 service nova] [instance: 48825592-dcdc-41d5-9fbf-500d1f31017b] Refreshing network info cache for port 04aea882-6981-45ab-a2d9-685c86089980 {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 837.881449] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-9ff8b121-368d-4886-9882-962e7ef81bcf tempest-ServerTagsTestJSON-19215696 tempest-ServerTagsTestJSON-19215696-project-member] [instance: 48825592-dcdc-41d5-9fbf-500d1f31017b] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:90:83:d0', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '31e77685-b4dd-4810-80ef-24115ea9ea62', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '04aea882-6981-45ab-a2d9-685c86089980', 'vif_model': 'vmxnet3'}] {{(pid=61923) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 837.892844] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-9ff8b121-368d-4886-9882-962e7ef81bcf tempest-ServerTagsTestJSON-19215696 tempest-ServerTagsTestJSON-19215696-project-member] Creating folder: Project (8e955c461948478b9a5e5d6841cc64c6). Parent ref: group-v292629. {{(pid=61923) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 837.893586] env[61923]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-15049283-09ff-4721-932f-52229727a8e9 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.904641] env[61923]: INFO nova.virt.vmwareapi.vm_util [None req-9ff8b121-368d-4886-9882-962e7ef81bcf tempest-ServerTagsTestJSON-19215696 tempest-ServerTagsTestJSON-19215696-project-member] Created folder: Project (8e955c461948478b9a5e5d6841cc64c6) in parent group-v292629. [ 837.905369] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-9ff8b121-368d-4886-9882-962e7ef81bcf tempest-ServerTagsTestJSON-19215696 tempest-ServerTagsTestJSON-19215696-project-member] Creating folder: Instances. Parent ref: group-v292660. {{(pid=61923) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 837.905735] env[61923]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-f66785b3-e69b-4362-8741-2522ca271f57 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.916135] env[61923]: INFO nova.virt.vmwareapi.vm_util [None req-9ff8b121-368d-4886-9882-962e7ef81bcf tempest-ServerTagsTestJSON-19215696 tempest-ServerTagsTestJSON-19215696-project-member] Created folder: Instances in parent group-v292660. [ 837.916769] env[61923]: DEBUG oslo.service.loopingcall [None req-9ff8b121-368d-4886-9882-962e7ef81bcf tempest-ServerTagsTestJSON-19215696 tempest-ServerTagsTestJSON-19215696-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61923) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 837.917098] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 48825592-dcdc-41d5-9fbf-500d1f31017b] Creating VM on the ESX host {{(pid=61923) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 837.917417] env[61923]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e908fb84-7f79-477a-8fdb-5a3b36bd9cac {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 837.943764] env[61923]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 837.943764] env[61923]: value = "task-1377539" [ 837.943764] env[61923]: _type = "Task" [ 837.943764] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 837.953289] env[61923]: DEBUG oslo_vmware.api [-] Task: {'id': task-1377539, 'name': CreateVM_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.046646] env[61923]: DEBUG oslo_vmware.api [None req-8e8efa20-a2ed-4da9-89f1-5ff799c0382c tempest-ServersNegativeTestMultiTenantJSON-1497120864 tempest-ServersNegativeTestMultiTenantJSON-1497120864-project-member] Task: {'id': task-1377536, 'name': PowerOnVM_Task, 'duration_secs': 0.44188} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 838.047685] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-8e8efa20-a2ed-4da9-89f1-5ff799c0382c tempest-ServersNegativeTestMultiTenantJSON-1497120864 tempest-ServersNegativeTestMultiTenantJSON-1497120864-project-member] [instance: 1d1b1dc4-c701-42b2-beba-344d622aef41] Powered on the VM {{(pid=61923) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 838.047685] env[61923]: INFO nova.compute.manager [None req-8e8efa20-a2ed-4da9-89f1-5ff799c0382c tempest-ServersNegativeTestMultiTenantJSON-1497120864 tempest-ServersNegativeTestMultiTenantJSON-1497120864-project-member] [instance: 1d1b1dc4-c701-42b2-beba-344d622aef41] Took 7.53 seconds to spawn the instance on the hypervisor. [ 838.047685] env[61923]: DEBUG nova.compute.manager [None req-8e8efa20-a2ed-4da9-89f1-5ff799c0382c tempest-ServersNegativeTestMultiTenantJSON-1497120864 tempest-ServersNegativeTestMultiTenantJSON-1497120864-project-member] [instance: 1d1b1dc4-c701-42b2-beba-344d622aef41] Checking state {{(pid=61923) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 838.048151] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87e18a83-aa79-4e4b-bd0d-be7dc1360d43 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.095530] env[61923]: DEBUG nova.compute.manager [req-d481d878-c7fd-4d20-9613-827ed51916d0 req-997dd5e9-3c34-4064-8ae4-99d7d4cd950d service nova] [instance: 422d06d6-6932-46c3-bb25-841e6f65c028] Received event network-vif-plugged-97caa1b8-427c-46b0-be7a-d59b932900c3 {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 838.095762] env[61923]: DEBUG oslo_concurrency.lockutils [req-d481d878-c7fd-4d20-9613-827ed51916d0 req-997dd5e9-3c34-4064-8ae4-99d7d4cd950d service nova] Acquiring lock "422d06d6-6932-46c3-bb25-841e6f65c028-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 838.095965] env[61923]: DEBUG oslo_concurrency.lockutils [req-d481d878-c7fd-4d20-9613-827ed51916d0 req-997dd5e9-3c34-4064-8ae4-99d7d4cd950d service nova] Lock "422d06d6-6932-46c3-bb25-841e6f65c028-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 838.096154] env[61923]: DEBUG oslo_concurrency.lockutils [req-d481d878-c7fd-4d20-9613-827ed51916d0 req-997dd5e9-3c34-4064-8ae4-99d7d4cd950d service nova] Lock "422d06d6-6932-46c3-bb25-841e6f65c028-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 838.096342] env[61923]: DEBUG nova.compute.manager [req-d481d878-c7fd-4d20-9613-827ed51916d0 req-997dd5e9-3c34-4064-8ae4-99d7d4cd950d service nova] [instance: 422d06d6-6932-46c3-bb25-841e6f65c028] No waiting events found dispatching network-vif-plugged-97caa1b8-427c-46b0-be7a-d59b932900c3 {{(pid=61923) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 838.096559] env[61923]: WARNING nova.compute.manager [req-d481d878-c7fd-4d20-9613-827ed51916d0 req-997dd5e9-3c34-4064-8ae4-99d7d4cd950d service nova] [instance: 422d06d6-6932-46c3-bb25-841e6f65c028] Received unexpected event network-vif-plugged-97caa1b8-427c-46b0-be7a-d59b932900c3 for instance with vm_state building and task_state spawning. [ 838.115275] env[61923]: DEBUG nova.network.neutron [None req-950e07c2-502c-43fa-8598-da5f189b07e3 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] [instance: 422d06d6-6932-46c3-bb25-841e6f65c028] Successfully updated port: 97caa1b8-427c-46b0-be7a-d59b932900c3 {{(pid=61923) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 838.135591] env[61923]: DEBUG oslo_vmware.api [None req-458eed44-c3ca-4049-bcef-fbcaf60c50ed tempest-ServersTestManualDisk-2012652654 tempest-ServersTestManualDisk-2012652654-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52fba138-9d1e-9807-b825-f1229b02b865, 'name': SearchDatastore_Task, 'duration_secs': 0.010689} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 838.139257] env[61923]: DEBUG oslo_concurrency.lockutils [None req-458eed44-c3ca-4049-bcef-fbcaf60c50ed tempest-ServersTestManualDisk-2012652654 tempest-ServersTestManualDisk-2012652654-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 838.139606] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-458eed44-c3ca-4049-bcef-fbcaf60c50ed tempest-ServersTestManualDisk-2012652654 tempest-ServersTestManualDisk-2012652654-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk to [datastore1] f1839f61-4314-48fe-8ab6-14b5e49d569d/f1839f61-4314-48fe-8ab6-14b5e49d569d.vmdk {{(pid=61923) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 838.139907] env[61923]: DEBUG oslo_vmware.api [None req-fc756c76-ebec-403f-b8ba-39ff93058631 tempest-ServerShowV257Test-986496257 tempest-ServerShowV257Test-986496257-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]526daeb1-56c5-8fcb-b81a-d99580fe8e30, 'name': SearchDatastore_Task, 'duration_secs': 0.01673} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 838.140403] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-0948c2d8-fd60-4a40-931c-0ec96e892de7 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.143137] env[61923]: DEBUG oslo_concurrency.lockutils [None req-fc756c76-ebec-403f-b8ba-39ff93058631 tempest-ServerShowV257Test-986496257 tempest-ServerShowV257Test-986496257-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 838.143364] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-fc756c76-ebec-403f-b8ba-39ff93058631 tempest-ServerShowV257Test-986496257 tempest-ServerShowV257Test-986496257-project-member] [instance: b8c60af7-3561-468c-a9a5-1c645080fc69] Processing image 0f153f63-ae0a-45b1-b7f5-7f9b673c947f {{(pid=61923) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 838.143583] env[61923]: DEBUG oslo_concurrency.lockutils [None req-fc756c76-ebec-403f-b8ba-39ff93058631 tempest-ServerShowV257Test-986496257 tempest-ServerShowV257Test-986496257-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 838.143727] env[61923]: DEBUG oslo_concurrency.lockutils [None req-fc756c76-ebec-403f-b8ba-39ff93058631 tempest-ServerShowV257Test-986496257 tempest-ServerShowV257Test-986496257-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 838.143896] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-fc756c76-ebec-403f-b8ba-39ff93058631 tempest-ServerShowV257Test-986496257 tempest-ServerShowV257Test-986496257-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61923) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 838.144153] env[61923]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e3aa064d-94ac-4c0d-83b1-1bf0258c9880 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.150509] env[61923]: DEBUG oslo_vmware.api [None req-458eed44-c3ca-4049-bcef-fbcaf60c50ed tempest-ServersTestManualDisk-2012652654 tempest-ServersTestManualDisk-2012652654-project-member] Waiting for the task: (returnval){ [ 838.150509] env[61923]: value = "task-1377540" [ 838.150509] env[61923]: _type = "Task" [ 838.150509] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 838.154258] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-fc756c76-ebec-403f-b8ba-39ff93058631 tempest-ServerShowV257Test-986496257 tempest-ServerShowV257Test-986496257-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61923) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 838.154430] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-fc756c76-ebec-403f-b8ba-39ff93058631 tempest-ServerShowV257Test-986496257 tempest-ServerShowV257Test-986496257-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61923) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 838.155393] env[61923]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9cef811b-987a-4060-b96f-e1b14ff71a34 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.160453] env[61923]: DEBUG oslo_vmware.api [None req-458eed44-c3ca-4049-bcef-fbcaf60c50ed tempest-ServersTestManualDisk-2012652654 tempest-ServersTestManualDisk-2012652654-project-member] Task: {'id': task-1377540, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.163385] env[61923]: DEBUG oslo_vmware.api [None req-fc756c76-ebec-403f-b8ba-39ff93058631 tempest-ServerShowV257Test-986496257 tempest-ServerShowV257Test-986496257-project-member] Waiting for the task: (returnval){ [ 838.163385] env[61923]: value = "session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]524fb07f-d1d0-e480-5e3d-baed57cb8bb1" [ 838.163385] env[61923]: _type = "Task" [ 838.163385] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 838.171923] env[61923]: DEBUG oslo_vmware.api [None req-fc756c76-ebec-403f-b8ba-39ff93058631 tempest-ServerShowV257Test-986496257 tempest-ServerShowV257Test-986496257-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]524fb07f-d1d0-e480-5e3d-baed57cb8bb1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.371311] env[61923]: DEBUG nova.compute.utils [None req-e62df7da-42b5-4263-bcf8-001a2d06427a tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Using /dev/sd instead of None {{(pid=61923) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 838.375840] env[61923]: DEBUG nova.compute.manager [None req-e62df7da-42b5-4263-bcf8-001a2d06427a tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: 73692517-1816-4e9b-ab2d-8265c683e83d] Allocating IP information in the background. {{(pid=61923) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 838.375993] env[61923]: DEBUG nova.network.neutron [None req-e62df7da-42b5-4263-bcf8-001a2d06427a tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: 73692517-1816-4e9b-ab2d-8265c683e83d] allocate_for_instance() {{(pid=61923) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 838.422448] env[61923]: DEBUG nova.policy [None req-e62df7da-42b5-4263-bcf8-001a2d06427a tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5b8a639e99914e75857b4571f0d58a3c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '69cc941eb6dd4780ac12aa29656c37f7', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61923) authorize /opt/stack/nova/nova/policy.py:201}} [ 838.455395] env[61923]: DEBUG oslo_vmware.api [-] Task: {'id': task-1377539, 'name': CreateVM_Task, 'duration_secs': 0.344397} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 838.455584] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 48825592-dcdc-41d5-9fbf-500d1f31017b] Created VM on the ESX host {{(pid=61923) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 838.456396] env[61923]: DEBUG oslo_concurrency.lockutils [None req-9ff8b121-368d-4886-9882-962e7ef81bcf tempest-ServerTagsTestJSON-19215696 tempest-ServerTagsTestJSON-19215696-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 838.456591] env[61923]: DEBUG oslo_concurrency.lockutils [None req-9ff8b121-368d-4886-9882-962e7ef81bcf tempest-ServerTagsTestJSON-19215696 tempest-ServerTagsTestJSON-19215696-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 838.456958] env[61923]: DEBUG oslo_concurrency.lockutils [None req-9ff8b121-368d-4886-9882-962e7ef81bcf tempest-ServerTagsTestJSON-19215696 tempest-ServerTagsTestJSON-19215696-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 838.457630] env[61923]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f199bfe9-0d7d-40c0-84f1-fae58cbb2932 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.462645] env[61923]: DEBUG oslo_vmware.api [None req-9ff8b121-368d-4886-9882-962e7ef81bcf tempest-ServerTagsTestJSON-19215696 tempest-ServerTagsTestJSON-19215696-project-member] Waiting for the task: (returnval){ [ 838.462645] env[61923]: value = "session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52fa0b4b-7243-5e7e-551b-3e4fe4ead9dc" [ 838.462645] env[61923]: _type = "Task" [ 838.462645] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 838.471518] env[61923]: DEBUG oslo_vmware.api [None req-9ff8b121-368d-4886-9882-962e7ef81bcf tempest-ServerTagsTestJSON-19215696 tempest-ServerTagsTestJSON-19215696-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52fa0b4b-7243-5e7e-551b-3e4fe4ead9dc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.567959] env[61923]: INFO nova.compute.manager [None req-8e8efa20-a2ed-4da9-89f1-5ff799c0382c tempest-ServersNegativeTestMultiTenantJSON-1497120864 tempest-ServersNegativeTestMultiTenantJSON-1497120864-project-member] [instance: 1d1b1dc4-c701-42b2-beba-344d622aef41] Took 24.69 seconds to build instance. [ 838.618214] env[61923]: DEBUG oslo_concurrency.lockutils [None req-950e07c2-502c-43fa-8598-da5f189b07e3 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Acquiring lock "refresh_cache-422d06d6-6932-46c3-bb25-841e6f65c028" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 838.618324] env[61923]: DEBUG oslo_concurrency.lockutils [None req-950e07c2-502c-43fa-8598-da5f189b07e3 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Acquired lock "refresh_cache-422d06d6-6932-46c3-bb25-841e6f65c028" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 838.618474] env[61923]: DEBUG nova.network.neutron [None req-950e07c2-502c-43fa-8598-da5f189b07e3 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] [instance: 422d06d6-6932-46c3-bb25-841e6f65c028] Building network info cache for instance {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 838.630564] env[61923]: DEBUG nova.network.neutron [req-6a9d4023-8331-4236-bbf6-f010fc1229bc req-3dc6dfe8-86c7-4fb1-8b62-ffdf51cee4a6 service nova] [instance: 48825592-dcdc-41d5-9fbf-500d1f31017b] Updated VIF entry in instance network info cache for port 04aea882-6981-45ab-a2d9-685c86089980. {{(pid=61923) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 838.630721] env[61923]: DEBUG nova.network.neutron [req-6a9d4023-8331-4236-bbf6-f010fc1229bc req-3dc6dfe8-86c7-4fb1-8b62-ffdf51cee4a6 service nova] [instance: 48825592-dcdc-41d5-9fbf-500d1f31017b] Updating instance_info_cache with network_info: [{"id": "04aea882-6981-45ab-a2d9-685c86089980", "address": "fa:16:3e:90:83:d0", "network": {"id": "739e14c3-4cfb-4507-8880-b608b8528ad5", "bridge": "br-int", "label": "tempest-ServerTagsTestJSON-1880597985-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8e955c461948478b9a5e5d6841cc64c6", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "31e77685-b4dd-4810-80ef-24115ea9ea62", "external-id": "nsx-vlan-transportzone-56", "segmentation_id": 56, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap04aea882-69", "ovs_interfaceid": "04aea882-6981-45ab-a2d9-685c86089980", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 838.661747] env[61923]: DEBUG oslo_vmware.api [None req-458eed44-c3ca-4049-bcef-fbcaf60c50ed tempest-ServersTestManualDisk-2012652654 tempest-ServersTestManualDisk-2012652654-project-member] Task: {'id': task-1377540, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.674701] env[61923]: DEBUG oslo_vmware.api [None req-fc756c76-ebec-403f-b8ba-39ff93058631 tempest-ServerShowV257Test-986496257 tempest-ServerShowV257Test-986496257-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]524fb07f-d1d0-e480-5e3d-baed57cb8bb1, 'name': SearchDatastore_Task, 'duration_secs': 0.009181} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 838.676294] env[61923]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-be4639e9-89e9-4953-8603-97ed76d507b0 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.682113] env[61923]: DEBUG oslo_vmware.api [None req-fc756c76-ebec-403f-b8ba-39ff93058631 tempest-ServerShowV257Test-986496257 tempest-ServerShowV257Test-986496257-project-member] Waiting for the task: (returnval){ [ 838.682113] env[61923]: value = "session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]526ab346-5506-ee87-3462-3daa022923a8" [ 838.682113] env[61923]: _type = "Task" [ 838.682113] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 838.691536] env[61923]: DEBUG oslo_vmware.api [None req-fc756c76-ebec-403f-b8ba-39ff93058631 tempest-ServerShowV257Test-986496257 tempest-ServerShowV257Test-986496257-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]526ab346-5506-ee87-3462-3daa022923a8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 838.704030] env[61923]: DEBUG nova.network.neutron [None req-e62df7da-42b5-4263-bcf8-001a2d06427a tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: 73692517-1816-4e9b-ab2d-8265c683e83d] Successfully created port: dbed34c7-9716-4c24-8f0e-2298de6cf4b9 {{(pid=61923) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 838.811228] env[61923]: DEBUG oslo_service.periodic_task [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61923) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 838.811531] env[61923]: DEBUG oslo_service.periodic_task [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=61923) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 838.876897] env[61923]: DEBUG nova.compute.manager [None req-e62df7da-42b5-4263-bcf8-001a2d06427a tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: 73692517-1816-4e9b-ab2d-8265c683e83d] Start building block device mappings for instance. {{(pid=61923) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 838.974919] env[61923]: DEBUG oslo_vmware.api [None req-9ff8b121-368d-4886-9882-962e7ef81bcf tempest-ServerTagsTestJSON-19215696 tempest-ServerTagsTestJSON-19215696-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52fa0b4b-7243-5e7e-551b-3e4fe4ead9dc, 'name': SearchDatastore_Task, 'duration_secs': 0.065657} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 838.975136] env[61923]: DEBUG oslo_concurrency.lockutils [None req-9ff8b121-368d-4886-9882-962e7ef81bcf tempest-ServerTagsTestJSON-19215696 tempest-ServerTagsTestJSON-19215696-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 838.975528] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-9ff8b121-368d-4886-9882-962e7ef81bcf tempest-ServerTagsTestJSON-19215696 tempest-ServerTagsTestJSON-19215696-project-member] [instance: 48825592-dcdc-41d5-9fbf-500d1f31017b] Processing image 0f153f63-ae0a-45b1-b7f5-7f9b673c947f {{(pid=61923) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 838.975609] env[61923]: DEBUG oslo_concurrency.lockutils [None req-9ff8b121-368d-4886-9882-962e7ef81bcf tempest-ServerTagsTestJSON-19215696 tempest-ServerTagsTestJSON-19215696-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 838.975686] env[61923]: DEBUG oslo_concurrency.lockutils [None req-9ff8b121-368d-4886-9882-962e7ef81bcf tempest-ServerTagsTestJSON-19215696 tempest-ServerTagsTestJSON-19215696-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 838.975853] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-9ff8b121-368d-4886-9882-962e7ef81bcf tempest-ServerTagsTestJSON-19215696 tempest-ServerTagsTestJSON-19215696-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61923) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 838.976122] env[61923]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-3e8920bc-5064-4a24-80c4-95cd8205e43f {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.991835] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-9ff8b121-368d-4886-9882-962e7ef81bcf tempest-ServerTagsTestJSON-19215696 tempest-ServerTagsTestJSON-19215696-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61923) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 838.991835] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-9ff8b121-368d-4886-9882-962e7ef81bcf tempest-ServerTagsTestJSON-19215696 tempest-ServerTagsTestJSON-19215696-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61923) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 838.994609] env[61923]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-af006667-5025-4158-9495-dd82a11350f5 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.999864] env[61923]: DEBUG oslo_vmware.api [None req-9ff8b121-368d-4886-9882-962e7ef81bcf tempest-ServerTagsTestJSON-19215696 tempest-ServerTagsTestJSON-19215696-project-member] Waiting for the task: (returnval){ [ 838.999864] env[61923]: value = "session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]522777cd-469d-9dc0-41dc-3d01bfdae9c5" [ 838.999864] env[61923]: _type = "Task" [ 838.999864] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 839.007375] env[61923]: DEBUG oslo_vmware.api [None req-9ff8b121-368d-4886-9882-962e7ef81bcf tempest-ServerTagsTestJSON-19215696 tempest-ServerTagsTestJSON-19215696-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]522777cd-469d-9dc0-41dc-3d01bfdae9c5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.070517] env[61923]: DEBUG oslo_concurrency.lockutils [None req-8e8efa20-a2ed-4da9-89f1-5ff799c0382c tempest-ServersNegativeTestMultiTenantJSON-1497120864 tempest-ServersNegativeTestMultiTenantJSON-1497120864-project-member] Lock "1d1b1dc4-c701-42b2-beba-344d622aef41" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 101.200s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 839.102948] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30acad05-16fd-4763-a893-b581077b0add {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.110780] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15e19c08-d770-4670-809a-b52741a0fd28 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.144464] env[61923]: DEBUG oslo_concurrency.lockutils [req-6a9d4023-8331-4236-bbf6-f010fc1229bc req-3dc6dfe8-86c7-4fb1-8b62-ffdf51cee4a6 service nova] Releasing lock "refresh_cache-48825592-dcdc-41d5-9fbf-500d1f31017b" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 839.145617] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0057d12-08dd-4b33-9db7-dcf8bbd12c40 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.156269] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fcbcc3fb-9f07-4dad-8299-f276102802ec {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.172786] env[61923]: DEBUG nova.compute.provider_tree [None req-513daa64-2b05-4c0d-94e9-55b95de63eec tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 839.174111] env[61923]: DEBUG oslo_vmware.api [None req-458eed44-c3ca-4049-bcef-fbcaf60c50ed tempest-ServersTestManualDisk-2012652654 tempest-ServersTestManualDisk-2012652654-project-member] Task: {'id': task-1377540, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.592271} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 839.174810] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-458eed44-c3ca-4049-bcef-fbcaf60c50ed tempest-ServersTestManualDisk-2012652654 tempest-ServersTestManualDisk-2012652654-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk to [datastore1] f1839f61-4314-48fe-8ab6-14b5e49d569d/f1839f61-4314-48fe-8ab6-14b5e49d569d.vmdk {{(pid=61923) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 839.174974] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-458eed44-c3ca-4049-bcef-fbcaf60c50ed tempest-ServersTestManualDisk-2012652654 tempest-ServersTestManualDisk-2012652654-project-member] [instance: f1839f61-4314-48fe-8ab6-14b5e49d569d] Extending root virtual disk to 1048576 {{(pid=61923) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 839.175693] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-2dbed801-deb3-4b22-91f3-3bb184f52813 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.177602] env[61923]: DEBUG nova.network.neutron [None req-950e07c2-502c-43fa-8598-da5f189b07e3 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] [instance: 422d06d6-6932-46c3-bb25-841e6f65c028] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 839.186717] env[61923]: DEBUG oslo_vmware.api [None req-458eed44-c3ca-4049-bcef-fbcaf60c50ed tempest-ServersTestManualDisk-2012652654 tempest-ServersTestManualDisk-2012652654-project-member] Waiting for the task: (returnval){ [ 839.186717] env[61923]: value = "task-1377541" [ 839.186717] env[61923]: _type = "Task" [ 839.186717] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 839.193534] env[61923]: DEBUG oslo_vmware.api [None req-fc756c76-ebec-403f-b8ba-39ff93058631 tempest-ServerShowV257Test-986496257 tempest-ServerShowV257Test-986496257-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]526ab346-5506-ee87-3462-3daa022923a8, 'name': SearchDatastore_Task, 'duration_secs': 0.022315} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 839.194161] env[61923]: DEBUG oslo_concurrency.lockutils [None req-fc756c76-ebec-403f-b8ba-39ff93058631 tempest-ServerShowV257Test-986496257 tempest-ServerShowV257Test-986496257-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 839.194453] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-fc756c76-ebec-403f-b8ba-39ff93058631 tempest-ServerShowV257Test-986496257 tempest-ServerShowV257Test-986496257-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk to [datastore2] b8c60af7-3561-468c-a9a5-1c645080fc69/b8c60af7-3561-468c-a9a5-1c645080fc69.vmdk {{(pid=61923) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 839.194714] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ec14949f-a79f-44cc-bd79-438a1295908c {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.199146] env[61923]: DEBUG oslo_vmware.api [None req-458eed44-c3ca-4049-bcef-fbcaf60c50ed tempest-ServersTestManualDisk-2012652654 tempest-ServersTestManualDisk-2012652654-project-member] Task: {'id': task-1377541, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.203709] env[61923]: DEBUG oslo_vmware.api [None req-fc756c76-ebec-403f-b8ba-39ff93058631 tempest-ServerShowV257Test-986496257 tempest-ServerShowV257Test-986496257-project-member] Waiting for the task: (returnval){ [ 839.203709] env[61923]: value = "task-1377542" [ 839.203709] env[61923]: _type = "Task" [ 839.203709] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 839.212975] env[61923]: DEBUG oslo_vmware.api [None req-fc756c76-ebec-403f-b8ba-39ff93058631 tempest-ServerShowV257Test-986496257 tempest-ServerShowV257Test-986496257-project-member] Task: {'id': task-1377542, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.319380] env[61923]: DEBUG oslo_service.periodic_task [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61923) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 839.319380] env[61923]: DEBUG nova.compute.manager [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Starting heal instance info cache {{(pid=61923) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 839.319380] env[61923]: DEBUG nova.compute.manager [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Rebuilding the list of instances to heal {{(pid=61923) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 839.418822] env[61923]: DEBUG nova.network.neutron [None req-950e07c2-502c-43fa-8598-da5f189b07e3 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] [instance: 422d06d6-6932-46c3-bb25-841e6f65c028] Updating instance_info_cache with network_info: [{"id": "97caa1b8-427c-46b0-be7a-d59b932900c3", "address": "fa:16:3e:fa:f8:d1", "network": {"id": "a8b7e1b4-1ec6-4226-bdd6-535dbce450d2", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1928758125-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "912300a09b80452e85e3cd13c4a644cc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7806fe18-2b89-4386-87b1-f22876f82af2", "external-id": "nsx-vlan-transportzone-727", "segmentation_id": 727, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap97caa1b8-42", "ovs_interfaceid": "97caa1b8-427c-46b0-be7a-d59b932900c3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 839.509872] env[61923]: DEBUG oslo_vmware.api [None req-9ff8b121-368d-4886-9882-962e7ef81bcf tempest-ServerTagsTestJSON-19215696 tempest-ServerTagsTestJSON-19215696-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]522777cd-469d-9dc0-41dc-3d01bfdae9c5, 'name': SearchDatastore_Task, 'duration_secs': 0.042582} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 839.511187] env[61923]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e0e25707-2a06-45c0-bdf9-a3f38438f718 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.516889] env[61923]: DEBUG oslo_vmware.api [None req-9ff8b121-368d-4886-9882-962e7ef81bcf tempest-ServerTagsTestJSON-19215696 tempest-ServerTagsTestJSON-19215696-project-member] Waiting for the task: (returnval){ [ 839.516889] env[61923]: value = "session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52183c7c-f7e4-ec70-de52-86047f4eee0a" [ 839.516889] env[61923]: _type = "Task" [ 839.516889] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 839.524482] env[61923]: DEBUG oslo_vmware.api [None req-9ff8b121-368d-4886-9882-962e7ef81bcf tempest-ServerTagsTestJSON-19215696 tempest-ServerTagsTestJSON-19215696-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52183c7c-f7e4-ec70-de52-86047f4eee0a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.573665] env[61923]: DEBUG nova.compute.manager [None req-cdbe0145-cd3c-4551-af87-0c7f8b53f26c tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] [instance: 444affa5-a7ed-4a17-9015-9fd5724aab64] Starting instance... {{(pid=61923) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 839.675898] env[61923]: DEBUG nova.scheduler.client.report [None req-513daa64-2b05-4c0d-94e9-55b95de63eec tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 839.697905] env[61923]: DEBUG oslo_vmware.api [None req-458eed44-c3ca-4049-bcef-fbcaf60c50ed tempest-ServersTestManualDisk-2012652654 tempest-ServersTestManualDisk-2012652654-project-member] Task: {'id': task-1377541, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.157872} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 839.698196] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-458eed44-c3ca-4049-bcef-fbcaf60c50ed tempest-ServersTestManualDisk-2012652654 tempest-ServersTestManualDisk-2012652654-project-member] [instance: f1839f61-4314-48fe-8ab6-14b5e49d569d] Extended root virtual disk {{(pid=61923) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 839.698972] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88f5df68-c4fe-4834-94a3-ac3b547dec71 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.722165] env[61923]: DEBUG nova.virt.vmwareapi.volumeops [None req-458eed44-c3ca-4049-bcef-fbcaf60c50ed tempest-ServersTestManualDisk-2012652654 tempest-ServersTestManualDisk-2012652654-project-member] [instance: f1839f61-4314-48fe-8ab6-14b5e49d569d] Reconfiguring VM instance instance-00000041 to attach disk [datastore1] f1839f61-4314-48fe-8ab6-14b5e49d569d/f1839f61-4314-48fe-8ab6-14b5e49d569d.vmdk or device None with type sparse {{(pid=61923) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 839.725247] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-70a95b01-6e0e-4acb-b5d5-393a38e1e2aa {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.745685] env[61923]: DEBUG oslo_vmware.api [None req-fc756c76-ebec-403f-b8ba-39ff93058631 tempest-ServerShowV257Test-986496257 tempest-ServerShowV257Test-986496257-project-member] Task: {'id': task-1377542, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.747144] env[61923]: DEBUG oslo_vmware.api [None req-458eed44-c3ca-4049-bcef-fbcaf60c50ed tempest-ServersTestManualDisk-2012652654 tempest-ServersTestManualDisk-2012652654-project-member] Waiting for the task: (returnval){ [ 839.747144] env[61923]: value = "task-1377543" [ 839.747144] env[61923]: _type = "Task" [ 839.747144] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 839.755627] env[61923]: DEBUG oslo_vmware.api [None req-458eed44-c3ca-4049-bcef-fbcaf60c50ed tempest-ServersTestManualDisk-2012652654 tempest-ServersTestManualDisk-2012652654-project-member] Task: {'id': task-1377543, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 839.823216] env[61923]: DEBUG nova.compute.manager [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] [instance: f1839f61-4314-48fe-8ab6-14b5e49d569d] Skipping network cache update for instance because it is Building. {{(pid=61923) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 839.823456] env[61923]: DEBUG nova.compute.manager [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] [instance: 48825592-dcdc-41d5-9fbf-500d1f31017b] Skipping network cache update for instance because it is Building. {{(pid=61923) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 839.823573] env[61923]: DEBUG nova.compute.manager [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] [instance: 422d06d6-6932-46c3-bb25-841e6f65c028] Skipping network cache update for instance because it is Building. {{(pid=61923) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 839.823643] env[61923]: DEBUG nova.compute.manager [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] [instance: 73692517-1816-4e9b-ab2d-8265c683e83d] Skipping network cache update for instance because it is Building. {{(pid=61923) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 839.823786] env[61923]: DEBUG nova.compute.manager [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] [instance: 59198f4d-4dde-4eaf-9f6c-a962cbe53c6e] Skipping network cache update for instance because it is Building. {{(pid=61923) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 839.823990] env[61923]: DEBUG oslo_concurrency.lockutils [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Acquiring lock "refresh_cache-b8c60af7-3561-468c-a9a5-1c645080fc69" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 839.824116] env[61923]: DEBUG oslo_concurrency.lockutils [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Acquired lock "refresh_cache-b8c60af7-3561-468c-a9a5-1c645080fc69" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 839.824252] env[61923]: DEBUG nova.network.neutron [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] [instance: b8c60af7-3561-468c-a9a5-1c645080fc69] Forcefully refreshing network info cache for instance {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 839.824396] env[61923]: DEBUG nova.objects.instance [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Lazy-loading 'info_cache' on Instance uuid b8c60af7-3561-468c-a9a5-1c645080fc69 {{(pid=61923) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 839.890295] env[61923]: DEBUG nova.compute.manager [None req-e62df7da-42b5-4263-bcf8-001a2d06427a tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: 73692517-1816-4e9b-ab2d-8265c683e83d] Start spawning the instance on the hypervisor. {{(pid=61923) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 839.921287] env[61923]: DEBUG nova.virt.hardware [None req-e62df7da-42b5-4263-bcf8-001a2d06427a tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-29T20:07:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-29T20:07:35Z,direct_url=,disk_format='vmdk',id=0f153f63-ae0a-45b1-b7f5-7f9b673c947f,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='4e7b5e3b3c3443c8bd5c70f8a15739f5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-29T20:07:36Z,virtual_size=,visibility=), allow threads: False {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 839.921620] env[61923]: DEBUG nova.virt.hardware [None req-e62df7da-42b5-4263-bcf8-001a2d06427a tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Flavor limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 839.921795] env[61923]: DEBUG nova.virt.hardware [None req-e62df7da-42b5-4263-bcf8-001a2d06427a tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Image limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 839.921966] env[61923]: DEBUG nova.virt.hardware [None req-e62df7da-42b5-4263-bcf8-001a2d06427a tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Flavor pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 839.922137] env[61923]: DEBUG nova.virt.hardware [None req-e62df7da-42b5-4263-bcf8-001a2d06427a tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Image pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 839.922275] env[61923]: DEBUG nova.virt.hardware [None req-e62df7da-42b5-4263-bcf8-001a2d06427a tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 839.922485] env[61923]: DEBUG nova.virt.hardware [None req-e62df7da-42b5-4263-bcf8-001a2d06427a tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 839.922645] env[61923]: DEBUG nova.virt.hardware [None req-e62df7da-42b5-4263-bcf8-001a2d06427a tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 839.922812] env[61923]: DEBUG nova.virt.hardware [None req-e62df7da-42b5-4263-bcf8-001a2d06427a tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Got 1 possible topologies {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 839.923041] env[61923]: DEBUG nova.virt.hardware [None req-e62df7da-42b5-4263-bcf8-001a2d06427a tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 839.923262] env[61923]: DEBUG nova.virt.hardware [None req-e62df7da-42b5-4263-bcf8-001a2d06427a tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 839.923783] env[61923]: DEBUG oslo_concurrency.lockutils [None req-950e07c2-502c-43fa-8598-da5f189b07e3 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Releasing lock "refresh_cache-422d06d6-6932-46c3-bb25-841e6f65c028" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 839.924050] env[61923]: DEBUG nova.compute.manager [None req-950e07c2-502c-43fa-8598-da5f189b07e3 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] [instance: 422d06d6-6932-46c3-bb25-841e6f65c028] Instance network_info: |[{"id": "97caa1b8-427c-46b0-be7a-d59b932900c3", "address": "fa:16:3e:fa:f8:d1", "network": {"id": "a8b7e1b4-1ec6-4226-bdd6-535dbce450d2", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1928758125-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "912300a09b80452e85e3cd13c4a644cc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7806fe18-2b89-4386-87b1-f22876f82af2", "external-id": "nsx-vlan-transportzone-727", "segmentation_id": 727, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap97caa1b8-42", "ovs_interfaceid": "97caa1b8-427c-46b0-be7a-d59b932900c3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61923) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 839.924881] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f236041-de3a-4631-844c-ba9687e47db5 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.927764] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-950e07c2-502c-43fa-8598-da5f189b07e3 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] [instance: 422d06d6-6932-46c3-bb25-841e6f65c028] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:fa:f8:d1', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '7806fe18-2b89-4386-87b1-f22876f82af2', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '97caa1b8-427c-46b0-be7a-d59b932900c3', 'vif_model': 'vmxnet3'}] {{(pid=61923) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 839.935249] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-950e07c2-502c-43fa-8598-da5f189b07e3 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Creating folder: Project (912300a09b80452e85e3cd13c4a644cc). Parent ref: group-v292629. {{(pid=61923) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 839.935959] env[61923]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-761464d8-0471-4afb-9c10-d173dbca910e {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.943599] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-111b9765-6692-48ff-94c2-25ba309f6c4f {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.948886] env[61923]: INFO nova.virt.vmwareapi.vm_util [None req-950e07c2-502c-43fa-8598-da5f189b07e3 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Created folder: Project (912300a09b80452e85e3cd13c4a644cc) in parent group-v292629. [ 839.949147] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-950e07c2-502c-43fa-8598-da5f189b07e3 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Creating folder: Instances. Parent ref: group-v292663. {{(pid=61923) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 839.949807] env[61923]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-cdbe17be-7844-48f8-9cc6-75d22a5cb245 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.974122] env[61923]: INFO nova.virt.vmwareapi.vm_util [None req-950e07c2-502c-43fa-8598-da5f189b07e3 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Created folder: Instances in parent group-v292663. [ 839.974403] env[61923]: DEBUG oslo.service.loopingcall [None req-950e07c2-502c-43fa-8598-da5f189b07e3 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61923) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 839.974611] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 422d06d6-6932-46c3-bb25-841e6f65c028] Creating VM on the ESX host {{(pid=61923) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 839.974832] env[61923]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d6b8d62c-d80b-4cc2-89cd-6ea89f87ede3 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 839.997763] env[61923]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 839.997763] env[61923]: value = "task-1377546" [ 839.997763] env[61923]: _type = "Task" [ 839.997763] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 840.008447] env[61923]: DEBUG oslo_vmware.api [-] Task: {'id': task-1377546, 'name': CreateVM_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.029412] env[61923]: DEBUG oslo_vmware.api [None req-9ff8b121-368d-4886-9882-962e7ef81bcf tempest-ServerTagsTestJSON-19215696 tempest-ServerTagsTestJSON-19215696-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52183c7c-f7e4-ec70-de52-86047f4eee0a, 'name': SearchDatastore_Task, 'duration_secs': 0.046606} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 840.029689] env[61923]: DEBUG oslo_concurrency.lockutils [None req-9ff8b121-368d-4886-9882-962e7ef81bcf tempest-ServerTagsTestJSON-19215696 tempest-ServerTagsTestJSON-19215696-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 840.029953] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-9ff8b121-368d-4886-9882-962e7ef81bcf tempest-ServerTagsTestJSON-19215696 tempest-ServerTagsTestJSON-19215696-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk to [datastore1] 48825592-dcdc-41d5-9fbf-500d1f31017b/48825592-dcdc-41d5-9fbf-500d1f31017b.vmdk {{(pid=61923) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 840.030250] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-6068101d-1037-4f17-8517-70192ad26fbe {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.041715] env[61923]: DEBUG oslo_vmware.api [None req-9ff8b121-368d-4886-9882-962e7ef81bcf tempest-ServerTagsTestJSON-19215696 tempest-ServerTagsTestJSON-19215696-project-member] Waiting for the task: (returnval){ [ 840.041715] env[61923]: value = "task-1377547" [ 840.041715] env[61923]: _type = "Task" [ 840.041715] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 840.052790] env[61923]: DEBUG oslo_vmware.api [None req-9ff8b121-368d-4886-9882-962e7ef81bcf tempest-ServerTagsTestJSON-19215696 tempest-ServerTagsTestJSON-19215696-project-member] Task: {'id': task-1377547, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.102208] env[61923]: DEBUG oslo_concurrency.lockutils [None req-cdbe0145-cd3c-4551-af87-0c7f8b53f26c tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 840.121932] env[61923]: DEBUG nova.compute.manager [req-1a317816-99f5-40ae-b198-6b8c03c4b429 req-4044a0f3-8ed2-491a-ae3d-441bda5b5a28 service nova] [instance: 422d06d6-6932-46c3-bb25-841e6f65c028] Received event network-changed-97caa1b8-427c-46b0-be7a-d59b932900c3 {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 840.121932] env[61923]: DEBUG nova.compute.manager [req-1a317816-99f5-40ae-b198-6b8c03c4b429 req-4044a0f3-8ed2-491a-ae3d-441bda5b5a28 service nova] [instance: 422d06d6-6932-46c3-bb25-841e6f65c028] Refreshing instance network info cache due to event network-changed-97caa1b8-427c-46b0-be7a-d59b932900c3. {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 840.122225] env[61923]: DEBUG oslo_concurrency.lockutils [req-1a317816-99f5-40ae-b198-6b8c03c4b429 req-4044a0f3-8ed2-491a-ae3d-441bda5b5a28 service nova] Acquiring lock "refresh_cache-422d06d6-6932-46c3-bb25-841e6f65c028" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 840.122447] env[61923]: DEBUG oslo_concurrency.lockutils [req-1a317816-99f5-40ae-b198-6b8c03c4b429 req-4044a0f3-8ed2-491a-ae3d-441bda5b5a28 service nova] Acquired lock "refresh_cache-422d06d6-6932-46c3-bb25-841e6f65c028" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 840.122627] env[61923]: DEBUG nova.network.neutron [req-1a317816-99f5-40ae-b198-6b8c03c4b429 req-4044a0f3-8ed2-491a-ae3d-441bda5b5a28 service nova] [instance: 422d06d6-6932-46c3-bb25-841e6f65c028] Refreshing network info cache for port 97caa1b8-427c-46b0-be7a-d59b932900c3 {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 840.167939] env[61923]: DEBUG oslo_concurrency.lockutils [None req-7d4ecc17-51d2-4753-8cb6-fd83e220b388 tempest-ServersNegativeTestMultiTenantJSON-1497120864 tempest-ServersNegativeTestMultiTenantJSON-1497120864-project-member] Acquiring lock "1d1b1dc4-c701-42b2-beba-344d622aef41" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 840.167939] env[61923]: DEBUG oslo_concurrency.lockutils [None req-7d4ecc17-51d2-4753-8cb6-fd83e220b388 tempest-ServersNegativeTestMultiTenantJSON-1497120864 tempest-ServersNegativeTestMultiTenantJSON-1497120864-project-member] Lock "1d1b1dc4-c701-42b2-beba-344d622aef41" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 840.168120] env[61923]: DEBUG oslo_concurrency.lockutils [None req-7d4ecc17-51d2-4753-8cb6-fd83e220b388 tempest-ServersNegativeTestMultiTenantJSON-1497120864 tempest-ServersNegativeTestMultiTenantJSON-1497120864-project-member] Acquiring lock "1d1b1dc4-c701-42b2-beba-344d622aef41-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 840.168311] env[61923]: DEBUG oslo_concurrency.lockutils [None req-7d4ecc17-51d2-4753-8cb6-fd83e220b388 tempest-ServersNegativeTestMultiTenantJSON-1497120864 tempest-ServersNegativeTestMultiTenantJSON-1497120864-project-member] Lock "1d1b1dc4-c701-42b2-beba-344d622aef41-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 840.168467] env[61923]: DEBUG oslo_concurrency.lockutils [None req-7d4ecc17-51d2-4753-8cb6-fd83e220b388 tempest-ServersNegativeTestMultiTenantJSON-1497120864 tempest-ServersNegativeTestMultiTenantJSON-1497120864-project-member] Lock "1d1b1dc4-c701-42b2-beba-344d622aef41-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 840.172271] env[61923]: INFO nova.compute.manager [None req-7d4ecc17-51d2-4753-8cb6-fd83e220b388 tempest-ServersNegativeTestMultiTenantJSON-1497120864 tempest-ServersNegativeTestMultiTenantJSON-1497120864-project-member] [instance: 1d1b1dc4-c701-42b2-beba-344d622aef41] Terminating instance [ 840.174652] env[61923]: DEBUG nova.compute.manager [None req-7d4ecc17-51d2-4753-8cb6-fd83e220b388 tempest-ServersNegativeTestMultiTenantJSON-1497120864 tempest-ServersNegativeTestMultiTenantJSON-1497120864-project-member] [instance: 1d1b1dc4-c701-42b2-beba-344d622aef41] Start destroying the instance on the hypervisor. {{(pid=61923) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 840.174879] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-7d4ecc17-51d2-4753-8cb6-fd83e220b388 tempest-ServersNegativeTestMultiTenantJSON-1497120864 tempest-ServersNegativeTestMultiTenantJSON-1497120864-project-member] [instance: 1d1b1dc4-c701-42b2-beba-344d622aef41] Destroying instance {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 840.175817] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-acaa66f1-1843-46c9-a169-8e46ea26661e {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.181547] env[61923]: DEBUG oslo_concurrency.lockutils [None req-513daa64-2b05-4c0d-94e9-55b95de63eec tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.320s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 840.182130] env[61923]: DEBUG nova.compute.manager [None req-513daa64-2b05-4c0d-94e9-55b95de63eec tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] [instance: 59198f4d-4dde-4eaf-9f6c-a962cbe53c6e] Start building networks asynchronously for instance. {{(pid=61923) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 840.187537] env[61923]: DEBUG oslo_concurrency.lockutils [None req-80740a2c-9779-4afc-be30-c4469cd3724d tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.272s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 840.190682] env[61923]: INFO nova.compute.claims [None req-80740a2c-9779-4afc-be30-c4469cd3724d tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] [instance: fc4fdf89-4615-4586-9dcd-ea90dc258361] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 840.192730] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-7d4ecc17-51d2-4753-8cb6-fd83e220b388 tempest-ServersNegativeTestMultiTenantJSON-1497120864 tempest-ServersNegativeTestMultiTenantJSON-1497120864-project-member] [instance: 1d1b1dc4-c701-42b2-beba-344d622aef41] Powering off the VM {{(pid=61923) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 840.193510] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d48e3a03-181a-42b3-8ab9-be105ac639d7 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.200705] env[61923]: DEBUG oslo_vmware.api [None req-7d4ecc17-51d2-4753-8cb6-fd83e220b388 tempest-ServersNegativeTestMultiTenantJSON-1497120864 tempest-ServersNegativeTestMultiTenantJSON-1497120864-project-member] Waiting for the task: (returnval){ [ 840.200705] env[61923]: value = "task-1377548" [ 840.200705] env[61923]: _type = "Task" [ 840.200705] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 840.213022] env[61923]: DEBUG oslo_vmware.api [None req-7d4ecc17-51d2-4753-8cb6-fd83e220b388 tempest-ServersNegativeTestMultiTenantJSON-1497120864 tempest-ServersNegativeTestMultiTenantJSON-1497120864-project-member] Task: {'id': task-1377548, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.222923] env[61923]: DEBUG oslo_vmware.api [None req-fc756c76-ebec-403f-b8ba-39ff93058631 tempest-ServerShowV257Test-986496257 tempest-ServerShowV257Test-986496257-project-member] Task: {'id': task-1377542, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.867852} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 840.223233] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-fc756c76-ebec-403f-b8ba-39ff93058631 tempest-ServerShowV257Test-986496257 tempest-ServerShowV257Test-986496257-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk to [datastore2] b8c60af7-3561-468c-a9a5-1c645080fc69/b8c60af7-3561-468c-a9a5-1c645080fc69.vmdk {{(pid=61923) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 840.224287] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-fc756c76-ebec-403f-b8ba-39ff93058631 tempest-ServerShowV257Test-986496257 tempest-ServerShowV257Test-986496257-project-member] [instance: b8c60af7-3561-468c-a9a5-1c645080fc69] Extending root virtual disk to 1048576 {{(pid=61923) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 840.224287] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-286d1f23-42a5-4aef-bc18-61db49a454f4 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.231422] env[61923]: DEBUG oslo_vmware.api [None req-fc756c76-ebec-403f-b8ba-39ff93058631 tempest-ServerShowV257Test-986496257 tempest-ServerShowV257Test-986496257-project-member] Waiting for the task: (returnval){ [ 840.231422] env[61923]: value = "task-1377549" [ 840.231422] env[61923]: _type = "Task" [ 840.231422] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 840.245025] env[61923]: DEBUG oslo_vmware.api [None req-fc756c76-ebec-403f-b8ba-39ff93058631 tempest-ServerShowV257Test-986496257 tempest-ServerShowV257Test-986496257-project-member] Task: {'id': task-1377549, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.260146] env[61923]: DEBUG oslo_vmware.api [None req-458eed44-c3ca-4049-bcef-fbcaf60c50ed tempest-ServersTestManualDisk-2012652654 tempest-ServersTestManualDisk-2012652654-project-member] Task: {'id': task-1377543, 'name': ReconfigVM_Task, 'duration_secs': 0.439453} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 840.260511] env[61923]: DEBUG nova.virt.vmwareapi.volumeops [None req-458eed44-c3ca-4049-bcef-fbcaf60c50ed tempest-ServersTestManualDisk-2012652654 tempest-ServersTestManualDisk-2012652654-project-member] [instance: f1839f61-4314-48fe-8ab6-14b5e49d569d] Reconfigured VM instance instance-00000041 to attach disk [datastore1] f1839f61-4314-48fe-8ab6-14b5e49d569d/f1839f61-4314-48fe-8ab6-14b5e49d569d.vmdk or device None with type sparse {{(pid=61923) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 840.261555] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c9560a8a-61a7-4891-b05f-e15594c6eb79 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.268893] env[61923]: DEBUG oslo_vmware.api [None req-458eed44-c3ca-4049-bcef-fbcaf60c50ed tempest-ServersTestManualDisk-2012652654 tempest-ServersTestManualDisk-2012652654-project-member] Waiting for the task: (returnval){ [ 840.268893] env[61923]: value = "task-1377550" [ 840.268893] env[61923]: _type = "Task" [ 840.268893] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 840.278811] env[61923]: DEBUG oslo_vmware.api [None req-458eed44-c3ca-4049-bcef-fbcaf60c50ed tempest-ServersTestManualDisk-2012652654 tempest-ServersTestManualDisk-2012652654-project-member] Task: {'id': task-1377550, 'name': Rename_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.302970] env[61923]: DEBUG nova.compute.manager [req-f34ff19f-19a0-445a-9f32-4078311b2820 req-8a4d07ed-8854-4733-9663-738e5af979f5 service nova] [instance: 73692517-1816-4e9b-ab2d-8265c683e83d] Received event network-vif-plugged-dbed34c7-9716-4c24-8f0e-2298de6cf4b9 {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 840.303269] env[61923]: DEBUG oslo_concurrency.lockutils [req-f34ff19f-19a0-445a-9f32-4078311b2820 req-8a4d07ed-8854-4733-9663-738e5af979f5 service nova] Acquiring lock "73692517-1816-4e9b-ab2d-8265c683e83d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 840.303553] env[61923]: DEBUG oslo_concurrency.lockutils [req-f34ff19f-19a0-445a-9f32-4078311b2820 req-8a4d07ed-8854-4733-9663-738e5af979f5 service nova] Lock "73692517-1816-4e9b-ab2d-8265c683e83d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 840.303787] env[61923]: DEBUG oslo_concurrency.lockutils [req-f34ff19f-19a0-445a-9f32-4078311b2820 req-8a4d07ed-8854-4733-9663-738e5af979f5 service nova] Lock "73692517-1816-4e9b-ab2d-8265c683e83d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 840.304026] env[61923]: DEBUG nova.compute.manager [req-f34ff19f-19a0-445a-9f32-4078311b2820 req-8a4d07ed-8854-4733-9663-738e5af979f5 service nova] [instance: 73692517-1816-4e9b-ab2d-8265c683e83d] No waiting events found dispatching network-vif-plugged-dbed34c7-9716-4c24-8f0e-2298de6cf4b9 {{(pid=61923) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 840.304277] env[61923]: WARNING nova.compute.manager [req-f34ff19f-19a0-445a-9f32-4078311b2820 req-8a4d07ed-8854-4733-9663-738e5af979f5 service nova] [instance: 73692517-1816-4e9b-ab2d-8265c683e83d] Received unexpected event network-vif-plugged-dbed34c7-9716-4c24-8f0e-2298de6cf4b9 for instance with vm_state building and task_state spawning. [ 840.510428] env[61923]: DEBUG oslo_vmware.api [-] Task: {'id': task-1377546, 'name': CreateVM_Task, 'duration_secs': 0.339317} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 840.510628] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 422d06d6-6932-46c3-bb25-841e6f65c028] Created VM on the ESX host {{(pid=61923) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 840.511475] env[61923]: DEBUG oslo_concurrency.lockutils [None req-950e07c2-502c-43fa-8598-da5f189b07e3 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 840.511475] env[61923]: DEBUG oslo_concurrency.lockutils [None req-950e07c2-502c-43fa-8598-da5f189b07e3 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 840.511856] env[61923]: DEBUG oslo_concurrency.lockutils [None req-950e07c2-502c-43fa-8598-da5f189b07e3 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 840.512124] env[61923]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-71797fa4-515d-492f-a6c4-38db6a1ab643 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.516948] env[61923]: DEBUG oslo_vmware.api [None req-950e07c2-502c-43fa-8598-da5f189b07e3 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Waiting for the task: (returnval){ [ 840.516948] env[61923]: value = "session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52da0868-dde5-5d8e-1eae-ab1f6e5d8597" [ 840.516948] env[61923]: _type = "Task" [ 840.516948] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 840.525300] env[61923]: DEBUG oslo_vmware.api [None req-950e07c2-502c-43fa-8598-da5f189b07e3 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52da0868-dde5-5d8e-1eae-ab1f6e5d8597, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.550915] env[61923]: DEBUG oslo_vmware.api [None req-9ff8b121-368d-4886-9882-962e7ef81bcf tempest-ServerTagsTestJSON-19215696 tempest-ServerTagsTestJSON-19215696-project-member] Task: {'id': task-1377547, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.695917] env[61923]: DEBUG nova.compute.utils [None req-513daa64-2b05-4c0d-94e9-55b95de63eec tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Using /dev/sd instead of None {{(pid=61923) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 840.697761] env[61923]: DEBUG nova.compute.manager [None req-513daa64-2b05-4c0d-94e9-55b95de63eec tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] [instance: 59198f4d-4dde-4eaf-9f6c-a962cbe53c6e] Allocating IP information in the background. {{(pid=61923) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 840.697761] env[61923]: DEBUG nova.network.neutron [None req-513daa64-2b05-4c0d-94e9-55b95de63eec tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] [instance: 59198f4d-4dde-4eaf-9f6c-a962cbe53c6e] allocate_for_instance() {{(pid=61923) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 840.712146] env[61923]: DEBUG oslo_vmware.api [None req-7d4ecc17-51d2-4753-8cb6-fd83e220b388 tempest-ServersNegativeTestMultiTenantJSON-1497120864 tempest-ServersNegativeTestMultiTenantJSON-1497120864-project-member] Task: {'id': task-1377548, 'name': PowerOffVM_Task} progress is 100%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.744910] env[61923]: DEBUG oslo_vmware.api [None req-fc756c76-ebec-403f-b8ba-39ff93058631 tempest-ServerShowV257Test-986496257 tempest-ServerShowV257Test-986496257-project-member] Task: {'id': task-1377549, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.074601} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 840.745513] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-fc756c76-ebec-403f-b8ba-39ff93058631 tempest-ServerShowV257Test-986496257 tempest-ServerShowV257Test-986496257-project-member] [instance: b8c60af7-3561-468c-a9a5-1c645080fc69] Extended root virtual disk {{(pid=61923) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 840.746883] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d748de7e-465b-4e5d-81a1-b9f489ff722d {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.782271] env[61923]: DEBUG nova.virt.vmwareapi.volumeops [None req-fc756c76-ebec-403f-b8ba-39ff93058631 tempest-ServerShowV257Test-986496257 tempest-ServerShowV257Test-986496257-project-member] [instance: b8c60af7-3561-468c-a9a5-1c645080fc69] Reconfiguring VM instance instance-0000003f to attach disk [datastore2] b8c60af7-3561-468c-a9a5-1c645080fc69/b8c60af7-3561-468c-a9a5-1c645080fc69.vmdk or device None with type sparse {{(pid=61923) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 840.782271] env[61923]: DEBUG nova.network.neutron [None req-e62df7da-42b5-4263-bcf8-001a2d06427a tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: 73692517-1816-4e9b-ab2d-8265c683e83d] Successfully updated port: dbed34c7-9716-4c24-8f0e-2298de6cf4b9 {{(pid=61923) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 840.783658] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3bf639ec-5bcd-4a01-840b-27466870b8af {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.810707] env[61923]: DEBUG oslo_concurrency.lockutils [None req-e62df7da-42b5-4263-bcf8-001a2d06427a tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Acquiring lock "refresh_cache-73692517-1816-4e9b-ab2d-8265c683e83d" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 840.810938] env[61923]: DEBUG oslo_concurrency.lockutils [None req-e62df7da-42b5-4263-bcf8-001a2d06427a tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Acquired lock "refresh_cache-73692517-1816-4e9b-ab2d-8265c683e83d" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 840.811193] env[61923]: DEBUG nova.network.neutron [None req-e62df7da-42b5-4263-bcf8-001a2d06427a tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: 73692517-1816-4e9b-ab2d-8265c683e83d] Building network info cache for instance {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 840.821056] env[61923]: DEBUG nova.policy [None req-513daa64-2b05-4c0d-94e9-55b95de63eec tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'eb00f8aff92b44e9a0e739366a1d2662', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7abfe95660904cd4b8824257795d8e8a', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61923) authorize /opt/stack/nova/nova/policy.py:201}} [ 840.831489] env[61923]: DEBUG oslo_vmware.api [None req-fc756c76-ebec-403f-b8ba-39ff93058631 tempest-ServerShowV257Test-986496257 tempest-ServerShowV257Test-986496257-project-member] Waiting for the task: (returnval){ [ 840.831489] env[61923]: value = "task-1377551" [ 840.831489] env[61923]: _type = "Task" [ 840.831489] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 840.840857] env[61923]: DEBUG oslo_vmware.api [None req-458eed44-c3ca-4049-bcef-fbcaf60c50ed tempest-ServersTestManualDisk-2012652654 tempest-ServersTestManualDisk-2012652654-project-member] Task: {'id': task-1377550, 'name': Rename_Task, 'duration_secs': 0.431154} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 840.845904] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-458eed44-c3ca-4049-bcef-fbcaf60c50ed tempest-ServersTestManualDisk-2012652654 tempest-ServersTestManualDisk-2012652654-project-member] [instance: f1839f61-4314-48fe-8ab6-14b5e49d569d] Powering on the VM {{(pid=61923) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 840.846791] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1a22a4ce-5534-47aa-a32a-e60f6cc3c2e2 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 840.856273] env[61923]: DEBUG oslo_vmware.api [None req-fc756c76-ebec-403f-b8ba-39ff93058631 tempest-ServerShowV257Test-986496257 tempest-ServerShowV257Test-986496257-project-member] Task: {'id': task-1377551, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 840.858247] env[61923]: DEBUG oslo_vmware.api [None req-458eed44-c3ca-4049-bcef-fbcaf60c50ed tempest-ServersTestManualDisk-2012652654 tempest-ServersTestManualDisk-2012652654-project-member] Waiting for the task: (returnval){ [ 840.858247] env[61923]: value = "task-1377552" [ 840.858247] env[61923]: _type = "Task" [ 840.858247] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 840.863869] env[61923]: DEBUG nova.network.neutron [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] [instance: b8c60af7-3561-468c-a9a5-1c645080fc69] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 840.872976] env[61923]: DEBUG oslo_vmware.api [None req-458eed44-c3ca-4049-bcef-fbcaf60c50ed tempest-ServersTestManualDisk-2012652654 tempest-ServersTestManualDisk-2012652654-project-member] Task: {'id': task-1377552, 'name': PowerOnVM_Task} progress is 33%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.027023] env[61923]: DEBUG oslo_vmware.api [None req-950e07c2-502c-43fa-8598-da5f189b07e3 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52da0868-dde5-5d8e-1eae-ab1f6e5d8597, 'name': SearchDatastore_Task, 'duration_secs': 0.011625} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 841.027590] env[61923]: DEBUG oslo_concurrency.lockutils [None req-950e07c2-502c-43fa-8598-da5f189b07e3 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 841.027590] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-950e07c2-502c-43fa-8598-da5f189b07e3 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] [instance: 422d06d6-6932-46c3-bb25-841e6f65c028] Processing image 0f153f63-ae0a-45b1-b7f5-7f9b673c947f {{(pid=61923) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 841.027809] env[61923]: DEBUG oslo_concurrency.lockutils [None req-950e07c2-502c-43fa-8598-da5f189b07e3 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 841.027920] env[61923]: DEBUG oslo_concurrency.lockutils [None req-950e07c2-502c-43fa-8598-da5f189b07e3 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 841.028062] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-950e07c2-502c-43fa-8598-da5f189b07e3 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61923) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 841.028323] env[61923]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-019d1118-6d82-46f3-8fb4-dd550b9b0f68 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.052708] env[61923]: DEBUG oslo_vmware.api [None req-9ff8b121-368d-4886-9882-962e7ef81bcf tempest-ServerTagsTestJSON-19215696 tempest-ServerTagsTestJSON-19215696-project-member] Task: {'id': task-1377547, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.581564} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 841.053914] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-9ff8b121-368d-4886-9882-962e7ef81bcf tempest-ServerTagsTestJSON-19215696 tempest-ServerTagsTestJSON-19215696-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk to [datastore1] 48825592-dcdc-41d5-9fbf-500d1f31017b/48825592-dcdc-41d5-9fbf-500d1f31017b.vmdk {{(pid=61923) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 841.054169] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-9ff8b121-368d-4886-9882-962e7ef81bcf tempest-ServerTagsTestJSON-19215696 tempest-ServerTagsTestJSON-19215696-project-member] [instance: 48825592-dcdc-41d5-9fbf-500d1f31017b] Extending root virtual disk to 1048576 {{(pid=61923) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 841.054406] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-950e07c2-502c-43fa-8598-da5f189b07e3 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61923) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 841.054554] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-950e07c2-502c-43fa-8598-da5f189b07e3 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61923) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 841.055248] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-b1495fda-9c29-4776-bb9c-128c81727372 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.057089] env[61923]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e3d9d4e4-99db-4b82-9d25-0777e0a974b8 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.062680] env[61923]: DEBUG oslo_vmware.api [None req-950e07c2-502c-43fa-8598-da5f189b07e3 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Waiting for the task: (returnval){ [ 841.062680] env[61923]: value = "session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52e850c2-25af-fc17-3726-17362c0bba5e" [ 841.062680] env[61923]: _type = "Task" [ 841.062680] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 841.068741] env[61923]: DEBUG oslo_vmware.api [None req-9ff8b121-368d-4886-9882-962e7ef81bcf tempest-ServerTagsTestJSON-19215696 tempest-ServerTagsTestJSON-19215696-project-member] Waiting for the task: (returnval){ [ 841.068741] env[61923]: value = "task-1377553" [ 841.068741] env[61923]: _type = "Task" [ 841.068741] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 841.075589] env[61923]: DEBUG oslo_vmware.api [None req-950e07c2-502c-43fa-8598-da5f189b07e3 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52e850c2-25af-fc17-3726-17362c0bba5e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.080527] env[61923]: DEBUG oslo_vmware.api [None req-9ff8b121-368d-4886-9882-962e7ef81bcf tempest-ServerTagsTestJSON-19215696 tempest-ServerTagsTestJSON-19215696-project-member] Task: {'id': task-1377553, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.088404] env[61923]: DEBUG nova.network.neutron [req-1a317816-99f5-40ae-b198-6b8c03c4b429 req-4044a0f3-8ed2-491a-ae3d-441bda5b5a28 service nova] [instance: 422d06d6-6932-46c3-bb25-841e6f65c028] Updated VIF entry in instance network info cache for port 97caa1b8-427c-46b0-be7a-d59b932900c3. {{(pid=61923) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 841.088770] env[61923]: DEBUG nova.network.neutron [req-1a317816-99f5-40ae-b198-6b8c03c4b429 req-4044a0f3-8ed2-491a-ae3d-441bda5b5a28 service nova] [instance: 422d06d6-6932-46c3-bb25-841e6f65c028] Updating instance_info_cache with network_info: [{"id": "97caa1b8-427c-46b0-be7a-d59b932900c3", "address": "fa:16:3e:fa:f8:d1", "network": {"id": "a8b7e1b4-1ec6-4226-bdd6-535dbce450d2", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1928758125-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "912300a09b80452e85e3cd13c4a644cc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7806fe18-2b89-4386-87b1-f22876f82af2", "external-id": "nsx-vlan-transportzone-727", "segmentation_id": 727, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap97caa1b8-42", "ovs_interfaceid": "97caa1b8-427c-46b0-be7a-d59b932900c3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 841.203129] env[61923]: DEBUG nova.compute.manager [None req-513daa64-2b05-4c0d-94e9-55b95de63eec tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] [instance: 59198f4d-4dde-4eaf-9f6c-a962cbe53c6e] Start building block device mappings for instance. {{(pid=61923) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 841.217894] env[61923]: DEBUG oslo_vmware.api [None req-7d4ecc17-51d2-4753-8cb6-fd83e220b388 tempest-ServersNegativeTestMultiTenantJSON-1497120864 tempest-ServersNegativeTestMultiTenantJSON-1497120864-project-member] Task: {'id': task-1377548, 'name': PowerOffVM_Task, 'duration_secs': 0.518459} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 841.218165] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-7d4ecc17-51d2-4753-8cb6-fd83e220b388 tempest-ServersNegativeTestMultiTenantJSON-1497120864 tempest-ServersNegativeTestMultiTenantJSON-1497120864-project-member] [instance: 1d1b1dc4-c701-42b2-beba-344d622aef41] Powered off the VM {{(pid=61923) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 841.218338] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-7d4ecc17-51d2-4753-8cb6-fd83e220b388 tempest-ServersNegativeTestMultiTenantJSON-1497120864 tempest-ServersNegativeTestMultiTenantJSON-1497120864-project-member] [instance: 1d1b1dc4-c701-42b2-beba-344d622aef41] Unregistering the VM {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 841.218589] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d0667aad-dc58-43f7-bb25-226b7384cc83 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.304599] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-7d4ecc17-51d2-4753-8cb6-fd83e220b388 tempest-ServersNegativeTestMultiTenantJSON-1497120864 tempest-ServersNegativeTestMultiTenantJSON-1497120864-project-member] [instance: 1d1b1dc4-c701-42b2-beba-344d622aef41] Unregistered the VM {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 841.304599] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-7d4ecc17-51d2-4753-8cb6-fd83e220b388 tempest-ServersNegativeTestMultiTenantJSON-1497120864 tempest-ServersNegativeTestMultiTenantJSON-1497120864-project-member] [instance: 1d1b1dc4-c701-42b2-beba-344d622aef41] Deleting contents of the VM from datastore datastore1 {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 841.304599] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-7d4ecc17-51d2-4753-8cb6-fd83e220b388 tempest-ServersNegativeTestMultiTenantJSON-1497120864 tempest-ServersNegativeTestMultiTenantJSON-1497120864-project-member] Deleting the datastore file [datastore1] 1d1b1dc4-c701-42b2-beba-344d622aef41 {{(pid=61923) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 841.304599] env[61923]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-169c7c75-fa7e-4f98-ab05-e06876e4be6d {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.312095] env[61923]: DEBUG oslo_vmware.api [None req-7d4ecc17-51d2-4753-8cb6-fd83e220b388 tempest-ServersNegativeTestMultiTenantJSON-1497120864 tempest-ServersNegativeTestMultiTenantJSON-1497120864-project-member] Waiting for the task: (returnval){ [ 841.312095] env[61923]: value = "task-1377555" [ 841.312095] env[61923]: _type = "Task" [ 841.312095] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 841.315748] env[61923]: DEBUG nova.network.neutron [None req-513daa64-2b05-4c0d-94e9-55b95de63eec tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] [instance: 59198f4d-4dde-4eaf-9f6c-a962cbe53c6e] Successfully created port: cf99020c-d7cc-4912-a021-8861628b7540 {{(pid=61923) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 841.329529] env[61923]: DEBUG oslo_vmware.api [None req-7d4ecc17-51d2-4753-8cb6-fd83e220b388 tempest-ServersNegativeTestMultiTenantJSON-1497120864 tempest-ServersNegativeTestMultiTenantJSON-1497120864-project-member] Task: {'id': task-1377555, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.352113] env[61923]: DEBUG oslo_vmware.api [None req-fc756c76-ebec-403f-b8ba-39ff93058631 tempest-ServerShowV257Test-986496257 tempest-ServerShowV257Test-986496257-project-member] Task: {'id': task-1377551, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.370427] env[61923]: DEBUG oslo_vmware.api [None req-458eed44-c3ca-4049-bcef-fbcaf60c50ed tempest-ServersTestManualDisk-2012652654 tempest-ServersTestManualDisk-2012652654-project-member] Task: {'id': task-1377552, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.415918] env[61923]: DEBUG nova.network.neutron [None req-e62df7da-42b5-4263-bcf8-001a2d06427a tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: 73692517-1816-4e9b-ab2d-8265c683e83d] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 841.439973] env[61923]: DEBUG nova.network.neutron [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] [instance: b8c60af7-3561-468c-a9a5-1c645080fc69] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 841.468852] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f25b01e-7872-4b75-ad97-cb32d9b883ee {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.487255] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d590d3e1-5f67-4540-9707-d56b48a20023 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.540302] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3500bf7f-668b-41e7-85cd-c13a169742b0 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.551114] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0f230bf-366b-4e82-9513-8e82ca6f9408 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.567884] env[61923]: DEBUG nova.compute.provider_tree [None req-80740a2c-9779-4afc-be30-c4469cd3724d tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 841.581316] env[61923]: DEBUG oslo_vmware.api [None req-950e07c2-502c-43fa-8598-da5f189b07e3 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52e850c2-25af-fc17-3726-17362c0bba5e, 'name': SearchDatastore_Task, 'duration_secs': 0.023018} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 841.585409] env[61923]: DEBUG oslo_vmware.api [None req-9ff8b121-368d-4886-9882-962e7ef81bcf tempest-ServerTagsTestJSON-19215696 tempest-ServerTagsTestJSON-19215696-project-member] Task: {'id': task-1377553, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.260406} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 841.585624] env[61923]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8e8f0e2b-6a8a-40f8-8e51-d20f9497869a {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.587857] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-9ff8b121-368d-4886-9882-962e7ef81bcf tempest-ServerTagsTestJSON-19215696 tempest-ServerTagsTestJSON-19215696-project-member] [instance: 48825592-dcdc-41d5-9fbf-500d1f31017b] Extended root virtual disk {{(pid=61923) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 841.588607] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a922a28d-6dcb-4715-9933-3bbc45f4bec3 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.591954] env[61923]: DEBUG oslo_concurrency.lockutils [req-1a317816-99f5-40ae-b198-6b8c03c4b429 req-4044a0f3-8ed2-491a-ae3d-441bda5b5a28 service nova] Releasing lock "refresh_cache-422d06d6-6932-46c3-bb25-841e6f65c028" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 841.593890] env[61923]: DEBUG oslo_vmware.api [None req-950e07c2-502c-43fa-8598-da5f189b07e3 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Waiting for the task: (returnval){ [ 841.593890] env[61923]: value = "session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52969adc-5087-ca21-2200-9be5c2c45e59" [ 841.593890] env[61923]: _type = "Task" [ 841.593890] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 841.613805] env[61923]: DEBUG nova.virt.vmwareapi.volumeops [None req-9ff8b121-368d-4886-9882-962e7ef81bcf tempest-ServerTagsTestJSON-19215696 tempest-ServerTagsTestJSON-19215696-project-member] [instance: 48825592-dcdc-41d5-9fbf-500d1f31017b] Reconfiguring VM instance instance-00000042 to attach disk [datastore1] 48825592-dcdc-41d5-9fbf-500d1f31017b/48825592-dcdc-41d5-9fbf-500d1f31017b.vmdk or device None with type sparse {{(pid=61923) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 841.616878] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c8d05962-6b5e-4c62-89bb-e6b7e058317f {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.634404] env[61923]: DEBUG oslo_vmware.api [None req-950e07c2-502c-43fa-8598-da5f189b07e3 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52969adc-5087-ca21-2200-9be5c2c45e59, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.640197] env[61923]: DEBUG oslo_vmware.api [None req-9ff8b121-368d-4886-9882-962e7ef81bcf tempest-ServerTagsTestJSON-19215696 tempest-ServerTagsTestJSON-19215696-project-member] Waiting for the task: (returnval){ [ 841.640197] env[61923]: value = "task-1377556" [ 841.640197] env[61923]: _type = "Task" [ 841.640197] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 841.648311] env[61923]: DEBUG oslo_vmware.api [None req-9ff8b121-368d-4886-9882-962e7ef81bcf tempest-ServerTagsTestJSON-19215696 tempest-ServerTagsTestJSON-19215696-project-member] Task: {'id': task-1377556, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.821696] env[61923]: DEBUG nova.network.neutron [None req-e62df7da-42b5-4263-bcf8-001a2d06427a tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: 73692517-1816-4e9b-ab2d-8265c683e83d] Updating instance_info_cache with network_info: [{"id": "dbed34c7-9716-4c24-8f0e-2298de6cf4b9", "address": "fa:16:3e:a0:e2:bc", "network": {"id": "d1c38272-0461-43e5-93ae-3964346f77d0", "bridge": "br-int", "label": "tempest-ServersTestJSON-151839165-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "69cc941eb6dd4780ac12aa29656c37f7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8ba3bd22-c936-470e-89bd-b3a5587e87a0", "external-id": "nsx-vlan-transportzone-605", "segmentation_id": 605, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdbed34c7-97", "ovs_interfaceid": "dbed34c7-9716-4c24-8f0e-2298de6cf4b9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 841.826380] env[61923]: DEBUG oslo_vmware.api [None req-7d4ecc17-51d2-4753-8cb6-fd83e220b388 tempest-ServersNegativeTestMultiTenantJSON-1497120864 tempest-ServersNegativeTestMultiTenantJSON-1497120864-project-member] Task: {'id': task-1377555, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.463386} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 841.826733] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-7d4ecc17-51d2-4753-8cb6-fd83e220b388 tempest-ServersNegativeTestMultiTenantJSON-1497120864 tempest-ServersNegativeTestMultiTenantJSON-1497120864-project-member] Deleted the datastore file {{(pid=61923) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 841.827078] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-7d4ecc17-51d2-4753-8cb6-fd83e220b388 tempest-ServersNegativeTestMultiTenantJSON-1497120864 tempest-ServersNegativeTestMultiTenantJSON-1497120864-project-member] [instance: 1d1b1dc4-c701-42b2-beba-344d622aef41] Deleted contents of the VM from datastore datastore1 {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 841.827373] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-7d4ecc17-51d2-4753-8cb6-fd83e220b388 tempest-ServersNegativeTestMultiTenantJSON-1497120864 tempest-ServersNegativeTestMultiTenantJSON-1497120864-project-member] [instance: 1d1b1dc4-c701-42b2-beba-344d622aef41] Instance destroyed {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 841.827577] env[61923]: INFO nova.compute.manager [None req-7d4ecc17-51d2-4753-8cb6-fd83e220b388 tempest-ServersNegativeTestMultiTenantJSON-1497120864 tempest-ServersNegativeTestMultiTenantJSON-1497120864-project-member] [instance: 1d1b1dc4-c701-42b2-beba-344d622aef41] Took 1.65 seconds to destroy the instance on the hypervisor. [ 841.827830] env[61923]: DEBUG oslo.service.loopingcall [None req-7d4ecc17-51d2-4753-8cb6-fd83e220b388 tempest-ServersNegativeTestMultiTenantJSON-1497120864 tempest-ServersNegativeTestMultiTenantJSON-1497120864-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61923) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 841.828359] env[61923]: DEBUG nova.compute.manager [-] [instance: 1d1b1dc4-c701-42b2-beba-344d622aef41] Deallocating network for instance {{(pid=61923) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 841.828359] env[61923]: DEBUG nova.network.neutron [-] [instance: 1d1b1dc4-c701-42b2-beba-344d622aef41] deallocate_for_instance() {{(pid=61923) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 841.852793] env[61923]: DEBUG oslo_vmware.api [None req-fc756c76-ebec-403f-b8ba-39ff93058631 tempest-ServerShowV257Test-986496257 tempest-ServerShowV257Test-986496257-project-member] Task: {'id': task-1377551, 'name': ReconfigVM_Task, 'duration_secs': 0.981692} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 841.853097] env[61923]: DEBUG nova.virt.vmwareapi.volumeops [None req-fc756c76-ebec-403f-b8ba-39ff93058631 tempest-ServerShowV257Test-986496257 tempest-ServerShowV257Test-986496257-project-member] [instance: b8c60af7-3561-468c-a9a5-1c645080fc69] Reconfigured VM instance instance-0000003f to attach disk [datastore2] b8c60af7-3561-468c-a9a5-1c645080fc69/b8c60af7-3561-468c-a9a5-1c645080fc69.vmdk or device None with type sparse {{(pid=61923) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 841.853737] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e0ad27a1-a6f6-4ed7-a977-cb6aaab21e07 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.863613] env[61923]: DEBUG oslo_vmware.api [None req-fc756c76-ebec-403f-b8ba-39ff93058631 tempest-ServerShowV257Test-986496257 tempest-ServerShowV257Test-986496257-project-member] Waiting for the task: (returnval){ [ 841.863613] env[61923]: value = "task-1377557" [ 841.863613] env[61923]: _type = "Task" [ 841.863613] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 841.870786] env[61923]: DEBUG oslo_vmware.api [None req-458eed44-c3ca-4049-bcef-fbcaf60c50ed tempest-ServersTestManualDisk-2012652654 tempest-ServersTestManualDisk-2012652654-project-member] Task: {'id': task-1377552, 'name': PowerOnVM_Task} progress is 94%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.875873] env[61923]: DEBUG oslo_vmware.api [None req-fc756c76-ebec-403f-b8ba-39ff93058631 tempest-ServerShowV257Test-986496257 tempest-ServerShowV257Test-986496257-project-member] Task: {'id': task-1377557, 'name': Rename_Task} progress is 10%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 841.943565] env[61923]: DEBUG oslo_concurrency.lockutils [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Releasing lock "refresh_cache-b8c60af7-3561-468c-a9a5-1c645080fc69" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 841.943753] env[61923]: DEBUG nova.compute.manager [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] [instance: b8c60af7-3561-468c-a9a5-1c645080fc69] Updated the network info_cache for instance {{(pid=61923) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9999}} [ 841.944015] env[61923]: DEBUG oslo_service.periodic_task [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61923) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 841.944223] env[61923]: DEBUG oslo_service.periodic_task [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61923) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 841.944397] env[61923]: DEBUG oslo_service.periodic_task [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61923) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 841.944513] env[61923]: DEBUG oslo_service.periodic_task [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61923) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 841.944653] env[61923]: DEBUG oslo_service.periodic_task [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61923) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 841.944863] env[61923]: DEBUG oslo_service.periodic_task [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61923) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 841.945457] env[61923]: DEBUG nova.compute.manager [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61923) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 841.945457] env[61923]: DEBUG oslo_service.periodic_task [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Running periodic task ComputeManager.update_available_resource {{(pid=61923) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 842.074201] env[61923]: DEBUG nova.scheduler.client.report [None req-80740a2c-9779-4afc-be30-c4469cd3724d tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 842.106394] env[61923]: DEBUG oslo_vmware.api [None req-950e07c2-502c-43fa-8598-da5f189b07e3 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52969adc-5087-ca21-2200-9be5c2c45e59, 'name': SearchDatastore_Task, 'duration_secs': 0.058838} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 842.106672] env[61923]: DEBUG oslo_concurrency.lockutils [None req-950e07c2-502c-43fa-8598-da5f189b07e3 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 842.106864] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-950e07c2-502c-43fa-8598-da5f189b07e3 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk to [datastore2] 422d06d6-6932-46c3-bb25-841e6f65c028/422d06d6-6932-46c3-bb25-841e6f65c028.vmdk {{(pid=61923) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 842.107131] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-7e887e12-c4f9-4c51-a9f0-b6fc853690cf {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.113735] env[61923]: DEBUG oslo_vmware.api [None req-950e07c2-502c-43fa-8598-da5f189b07e3 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Waiting for the task: (returnval){ [ 842.113735] env[61923]: value = "task-1377558" [ 842.113735] env[61923]: _type = "Task" [ 842.113735] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 842.121849] env[61923]: DEBUG oslo_vmware.api [None req-950e07c2-502c-43fa-8598-da5f189b07e3 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Task: {'id': task-1377558, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.149518] env[61923]: DEBUG oslo_vmware.api [None req-9ff8b121-368d-4886-9882-962e7ef81bcf tempest-ServerTagsTestJSON-19215696 tempest-ServerTagsTestJSON-19215696-project-member] Task: {'id': task-1377556, 'name': ReconfigVM_Task, 'duration_secs': 0.306634} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 842.149843] env[61923]: DEBUG nova.virt.vmwareapi.volumeops [None req-9ff8b121-368d-4886-9882-962e7ef81bcf tempest-ServerTagsTestJSON-19215696 tempest-ServerTagsTestJSON-19215696-project-member] [instance: 48825592-dcdc-41d5-9fbf-500d1f31017b] Reconfigured VM instance instance-00000042 to attach disk [datastore1] 48825592-dcdc-41d5-9fbf-500d1f31017b/48825592-dcdc-41d5-9fbf-500d1f31017b.vmdk or device None with type sparse {{(pid=61923) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 842.150487] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-1a50e7f6-843d-448f-a28c-bcb8b87bb065 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.156323] env[61923]: DEBUG oslo_vmware.api [None req-9ff8b121-368d-4886-9882-962e7ef81bcf tempest-ServerTagsTestJSON-19215696 tempest-ServerTagsTestJSON-19215696-project-member] Waiting for the task: (returnval){ [ 842.156323] env[61923]: value = "task-1377559" [ 842.156323] env[61923]: _type = "Task" [ 842.156323] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 842.164512] env[61923]: DEBUG oslo_vmware.api [None req-9ff8b121-368d-4886-9882-962e7ef81bcf tempest-ServerTagsTestJSON-19215696 tempest-ServerTagsTestJSON-19215696-project-member] Task: {'id': task-1377559, 'name': Rename_Task} progress is 5%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.218724] env[61923]: DEBUG nova.compute.manager [None req-513daa64-2b05-4c0d-94e9-55b95de63eec tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] [instance: 59198f4d-4dde-4eaf-9f6c-a962cbe53c6e] Start spawning the instance on the hypervisor. {{(pid=61923) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 842.248217] env[61923]: DEBUG nova.virt.hardware [None req-513daa64-2b05-4c0d-94e9-55b95de63eec tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-29T20:07:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-29T20:07:35Z,direct_url=,disk_format='vmdk',id=0f153f63-ae0a-45b1-b7f5-7f9b673c947f,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='4e7b5e3b3c3443c8bd5c70f8a15739f5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-29T20:07:36Z,virtual_size=,visibility=), allow threads: False {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 842.248476] env[61923]: DEBUG nova.virt.hardware [None req-513daa64-2b05-4c0d-94e9-55b95de63eec tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Flavor limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 842.248666] env[61923]: DEBUG nova.virt.hardware [None req-513daa64-2b05-4c0d-94e9-55b95de63eec tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Image limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 842.248966] env[61923]: DEBUG nova.virt.hardware [None req-513daa64-2b05-4c0d-94e9-55b95de63eec tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Flavor pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 842.249187] env[61923]: DEBUG nova.virt.hardware [None req-513daa64-2b05-4c0d-94e9-55b95de63eec tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Image pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 842.249425] env[61923]: DEBUG nova.virt.hardware [None req-513daa64-2b05-4c0d-94e9-55b95de63eec tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 842.249600] env[61923]: DEBUG nova.virt.hardware [None req-513daa64-2b05-4c0d-94e9-55b95de63eec tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 842.249798] env[61923]: DEBUG nova.virt.hardware [None req-513daa64-2b05-4c0d-94e9-55b95de63eec tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 842.249995] env[61923]: DEBUG nova.virt.hardware [None req-513daa64-2b05-4c0d-94e9-55b95de63eec tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Got 1 possible topologies {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 842.250210] env[61923]: DEBUG nova.virt.hardware [None req-513daa64-2b05-4c0d-94e9-55b95de63eec tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 842.250400] env[61923]: DEBUG nova.virt.hardware [None req-513daa64-2b05-4c0d-94e9-55b95de63eec tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 842.251293] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9d2e443-dbc7-4143-9a8b-7db5bbf1eda4 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.255983] env[61923]: DEBUG nova.compute.manager [req-95afea4b-0803-4361-b7e1-8266a93cb7f6 req-db5b8fab-4416-4a51-89b5-ed2c16d5ad2c service nova] [instance: 73692517-1816-4e9b-ab2d-8265c683e83d] Received event network-changed-dbed34c7-9716-4c24-8f0e-2298de6cf4b9 {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 842.256650] env[61923]: DEBUG nova.compute.manager [req-95afea4b-0803-4361-b7e1-8266a93cb7f6 req-db5b8fab-4416-4a51-89b5-ed2c16d5ad2c service nova] [instance: 73692517-1816-4e9b-ab2d-8265c683e83d] Refreshing instance network info cache due to event network-changed-dbed34c7-9716-4c24-8f0e-2298de6cf4b9. {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 842.256846] env[61923]: DEBUG oslo_concurrency.lockutils [req-95afea4b-0803-4361-b7e1-8266a93cb7f6 req-db5b8fab-4416-4a51-89b5-ed2c16d5ad2c service nova] Acquiring lock "refresh_cache-73692517-1816-4e9b-ab2d-8265c683e83d" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 842.263178] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1025c6cc-1475-415c-8a6b-0468d8197271 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.326017] env[61923]: DEBUG oslo_concurrency.lockutils [None req-e62df7da-42b5-4263-bcf8-001a2d06427a tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Releasing lock "refresh_cache-73692517-1816-4e9b-ab2d-8265c683e83d" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 842.326384] env[61923]: DEBUG nova.compute.manager [None req-e62df7da-42b5-4263-bcf8-001a2d06427a tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: 73692517-1816-4e9b-ab2d-8265c683e83d] Instance network_info: |[{"id": "dbed34c7-9716-4c24-8f0e-2298de6cf4b9", "address": "fa:16:3e:a0:e2:bc", "network": {"id": "d1c38272-0461-43e5-93ae-3964346f77d0", "bridge": "br-int", "label": "tempest-ServersTestJSON-151839165-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "69cc941eb6dd4780ac12aa29656c37f7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8ba3bd22-c936-470e-89bd-b3a5587e87a0", "external-id": "nsx-vlan-transportzone-605", "segmentation_id": 605, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdbed34c7-97", "ovs_interfaceid": "dbed34c7-9716-4c24-8f0e-2298de6cf4b9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61923) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 842.331065] env[61923]: DEBUG oslo_concurrency.lockutils [req-95afea4b-0803-4361-b7e1-8266a93cb7f6 req-db5b8fab-4416-4a51-89b5-ed2c16d5ad2c service nova] Acquired lock "refresh_cache-73692517-1816-4e9b-ab2d-8265c683e83d" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 842.331065] env[61923]: DEBUG nova.network.neutron [req-95afea4b-0803-4361-b7e1-8266a93cb7f6 req-db5b8fab-4416-4a51-89b5-ed2c16d5ad2c service nova] [instance: 73692517-1816-4e9b-ab2d-8265c683e83d] Refreshing network info cache for port dbed34c7-9716-4c24-8f0e-2298de6cf4b9 {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 842.331065] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-e62df7da-42b5-4263-bcf8-001a2d06427a tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: 73692517-1816-4e9b-ab2d-8265c683e83d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:a0:e2:bc', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '8ba3bd22-c936-470e-89bd-b3a5587e87a0', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'dbed34c7-9716-4c24-8f0e-2298de6cf4b9', 'vif_model': 'vmxnet3'}] {{(pid=61923) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 842.339061] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-e62df7da-42b5-4263-bcf8-001a2d06427a tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Creating folder: Project (69cc941eb6dd4780ac12aa29656c37f7). Parent ref: group-v292629. {{(pid=61923) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 842.342318] env[61923]: DEBUG nova.network.neutron [-] [instance: 1d1b1dc4-c701-42b2-beba-344d622aef41] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 842.343409] env[61923]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-c6567675-9b73-4bed-9622-dee014f68164 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.353291] env[61923]: DEBUG nova.compute.manager [req-368ff499-4edf-42ea-8dbc-ce6f7688a61f req-95eb5e71-f848-4490-bb3e-554df6660253 service nova] [instance: 1d1b1dc4-c701-42b2-beba-344d622aef41] Received event network-vif-deleted-8118febd-f949-4bed-9959-775f44cfbd07 {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 842.353509] env[61923]: INFO nova.compute.manager [req-368ff499-4edf-42ea-8dbc-ce6f7688a61f req-95eb5e71-f848-4490-bb3e-554df6660253 service nova] [instance: 1d1b1dc4-c701-42b2-beba-344d622aef41] Neutron deleted interface 8118febd-f949-4bed-9959-775f44cfbd07; detaching it from the instance and deleting it from the info cache [ 842.353676] env[61923]: DEBUG nova.network.neutron [req-368ff499-4edf-42ea-8dbc-ce6f7688a61f req-95eb5e71-f848-4490-bb3e-554df6660253 service nova] [instance: 1d1b1dc4-c701-42b2-beba-344d622aef41] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 842.360433] env[61923]: INFO nova.virt.vmwareapi.vm_util [None req-e62df7da-42b5-4263-bcf8-001a2d06427a tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Created folder: Project (69cc941eb6dd4780ac12aa29656c37f7) in parent group-v292629. [ 842.360693] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-e62df7da-42b5-4263-bcf8-001a2d06427a tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Creating folder: Instances. Parent ref: group-v292666. {{(pid=61923) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 842.363066] env[61923]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-cc0eebf5-7037-4927-9c21-06d7d9058b5d {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.378700] env[61923]: DEBUG oslo_vmware.api [None req-fc756c76-ebec-403f-b8ba-39ff93058631 tempest-ServerShowV257Test-986496257 tempest-ServerShowV257Test-986496257-project-member] Task: {'id': task-1377557, 'name': Rename_Task, 'duration_secs': 0.148717} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 842.384971] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-fc756c76-ebec-403f-b8ba-39ff93058631 tempest-ServerShowV257Test-986496257 tempest-ServerShowV257Test-986496257-project-member] [instance: b8c60af7-3561-468c-a9a5-1c645080fc69] Powering on the VM {{(pid=61923) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 842.385264] env[61923]: INFO nova.virt.vmwareapi.vm_util [None req-e62df7da-42b5-4263-bcf8-001a2d06427a tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Created folder: Instances in parent group-v292666. [ 842.385515] env[61923]: DEBUG oslo.service.loopingcall [None req-e62df7da-42b5-4263-bcf8-001a2d06427a tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61923) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 842.386051] env[61923]: DEBUG oslo_vmware.api [None req-458eed44-c3ca-4049-bcef-fbcaf60c50ed tempest-ServersTestManualDisk-2012652654 tempest-ServersTestManualDisk-2012652654-project-member] Task: {'id': task-1377552, 'name': PowerOnVM_Task, 'duration_secs': 1.051056} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 842.386285] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-68261361-66e2-43a8-94e5-0a282b6f5ea9 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.388140] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 73692517-1816-4e9b-ab2d-8265c683e83d] Creating VM on the ESX host {{(pid=61923) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 842.388441] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-458eed44-c3ca-4049-bcef-fbcaf60c50ed tempest-ServersTestManualDisk-2012652654 tempest-ServersTestManualDisk-2012652654-project-member] [instance: f1839f61-4314-48fe-8ab6-14b5e49d569d] Powered on the VM {{(pid=61923) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 842.388657] env[61923]: INFO nova.compute.manager [None req-458eed44-c3ca-4049-bcef-fbcaf60c50ed tempest-ServersTestManualDisk-2012652654 tempest-ServersTestManualDisk-2012652654-project-member] [instance: f1839f61-4314-48fe-8ab6-14b5e49d569d] Took 9.54 seconds to spawn the instance on the hypervisor. [ 842.388861] env[61923]: DEBUG nova.compute.manager [None req-458eed44-c3ca-4049-bcef-fbcaf60c50ed tempest-ServersTestManualDisk-2012652654 tempest-ServersTestManualDisk-2012652654-project-member] [instance: f1839f61-4314-48fe-8ab6-14b5e49d569d] Checking state {{(pid=61923) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 842.389709] env[61923]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-5772e91a-92b0-46ce-af67-f1296cb4050d {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.405911] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04cd0320-7ba0-486c-bd84-d386203a7d48 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.419110] env[61923]: DEBUG oslo_vmware.api [None req-fc756c76-ebec-403f-b8ba-39ff93058631 tempest-ServerShowV257Test-986496257 tempest-ServerShowV257Test-986496257-project-member] Waiting for the task: (returnval){ [ 842.419110] env[61923]: value = "task-1377562" [ 842.419110] env[61923]: _type = "Task" [ 842.419110] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 842.423240] env[61923]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 842.423240] env[61923]: value = "task-1377563" [ 842.423240] env[61923]: _type = "Task" [ 842.423240] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 842.436038] env[61923]: DEBUG oslo_vmware.api [None req-fc756c76-ebec-403f-b8ba-39ff93058631 tempest-ServerShowV257Test-986496257 tempest-ServerShowV257Test-986496257-project-member] Task: {'id': task-1377562, 'name': PowerOnVM_Task} progress is 33%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.440300] env[61923]: DEBUG oslo_vmware.api [-] Task: {'id': task-1377563, 'name': CreateVM_Task} progress is 6%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.449311] env[61923]: DEBUG oslo_concurrency.lockutils [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 842.580071] env[61923]: DEBUG oslo_concurrency.lockutils [None req-80740a2c-9779-4afc-be30-c4469cd3724d tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.392s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 842.580634] env[61923]: DEBUG nova.compute.manager [None req-80740a2c-9779-4afc-be30-c4469cd3724d tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] [instance: fc4fdf89-4615-4586-9dcd-ea90dc258361] Start building networks asynchronously for instance. {{(pid=61923) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 842.583915] env[61923]: DEBUG oslo_concurrency.lockutils [None req-1214e428-bf00-4cbe-9e1e-7c8be1612c36 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.844s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 842.585537] env[61923]: INFO nova.compute.claims [None req-1214e428-bf00-4cbe-9e1e-7c8be1612c36 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] [instance: af3c317a-4007-4cea-a060-1e7dde5ce49e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 842.624506] env[61923]: DEBUG oslo_vmware.api [None req-950e07c2-502c-43fa-8598-da5f189b07e3 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Task: {'id': task-1377558, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.631651] env[61923]: DEBUG nova.network.neutron [req-95afea4b-0803-4361-b7e1-8266a93cb7f6 req-db5b8fab-4416-4a51-89b5-ed2c16d5ad2c service nova] [instance: 73692517-1816-4e9b-ab2d-8265c683e83d] Updated VIF entry in instance network info cache for port dbed34c7-9716-4c24-8f0e-2298de6cf4b9. {{(pid=61923) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 842.631987] env[61923]: DEBUG nova.network.neutron [req-95afea4b-0803-4361-b7e1-8266a93cb7f6 req-db5b8fab-4416-4a51-89b5-ed2c16d5ad2c service nova] [instance: 73692517-1816-4e9b-ab2d-8265c683e83d] Updating instance_info_cache with network_info: [{"id": "dbed34c7-9716-4c24-8f0e-2298de6cf4b9", "address": "fa:16:3e:a0:e2:bc", "network": {"id": "d1c38272-0461-43e5-93ae-3964346f77d0", "bridge": "br-int", "label": "tempest-ServersTestJSON-151839165-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "69cc941eb6dd4780ac12aa29656c37f7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8ba3bd22-c936-470e-89bd-b3a5587e87a0", "external-id": "nsx-vlan-transportzone-605", "segmentation_id": 605, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdbed34c7-97", "ovs_interfaceid": "dbed34c7-9716-4c24-8f0e-2298de6cf4b9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 842.666824] env[61923]: DEBUG oslo_vmware.api [None req-9ff8b121-368d-4886-9882-962e7ef81bcf tempest-ServerTagsTestJSON-19215696 tempest-ServerTagsTestJSON-19215696-project-member] Task: {'id': task-1377559, 'name': Rename_Task, 'duration_secs': 0.13716} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 842.667118] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-9ff8b121-368d-4886-9882-962e7ef81bcf tempest-ServerTagsTestJSON-19215696 tempest-ServerTagsTestJSON-19215696-project-member] [instance: 48825592-dcdc-41d5-9fbf-500d1f31017b] Powering on the VM {{(pid=61923) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 842.667365] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5a676e7f-a295-4fb0-819f-d360b2b4fee5 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.674064] env[61923]: DEBUG oslo_vmware.api [None req-9ff8b121-368d-4886-9882-962e7ef81bcf tempest-ServerTagsTestJSON-19215696 tempest-ServerTagsTestJSON-19215696-project-member] Waiting for the task: (returnval){ [ 842.674064] env[61923]: value = "task-1377564" [ 842.674064] env[61923]: _type = "Task" [ 842.674064] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 842.682847] env[61923]: DEBUG oslo_vmware.api [None req-9ff8b121-368d-4886-9882-962e7ef81bcf tempest-ServerTagsTestJSON-19215696 tempest-ServerTagsTestJSON-19215696-project-member] Task: {'id': task-1377564, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.846998] env[61923]: INFO nova.compute.manager [-] [instance: 1d1b1dc4-c701-42b2-beba-344d622aef41] Took 1.02 seconds to deallocate network for instance. [ 842.862917] env[61923]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-1ba79911-30e3-4352-816b-3a59a5e61169 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.873035] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e58c98a-d8d9-4b0c-94e0-477930878c9b {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.898524] env[61923]: DEBUG nova.compute.manager [req-368ff499-4edf-42ea-8dbc-ce6f7688a61f req-95eb5e71-f848-4490-bb3e-554df6660253 service nova] [instance: 1d1b1dc4-c701-42b2-beba-344d622aef41] Detach interface failed, port_id=8118febd-f949-4bed-9959-775f44cfbd07, reason: Instance 1d1b1dc4-c701-42b2-beba-344d622aef41 could not be found. {{(pid=61923) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 842.934421] env[61923]: INFO nova.compute.manager [None req-458eed44-c3ca-4049-bcef-fbcaf60c50ed tempest-ServersTestManualDisk-2012652654 tempest-ServersTestManualDisk-2012652654-project-member] [instance: f1839f61-4314-48fe-8ab6-14b5e49d569d] Took 27.27 seconds to build instance. [ 842.939914] env[61923]: DEBUG oslo_vmware.api [None req-fc756c76-ebec-403f-b8ba-39ff93058631 tempest-ServerShowV257Test-986496257 tempest-ServerShowV257Test-986496257-project-member] Task: {'id': task-1377562, 'name': PowerOnVM_Task} progress is 88%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 842.944143] env[61923]: DEBUG oslo_vmware.api [-] Task: {'id': task-1377563, 'name': CreateVM_Task} progress is 99%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.096793] env[61923]: DEBUG nova.compute.utils [None req-80740a2c-9779-4afc-be30-c4469cd3724d tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] Using /dev/sd instead of None {{(pid=61923) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 843.098507] env[61923]: DEBUG nova.compute.manager [None req-80740a2c-9779-4afc-be30-c4469cd3724d tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] [instance: fc4fdf89-4615-4586-9dcd-ea90dc258361] Allocating IP information in the background. {{(pid=61923) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 843.098679] env[61923]: DEBUG nova.network.neutron [None req-80740a2c-9779-4afc-be30-c4469cd3724d tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] [instance: fc4fdf89-4615-4586-9dcd-ea90dc258361] allocate_for_instance() {{(pid=61923) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 843.124739] env[61923]: DEBUG oslo_vmware.api [None req-950e07c2-502c-43fa-8598-da5f189b07e3 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Task: {'id': task-1377558, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.518043} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 843.125329] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-950e07c2-502c-43fa-8598-da5f189b07e3 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk to [datastore2] 422d06d6-6932-46c3-bb25-841e6f65c028/422d06d6-6932-46c3-bb25-841e6f65c028.vmdk {{(pid=61923) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 843.125541] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-950e07c2-502c-43fa-8598-da5f189b07e3 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] [instance: 422d06d6-6932-46c3-bb25-841e6f65c028] Extending root virtual disk to 1048576 {{(pid=61923) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 843.125785] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-2b8357a0-8c82-4103-a6e6-1c679d449061 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.132247] env[61923]: DEBUG oslo_vmware.api [None req-950e07c2-502c-43fa-8598-da5f189b07e3 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Waiting for the task: (returnval){ [ 843.132247] env[61923]: value = "task-1377565" [ 843.132247] env[61923]: _type = "Task" [ 843.132247] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 843.135583] env[61923]: DEBUG oslo_concurrency.lockutils [req-95afea4b-0803-4361-b7e1-8266a93cb7f6 req-db5b8fab-4416-4a51-89b5-ed2c16d5ad2c service nova] Releasing lock "refresh_cache-73692517-1816-4e9b-ab2d-8265c683e83d" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 843.140174] env[61923]: DEBUG oslo_vmware.api [None req-950e07c2-502c-43fa-8598-da5f189b07e3 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Task: {'id': task-1377565, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.189020] env[61923]: DEBUG oslo_vmware.api [None req-9ff8b121-368d-4886-9882-962e7ef81bcf tempest-ServerTagsTestJSON-19215696 tempest-ServerTagsTestJSON-19215696-project-member] Task: {'id': task-1377564, 'name': PowerOnVM_Task} progress is 81%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.211119] env[61923]: DEBUG nova.policy [None req-80740a2c-9779-4afc-be30-c4469cd3724d tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e423cb2e47934db39ae2c17a22ee0ceb', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c304b5b5d4df485fbda55a570494a62f', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61923) authorize /opt/stack/nova/nova/policy.py:201}} [ 843.358167] env[61923]: DEBUG oslo_concurrency.lockutils [None req-7d4ecc17-51d2-4753-8cb6-fd83e220b388 tempest-ServersNegativeTestMultiTenantJSON-1497120864 tempest-ServersNegativeTestMultiTenantJSON-1497120864-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 843.436038] env[61923]: DEBUG oslo_concurrency.lockutils [None req-458eed44-c3ca-4049-bcef-fbcaf60c50ed tempest-ServersTestManualDisk-2012652654 tempest-ServersTestManualDisk-2012652654-project-member] Lock "f1839f61-4314-48fe-8ab6-14b5e49d569d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 103.251s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 843.440613] env[61923]: DEBUG oslo_vmware.api [None req-fc756c76-ebec-403f-b8ba-39ff93058631 tempest-ServerShowV257Test-986496257 tempest-ServerShowV257Test-986496257-project-member] Task: {'id': task-1377562, 'name': PowerOnVM_Task, 'duration_secs': 0.876272} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 843.446261] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-fc756c76-ebec-403f-b8ba-39ff93058631 tempest-ServerShowV257Test-986496257 tempest-ServerShowV257Test-986496257-project-member] [instance: b8c60af7-3561-468c-a9a5-1c645080fc69] Powered on the VM {{(pid=61923) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 843.446261] env[61923]: DEBUG nova.compute.manager [None req-fc756c76-ebec-403f-b8ba-39ff93058631 tempest-ServerShowV257Test-986496257 tempest-ServerShowV257Test-986496257-project-member] [instance: b8c60af7-3561-468c-a9a5-1c645080fc69] Checking state {{(pid=61923) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 843.446261] env[61923]: DEBUG oslo_vmware.api [-] Task: {'id': task-1377563, 'name': CreateVM_Task, 'duration_secs': 0.646583} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 843.446261] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe7f926e-7e1a-46f7-8e7c-e2fe418ba71e {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.448630] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 73692517-1816-4e9b-ab2d-8265c683e83d] Created VM on the ESX host {{(pid=61923) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 843.449421] env[61923]: DEBUG oslo_concurrency.lockutils [None req-e62df7da-42b5-4263-bcf8-001a2d06427a tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 843.450282] env[61923]: DEBUG oslo_concurrency.lockutils [None req-e62df7da-42b5-4263-bcf8-001a2d06427a tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 843.450282] env[61923]: DEBUG oslo_concurrency.lockutils [None req-e62df7da-42b5-4263-bcf8-001a2d06427a tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 843.451041] env[61923]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-55181800-b4b6-40d3-9851-7b721a89fac7 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.460569] env[61923]: DEBUG oslo_vmware.api [None req-e62df7da-42b5-4263-bcf8-001a2d06427a tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Waiting for the task: (returnval){ [ 843.460569] env[61923]: value = "session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52219b0a-effa-d532-43be-2d06fa2e622a" [ 843.460569] env[61923]: _type = "Task" [ 843.460569] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 843.468875] env[61923]: DEBUG oslo_vmware.api [None req-e62df7da-42b5-4263-bcf8-001a2d06427a tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52219b0a-effa-d532-43be-2d06fa2e622a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.603943] env[61923]: DEBUG nova.compute.manager [None req-80740a2c-9779-4afc-be30-c4469cd3724d tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] [instance: fc4fdf89-4615-4586-9dcd-ea90dc258361] Start building block device mappings for instance. {{(pid=61923) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 843.628076] env[61923]: DEBUG nova.network.neutron [None req-513daa64-2b05-4c0d-94e9-55b95de63eec tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] [instance: 59198f4d-4dde-4eaf-9f6c-a962cbe53c6e] Successfully updated port: cf99020c-d7cc-4912-a021-8861628b7540 {{(pid=61923) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 843.642615] env[61923]: DEBUG oslo_vmware.api [None req-950e07c2-502c-43fa-8598-da5f189b07e3 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Task: {'id': task-1377565, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.070066} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 843.642615] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-950e07c2-502c-43fa-8598-da5f189b07e3 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] [instance: 422d06d6-6932-46c3-bb25-841e6f65c028] Extended root virtual disk {{(pid=61923) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 843.643403] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-452fe6bb-c031-4ec9-a866-411ed51929db {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.671520] env[61923]: DEBUG nova.virt.vmwareapi.volumeops [None req-950e07c2-502c-43fa-8598-da5f189b07e3 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] [instance: 422d06d6-6932-46c3-bb25-841e6f65c028] Reconfiguring VM instance instance-00000043 to attach disk [datastore2] 422d06d6-6932-46c3-bb25-841e6f65c028/422d06d6-6932-46c3-bb25-841e6f65c028.vmdk or device None with type sparse {{(pid=61923) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 843.675594] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ff0b0af4-d543-482d-9654-1d7ee4140586 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.701893] env[61923]: DEBUG oslo_vmware.api [None req-9ff8b121-368d-4886-9882-962e7ef81bcf tempest-ServerTagsTestJSON-19215696 tempest-ServerTagsTestJSON-19215696-project-member] Task: {'id': task-1377564, 'name': PowerOnVM_Task, 'duration_secs': 0.909131} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 843.703312] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-9ff8b121-368d-4886-9882-962e7ef81bcf tempest-ServerTagsTestJSON-19215696 tempest-ServerTagsTestJSON-19215696-project-member] [instance: 48825592-dcdc-41d5-9fbf-500d1f31017b] Powered on the VM {{(pid=61923) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 843.703525] env[61923]: INFO nova.compute.manager [None req-9ff8b121-368d-4886-9882-962e7ef81bcf tempest-ServerTagsTestJSON-19215696 tempest-ServerTagsTestJSON-19215696-project-member] [instance: 48825592-dcdc-41d5-9fbf-500d1f31017b] Took 8.48 seconds to spawn the instance on the hypervisor. [ 843.703702] env[61923]: DEBUG nova.compute.manager [None req-9ff8b121-368d-4886-9882-962e7ef81bcf tempest-ServerTagsTestJSON-19215696 tempest-ServerTagsTestJSON-19215696-project-member] [instance: 48825592-dcdc-41d5-9fbf-500d1f31017b] Checking state {{(pid=61923) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 843.704287] env[61923]: DEBUG oslo_vmware.api [None req-950e07c2-502c-43fa-8598-da5f189b07e3 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Waiting for the task: (returnval){ [ 843.704287] env[61923]: value = "task-1377566" [ 843.704287] env[61923]: _type = "Task" [ 843.704287] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 843.707836] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d79d68e-f048-4a06-bdae-09e712c25682 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.719286] env[61923]: DEBUG oslo_vmware.api [None req-950e07c2-502c-43fa-8598-da5f189b07e3 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Task: {'id': task-1377566, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 843.924531] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68a1d4bb-5710-41b3-9576-2fcbd2bc95e9 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.932533] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4d56942-bfea-49c0-8144-bf0c6b251230 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.965644] env[61923]: DEBUG nova.compute.manager [None req-cdbe0145-cd3c-4551-af87-0c7f8b53f26c tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] [instance: e413c45d-cd89-44d4-9102-3d188907e7bb] Starting instance... {{(pid=61923) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 843.979103] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0bef42e5-93e3-4652-9498-297eeffea3a2 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.982910] env[61923]: DEBUG nova.network.neutron [None req-80740a2c-9779-4afc-be30-c4469cd3724d tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] [instance: fc4fdf89-4615-4586-9dcd-ea90dc258361] Successfully created port: 8781f580-cc8d-4677-b6b3-2e71a0b907bc {{(pid=61923) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 843.986748] env[61923]: DEBUG oslo_concurrency.lockutils [None req-fc756c76-ebec-403f-b8ba-39ff93058631 tempest-ServerShowV257Test-986496257 tempest-ServerShowV257Test-986496257-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 843.994542] env[61923]: DEBUG oslo_vmware.api [None req-e62df7da-42b5-4263-bcf8-001a2d06427a tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52219b0a-effa-d532-43be-2d06fa2e622a, 'name': SearchDatastore_Task, 'duration_secs': 0.013863} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 843.996748] env[61923]: DEBUG oslo_concurrency.lockutils [None req-e62df7da-42b5-4263-bcf8-001a2d06427a tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 843.996993] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-e62df7da-42b5-4263-bcf8-001a2d06427a tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: 73692517-1816-4e9b-ab2d-8265c683e83d] Processing image 0f153f63-ae0a-45b1-b7f5-7f9b673c947f {{(pid=61923) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 843.997251] env[61923]: DEBUG oslo_concurrency.lockutils [None req-e62df7da-42b5-4263-bcf8-001a2d06427a tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 843.997396] env[61923]: DEBUG oslo_concurrency.lockutils [None req-e62df7da-42b5-4263-bcf8-001a2d06427a tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 843.997572] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-e62df7da-42b5-4263-bcf8-001a2d06427a tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61923) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 843.999044] env[61923]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-4fa72650-88ed-4554-bc86-0a59fc9c7adb {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.002273] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff59f99b-f4b8-4ef6-8ba7-687f97e80c7c {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.020795] env[61923]: DEBUG nova.compute.provider_tree [None req-1214e428-bf00-4cbe-9e1e-7c8be1612c36 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 844.023932] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-e62df7da-42b5-4263-bcf8-001a2d06427a tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61923) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 844.024262] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-e62df7da-42b5-4263-bcf8-001a2d06427a tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61923) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 844.025300] env[61923]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5b3e2e84-023a-48b5-92b4-a2f50444d171 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.031658] env[61923]: DEBUG oslo_vmware.api [None req-e62df7da-42b5-4263-bcf8-001a2d06427a tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Waiting for the task: (returnval){ [ 844.031658] env[61923]: value = "session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52c79bf9-120f-b3b9-884d-be9d77e87635" [ 844.031658] env[61923]: _type = "Task" [ 844.031658] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 844.039345] env[61923]: DEBUG oslo_vmware.api [None req-e62df7da-42b5-4263-bcf8-001a2d06427a tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52c79bf9-120f-b3b9-884d-be9d77e87635, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.130261] env[61923]: DEBUG oslo_concurrency.lockutils [None req-513daa64-2b05-4c0d-94e9-55b95de63eec tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Acquiring lock "refresh_cache-59198f4d-4dde-4eaf-9f6c-a962cbe53c6e" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 844.130544] env[61923]: DEBUG oslo_concurrency.lockutils [None req-513daa64-2b05-4c0d-94e9-55b95de63eec tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Acquired lock "refresh_cache-59198f4d-4dde-4eaf-9f6c-a962cbe53c6e" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 844.130846] env[61923]: DEBUG nova.network.neutron [None req-513daa64-2b05-4c0d-94e9-55b95de63eec tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] [instance: 59198f4d-4dde-4eaf-9f6c-a962cbe53c6e] Building network info cache for instance {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 844.219757] env[61923]: DEBUG oslo_vmware.api [None req-950e07c2-502c-43fa-8598-da5f189b07e3 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Task: {'id': task-1377566, 'name': ReconfigVM_Task, 'duration_secs': 0.312873} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 844.220019] env[61923]: DEBUG nova.virt.vmwareapi.volumeops [None req-950e07c2-502c-43fa-8598-da5f189b07e3 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] [instance: 422d06d6-6932-46c3-bb25-841e6f65c028] Reconfigured VM instance instance-00000043 to attach disk [datastore2] 422d06d6-6932-46c3-bb25-841e6f65c028/422d06d6-6932-46c3-bb25-841e6f65c028.vmdk or device None with type sparse {{(pid=61923) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 844.220655] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ded54b2c-7a35-4222-abd6-dde3e7730c01 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.233315] env[61923]: DEBUG oslo_vmware.api [None req-950e07c2-502c-43fa-8598-da5f189b07e3 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Waiting for the task: (returnval){ [ 844.233315] env[61923]: value = "task-1377567" [ 844.233315] env[61923]: _type = "Task" [ 844.233315] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 844.235382] env[61923]: INFO nova.compute.manager [None req-9ff8b121-368d-4886-9882-962e7ef81bcf tempest-ServerTagsTestJSON-19215696 tempest-ServerTagsTestJSON-19215696-project-member] [instance: 48825592-dcdc-41d5-9fbf-500d1f31017b] Took 26.71 seconds to build instance. [ 844.242294] env[61923]: DEBUG oslo_vmware.api [None req-950e07c2-502c-43fa-8598-da5f189b07e3 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Task: {'id': task-1377567, 'name': Rename_Task} progress is 6%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.281755] env[61923]: DEBUG nova.compute.manager [req-01326319-7ed1-4ad3-8e4d-7861d02c02d5 req-85a488d6-2ebf-432b-8efa-aba50ff6327f service nova] [instance: f1839f61-4314-48fe-8ab6-14b5e49d569d] Received event network-changed-f36077f4-1c49-48e7-9c41-0f45a58883e7 {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 844.281935] env[61923]: DEBUG nova.compute.manager [req-01326319-7ed1-4ad3-8e4d-7861d02c02d5 req-85a488d6-2ebf-432b-8efa-aba50ff6327f service nova] [instance: f1839f61-4314-48fe-8ab6-14b5e49d569d] Refreshing instance network info cache due to event network-changed-f36077f4-1c49-48e7-9c41-0f45a58883e7. {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 844.282103] env[61923]: DEBUG oslo_concurrency.lockutils [req-01326319-7ed1-4ad3-8e4d-7861d02c02d5 req-85a488d6-2ebf-432b-8efa-aba50ff6327f service nova] Acquiring lock "refresh_cache-f1839f61-4314-48fe-8ab6-14b5e49d569d" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 844.282249] env[61923]: DEBUG oslo_concurrency.lockutils [req-01326319-7ed1-4ad3-8e4d-7861d02c02d5 req-85a488d6-2ebf-432b-8efa-aba50ff6327f service nova] Acquired lock "refresh_cache-f1839f61-4314-48fe-8ab6-14b5e49d569d" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 844.282408] env[61923]: DEBUG nova.network.neutron [req-01326319-7ed1-4ad3-8e4d-7861d02c02d5 req-85a488d6-2ebf-432b-8efa-aba50ff6327f service nova] [instance: f1839f61-4314-48fe-8ab6-14b5e49d569d] Refreshing network info cache for port f36077f4-1c49-48e7-9c41-0f45a58883e7 {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 844.369530] env[61923]: DEBUG nova.network.neutron [None req-80740a2c-9779-4afc-be30-c4469cd3724d tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] [instance: fc4fdf89-4615-4586-9dcd-ea90dc258361] Successfully created port: a36809ea-2e4a-48ff-ac61-305da772b282 {{(pid=61923) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 844.430519] env[61923]: DEBUG nova.compute.manager [req-f275f677-8863-4edb-838f-1cc42b444f5a req-8700db0e-e32a-4910-9080-1e0ff2c97ad8 service nova] [instance: 59198f4d-4dde-4eaf-9f6c-a962cbe53c6e] Received event network-vif-plugged-cf99020c-d7cc-4912-a021-8861628b7540 {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 844.430854] env[61923]: DEBUG oslo_concurrency.lockutils [req-f275f677-8863-4edb-838f-1cc42b444f5a req-8700db0e-e32a-4910-9080-1e0ff2c97ad8 service nova] Acquiring lock "59198f4d-4dde-4eaf-9f6c-a962cbe53c6e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 844.431185] env[61923]: DEBUG oslo_concurrency.lockutils [req-f275f677-8863-4edb-838f-1cc42b444f5a req-8700db0e-e32a-4910-9080-1e0ff2c97ad8 service nova] Lock "59198f4d-4dde-4eaf-9f6c-a962cbe53c6e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 844.431446] env[61923]: DEBUG oslo_concurrency.lockutils [req-f275f677-8863-4edb-838f-1cc42b444f5a req-8700db0e-e32a-4910-9080-1e0ff2c97ad8 service nova] Lock "59198f4d-4dde-4eaf-9f6c-a962cbe53c6e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 844.431710] env[61923]: DEBUG nova.compute.manager [req-f275f677-8863-4edb-838f-1cc42b444f5a req-8700db0e-e32a-4910-9080-1e0ff2c97ad8 service nova] [instance: 59198f4d-4dde-4eaf-9f6c-a962cbe53c6e] No waiting events found dispatching network-vif-plugged-cf99020c-d7cc-4912-a021-8861628b7540 {{(pid=61923) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 844.431992] env[61923]: WARNING nova.compute.manager [req-f275f677-8863-4edb-838f-1cc42b444f5a req-8700db0e-e32a-4910-9080-1e0ff2c97ad8 service nova] [instance: 59198f4d-4dde-4eaf-9f6c-a962cbe53c6e] Received unexpected event network-vif-plugged-cf99020c-d7cc-4912-a021-8861628b7540 for instance with vm_state building and task_state spawning. [ 844.432271] env[61923]: DEBUG nova.compute.manager [req-f275f677-8863-4edb-838f-1cc42b444f5a req-8700db0e-e32a-4910-9080-1e0ff2c97ad8 service nova] [instance: 59198f4d-4dde-4eaf-9f6c-a962cbe53c6e] Received event network-changed-cf99020c-d7cc-4912-a021-8861628b7540 {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 844.432500] env[61923]: DEBUG nova.compute.manager [req-f275f677-8863-4edb-838f-1cc42b444f5a req-8700db0e-e32a-4910-9080-1e0ff2c97ad8 service nova] [instance: 59198f4d-4dde-4eaf-9f6c-a962cbe53c6e] Refreshing instance network info cache due to event network-changed-cf99020c-d7cc-4912-a021-8861628b7540. {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 844.432733] env[61923]: DEBUG oslo_concurrency.lockutils [req-f275f677-8863-4edb-838f-1cc42b444f5a req-8700db0e-e32a-4910-9080-1e0ff2c97ad8 service nova] Acquiring lock "refresh_cache-59198f4d-4dde-4eaf-9f6c-a962cbe53c6e" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 844.487841] env[61923]: DEBUG oslo_concurrency.lockutils [None req-cdbe0145-cd3c-4551-af87-0c7f8b53f26c tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 844.510973] env[61923]: DEBUG oslo_concurrency.lockutils [None req-62f42696-8467-4135-b6f9-a4440e44fead tempest-ServerShowV257Test-986496257 tempest-ServerShowV257Test-986496257-project-member] Acquiring lock "b8c60af7-3561-468c-a9a5-1c645080fc69" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 844.511255] env[61923]: DEBUG oslo_concurrency.lockutils [None req-62f42696-8467-4135-b6f9-a4440e44fead tempest-ServerShowV257Test-986496257 tempest-ServerShowV257Test-986496257-project-member] Lock "b8c60af7-3561-468c-a9a5-1c645080fc69" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 844.511458] env[61923]: DEBUG oslo_concurrency.lockutils [None req-62f42696-8467-4135-b6f9-a4440e44fead tempest-ServerShowV257Test-986496257 tempest-ServerShowV257Test-986496257-project-member] Acquiring lock "b8c60af7-3561-468c-a9a5-1c645080fc69-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 844.511641] env[61923]: DEBUG oslo_concurrency.lockutils [None req-62f42696-8467-4135-b6f9-a4440e44fead tempest-ServerShowV257Test-986496257 tempest-ServerShowV257Test-986496257-project-member] Lock "b8c60af7-3561-468c-a9a5-1c645080fc69-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 844.511816] env[61923]: DEBUG oslo_concurrency.lockutils [None req-62f42696-8467-4135-b6f9-a4440e44fead tempest-ServerShowV257Test-986496257 tempest-ServerShowV257Test-986496257-project-member] Lock "b8c60af7-3561-468c-a9a5-1c645080fc69-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 844.513907] env[61923]: INFO nova.compute.manager [None req-62f42696-8467-4135-b6f9-a4440e44fead tempest-ServerShowV257Test-986496257 tempest-ServerShowV257Test-986496257-project-member] [instance: b8c60af7-3561-468c-a9a5-1c645080fc69] Terminating instance [ 844.515696] env[61923]: DEBUG oslo_concurrency.lockutils [None req-62f42696-8467-4135-b6f9-a4440e44fead tempest-ServerShowV257Test-986496257 tempest-ServerShowV257Test-986496257-project-member] Acquiring lock "refresh_cache-b8c60af7-3561-468c-a9a5-1c645080fc69" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 844.515850] env[61923]: DEBUG oslo_concurrency.lockutils [None req-62f42696-8467-4135-b6f9-a4440e44fead tempest-ServerShowV257Test-986496257 tempest-ServerShowV257Test-986496257-project-member] Acquired lock "refresh_cache-b8c60af7-3561-468c-a9a5-1c645080fc69" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 844.516027] env[61923]: DEBUG nova.network.neutron [None req-62f42696-8467-4135-b6f9-a4440e44fead tempest-ServerShowV257Test-986496257 tempest-ServerShowV257Test-986496257-project-member] [instance: b8c60af7-3561-468c-a9a5-1c645080fc69] Building network info cache for instance {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 844.526545] env[61923]: DEBUG nova.scheduler.client.report [None req-1214e428-bf00-4cbe-9e1e-7c8be1612c36 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 844.542342] env[61923]: DEBUG oslo_vmware.api [None req-e62df7da-42b5-4263-bcf8-001a2d06427a tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52c79bf9-120f-b3b9-884d-be9d77e87635, 'name': SearchDatastore_Task, 'duration_secs': 0.019195} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 844.543531] env[61923]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1e00e992-d98b-4ac3-bf09-e337c169c145 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.550246] env[61923]: DEBUG oslo_vmware.api [None req-e62df7da-42b5-4263-bcf8-001a2d06427a tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Waiting for the task: (returnval){ [ 844.550246] env[61923]: value = "session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52499389-8e2a-28af-c8aa-590475b7f839" [ 844.550246] env[61923]: _type = "Task" [ 844.550246] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 844.558975] env[61923]: DEBUG oslo_vmware.api [None req-e62df7da-42b5-4263-bcf8-001a2d06427a tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52499389-8e2a-28af-c8aa-590475b7f839, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.612693] env[61923]: DEBUG nova.compute.manager [None req-80740a2c-9779-4afc-be30-c4469cd3724d tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] [instance: fc4fdf89-4615-4586-9dcd-ea90dc258361] Start spawning the instance on the hypervisor. {{(pid=61923) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 844.652667] env[61923]: DEBUG nova.virt.hardware [None req-80740a2c-9779-4afc-be30-c4469cd3724d tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-29T20:07:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-29T20:07:35Z,direct_url=,disk_format='vmdk',id=0f153f63-ae0a-45b1-b7f5-7f9b673c947f,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='4e7b5e3b3c3443c8bd5c70f8a15739f5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-29T20:07:36Z,virtual_size=,visibility=), allow threads: False {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 844.653032] env[61923]: DEBUG nova.virt.hardware [None req-80740a2c-9779-4afc-be30-c4469cd3724d tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] Flavor limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 844.653899] env[61923]: DEBUG nova.virt.hardware [None req-80740a2c-9779-4afc-be30-c4469cd3724d tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] Image limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 844.654077] env[61923]: DEBUG nova.virt.hardware [None req-80740a2c-9779-4afc-be30-c4469cd3724d tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] Flavor pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 844.654281] env[61923]: DEBUG nova.virt.hardware [None req-80740a2c-9779-4afc-be30-c4469cd3724d tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] Image pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 844.654467] env[61923]: DEBUG nova.virt.hardware [None req-80740a2c-9779-4afc-be30-c4469cd3724d tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 844.654695] env[61923]: DEBUG nova.virt.hardware [None req-80740a2c-9779-4afc-be30-c4469cd3724d tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 844.655075] env[61923]: DEBUG nova.virt.hardware [None req-80740a2c-9779-4afc-be30-c4469cd3724d tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 844.655201] env[61923]: DEBUG nova.virt.hardware [None req-80740a2c-9779-4afc-be30-c4469cd3724d tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] Got 1 possible topologies {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 844.655428] env[61923]: DEBUG nova.virt.hardware [None req-80740a2c-9779-4afc-be30-c4469cd3724d tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 844.655681] env[61923]: DEBUG nova.virt.hardware [None req-80740a2c-9779-4afc-be30-c4469cd3724d tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 844.660019] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09f9553b-164c-4ef5-a893-effd95e203a8 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.665838] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50729446-9c54-417e-8883-ccb9a1604b97 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.671053] env[61923]: DEBUG nova.network.neutron [None req-513daa64-2b05-4c0d-94e9-55b95de63eec tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] [instance: 59198f4d-4dde-4eaf-9f6c-a962cbe53c6e] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 844.737424] env[61923]: DEBUG oslo_concurrency.lockutils [None req-9ff8b121-368d-4886-9882-962e7ef81bcf tempest-ServerTagsTestJSON-19215696 tempest-ServerTagsTestJSON-19215696-project-member] Lock "48825592-dcdc-41d5-9fbf-500d1f31017b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 88.433s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 844.742742] env[61923]: DEBUG oslo_vmware.api [None req-950e07c2-502c-43fa-8598-da5f189b07e3 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Task: {'id': task-1377567, 'name': Rename_Task, 'duration_secs': 0.225229} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 844.745731] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-950e07c2-502c-43fa-8598-da5f189b07e3 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] [instance: 422d06d6-6932-46c3-bb25-841e6f65c028] Powering on the VM {{(pid=61923) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 844.745989] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-4b25cd78-2aa8-4d0a-b581-6dcc91c83649 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.753799] env[61923]: DEBUG oslo_vmware.api [None req-950e07c2-502c-43fa-8598-da5f189b07e3 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Waiting for the task: (returnval){ [ 844.753799] env[61923]: value = "task-1377568" [ 844.753799] env[61923]: _type = "Task" [ 844.753799] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 844.765413] env[61923]: DEBUG oslo_vmware.api [None req-950e07c2-502c-43fa-8598-da5f189b07e3 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Task: {'id': task-1377568, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 844.814805] env[61923]: DEBUG nova.network.neutron [None req-80740a2c-9779-4afc-be30-c4469cd3724d tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] [instance: fc4fdf89-4615-4586-9dcd-ea90dc258361] Successfully created port: 0738a978-9439-4919-8894-4b264480593e {{(pid=61923) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 844.997544] env[61923]: DEBUG nova.network.neutron [None req-513daa64-2b05-4c0d-94e9-55b95de63eec tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] [instance: 59198f4d-4dde-4eaf-9f6c-a962cbe53c6e] Updating instance_info_cache with network_info: [{"id": "cf99020c-d7cc-4912-a021-8861628b7540", "address": "fa:16:3e:89:5c:35", "network": {"id": "b8c2e837-0d64-4b07-9cc5-a66c8e8db1e9", "bridge": "br-int", "label": "tempest-ImagesTestJSON-133215580-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7abfe95660904cd4b8824257795d8e8a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f2c424c9-6446-4b2a-af8c-4d9c29117c39", "external-id": "nsx-vlan-transportzone-437", "segmentation_id": 437, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcf99020c-d7", "ovs_interfaceid": "cf99020c-d7cc-4912-a021-8861628b7540", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 845.032015] env[61923]: DEBUG oslo_concurrency.lockutils [None req-1214e428-bf00-4cbe-9e1e-7c8be1612c36 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.448s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 845.032643] env[61923]: DEBUG nova.compute.manager [None req-1214e428-bf00-4cbe-9e1e-7c8be1612c36 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] [instance: af3c317a-4007-4cea-a060-1e7dde5ce49e] Start building networks asynchronously for instance. {{(pid=61923) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 845.035456] env[61923]: DEBUG oslo_concurrency.lockutils [None req-52eadf86-cc45-4cae-af2a-645ac3cb8658 tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.410s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 845.037031] env[61923]: INFO nova.compute.claims [None req-52eadf86-cc45-4cae-af2a-645ac3cb8658 tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] [instance: 87ae37dc-d0d0-4b76-afdd-0ba3e8f6ce0b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 845.047100] env[61923]: DEBUG nova.network.neutron [req-01326319-7ed1-4ad3-8e4d-7861d02c02d5 req-85a488d6-2ebf-432b-8efa-aba50ff6327f service nova] [instance: f1839f61-4314-48fe-8ab6-14b5e49d569d] Updated VIF entry in instance network info cache for port f36077f4-1c49-48e7-9c41-0f45a58883e7. {{(pid=61923) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 845.047100] env[61923]: DEBUG nova.network.neutron [req-01326319-7ed1-4ad3-8e4d-7861d02c02d5 req-85a488d6-2ebf-432b-8efa-aba50ff6327f service nova] [instance: f1839f61-4314-48fe-8ab6-14b5e49d569d] Updating instance_info_cache with network_info: [{"id": "f36077f4-1c49-48e7-9c41-0f45a58883e7", "address": "fa:16:3e:f2:ba:61", "network": {"id": "df531efc-77d0-447e-b85e-6323fbc26dfb", "bridge": "br-int", "label": "tempest-ServersTestManualDisk-1619780476-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.232", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5357434049fb4acd910988fc5ff0d2c4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "20fbc86c-0935-406c-9659-3ffc33fc0d08", "external-id": "nsx-vlan-transportzone-191", "segmentation_id": 191, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf36077f4-1c", "ovs_interfaceid": "f36077f4-1c49-48e7-9c41-0f45a58883e7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 845.047279] env[61923]: DEBUG nova.network.neutron [None req-62f42696-8467-4135-b6f9-a4440e44fead tempest-ServerShowV257Test-986496257 tempest-ServerShowV257Test-986496257-project-member] [instance: b8c60af7-3561-468c-a9a5-1c645080fc69] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 845.062013] env[61923]: DEBUG oslo_vmware.api [None req-e62df7da-42b5-4263-bcf8-001a2d06427a tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52499389-8e2a-28af-c8aa-590475b7f839, 'name': SearchDatastore_Task, 'duration_secs': 0.011047} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 845.062266] env[61923]: DEBUG oslo_concurrency.lockutils [None req-e62df7da-42b5-4263-bcf8-001a2d06427a tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 845.062541] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-e62df7da-42b5-4263-bcf8-001a2d06427a tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk to [datastore2] 73692517-1816-4e9b-ab2d-8265c683e83d/73692517-1816-4e9b-ab2d-8265c683e83d.vmdk {{(pid=61923) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 845.062796] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-875ceb97-1455-4dbb-aefe-55b688b7cc35 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.072425] env[61923]: DEBUG oslo_vmware.api [None req-e62df7da-42b5-4263-bcf8-001a2d06427a tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Waiting for the task: (returnval){ [ 845.072425] env[61923]: value = "task-1377569" [ 845.072425] env[61923]: _type = "Task" [ 845.072425] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 845.084330] env[61923]: DEBUG oslo_vmware.api [None req-e62df7da-42b5-4263-bcf8-001a2d06427a tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Task: {'id': task-1377569, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.157721] env[61923]: DEBUG nova.network.neutron [None req-62f42696-8467-4135-b6f9-a4440e44fead tempest-ServerShowV257Test-986496257 tempest-ServerShowV257Test-986496257-project-member] [instance: b8c60af7-3561-468c-a9a5-1c645080fc69] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 845.246712] env[61923]: DEBUG nova.compute.manager [None req-13baf863-98e6-41c7-ae75-48fd1a9cef32 tempest-ServerAddressesTestJSON-1846886542 tempest-ServerAddressesTestJSON-1846886542-project-member] [instance: 5b73e07a-d6ef-4dcf-bdae-eea91d2aeb6f] Starting instance... {{(pid=61923) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 845.264414] env[61923]: DEBUG oslo_vmware.api [None req-950e07c2-502c-43fa-8598-da5f189b07e3 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Task: {'id': task-1377568, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.500817] env[61923]: DEBUG oslo_concurrency.lockutils [None req-513daa64-2b05-4c0d-94e9-55b95de63eec tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Releasing lock "refresh_cache-59198f4d-4dde-4eaf-9f6c-a962cbe53c6e" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 845.501246] env[61923]: DEBUG nova.compute.manager [None req-513daa64-2b05-4c0d-94e9-55b95de63eec tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] [instance: 59198f4d-4dde-4eaf-9f6c-a962cbe53c6e] Instance network_info: |[{"id": "cf99020c-d7cc-4912-a021-8861628b7540", "address": "fa:16:3e:89:5c:35", "network": {"id": "b8c2e837-0d64-4b07-9cc5-a66c8e8db1e9", "bridge": "br-int", "label": "tempest-ImagesTestJSON-133215580-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7abfe95660904cd4b8824257795d8e8a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f2c424c9-6446-4b2a-af8c-4d9c29117c39", "external-id": "nsx-vlan-transportzone-437", "segmentation_id": 437, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcf99020c-d7", "ovs_interfaceid": "cf99020c-d7cc-4912-a021-8861628b7540", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61923) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 845.501639] env[61923]: DEBUG oslo_concurrency.lockutils [req-f275f677-8863-4edb-838f-1cc42b444f5a req-8700db0e-e32a-4910-9080-1e0ff2c97ad8 service nova] Acquired lock "refresh_cache-59198f4d-4dde-4eaf-9f6c-a962cbe53c6e" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 845.501858] env[61923]: DEBUG nova.network.neutron [req-f275f677-8863-4edb-838f-1cc42b444f5a req-8700db0e-e32a-4910-9080-1e0ff2c97ad8 service nova] [instance: 59198f4d-4dde-4eaf-9f6c-a962cbe53c6e] Refreshing network info cache for port cf99020c-d7cc-4912-a021-8861628b7540 {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 845.503466] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-513daa64-2b05-4c0d-94e9-55b95de63eec tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] [instance: 59198f4d-4dde-4eaf-9f6c-a962cbe53c6e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:89:5c:35', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'f2c424c9-6446-4b2a-af8c-4d9c29117c39', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'cf99020c-d7cc-4912-a021-8861628b7540', 'vif_model': 'vmxnet3'}] {{(pid=61923) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 845.513581] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-513daa64-2b05-4c0d-94e9-55b95de63eec tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Creating folder: Project (7abfe95660904cd4b8824257795d8e8a). Parent ref: group-v292629. {{(pid=61923) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 845.516749] env[61923]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-b2262643-d26a-4825-b73f-1d6f7e59d59c {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.529926] env[61923]: INFO nova.virt.vmwareapi.vm_util [None req-513daa64-2b05-4c0d-94e9-55b95de63eec tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Created folder: Project (7abfe95660904cd4b8824257795d8e8a) in parent group-v292629. [ 845.530169] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-513daa64-2b05-4c0d-94e9-55b95de63eec tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Creating folder: Instances. Parent ref: group-v292669. {{(pid=61923) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 845.530420] env[61923]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-9f8ba308-5524-4d9d-9183-a07a792f98c3 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.542653] env[61923]: DEBUG nova.compute.utils [None req-1214e428-bf00-4cbe-9e1e-7c8be1612c36 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] Using /dev/sd instead of None {{(pid=61923) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 845.548061] env[61923]: INFO nova.virt.vmwareapi.vm_util [None req-513daa64-2b05-4c0d-94e9-55b95de63eec tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Created folder: Instances in parent group-v292669. [ 845.548061] env[61923]: DEBUG oslo.service.loopingcall [None req-513daa64-2b05-4c0d-94e9-55b95de63eec tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61923) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 845.548061] env[61923]: DEBUG nova.compute.manager [None req-1214e428-bf00-4cbe-9e1e-7c8be1612c36 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] [instance: af3c317a-4007-4cea-a060-1e7dde5ce49e] Allocating IP information in the background. {{(pid=61923) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 845.548061] env[61923]: DEBUG nova.network.neutron [None req-1214e428-bf00-4cbe-9e1e-7c8be1612c36 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] [instance: af3c317a-4007-4cea-a060-1e7dde5ce49e] allocate_for_instance() {{(pid=61923) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 845.550721] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 59198f4d-4dde-4eaf-9f6c-a962cbe53c6e] Creating VM on the ESX host {{(pid=61923) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 845.550721] env[61923]: DEBUG oslo_concurrency.lockutils [req-01326319-7ed1-4ad3-8e4d-7861d02c02d5 req-85a488d6-2ebf-432b-8efa-aba50ff6327f service nova] Releasing lock "refresh_cache-f1839f61-4314-48fe-8ab6-14b5e49d569d" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 845.550721] env[61923]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-a09309d4-897f-47d5-ae8a-0f7a4af18780 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.584635] env[61923]: DEBUG oslo_vmware.api [None req-e62df7da-42b5-4263-bcf8-001a2d06427a tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Task: {'id': task-1377569, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.585969] env[61923]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 845.585969] env[61923]: value = "task-1377572" [ 845.585969] env[61923]: _type = "Task" [ 845.585969] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 845.595190] env[61923]: DEBUG oslo_vmware.api [-] Task: {'id': task-1377572, 'name': CreateVM_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.618501] env[61923]: DEBUG nova.policy [None req-1214e428-bf00-4cbe-9e1e-7c8be1612c36 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '629d5a54afa44599a4c0756d997cabc1', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '87f548d252244b759e3f157af9ab875d', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61923) authorize /opt/stack/nova/nova/policy.py:201}} [ 845.659562] env[61923]: DEBUG oslo_concurrency.lockutils [None req-f39c3939-2621-40b9-927d-ae5b7aafe54d tempest-ServerTagsTestJSON-19215696 tempest-ServerTagsTestJSON-19215696-project-member] Acquiring lock "48825592-dcdc-41d5-9fbf-500d1f31017b" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 845.659861] env[61923]: DEBUG oslo_concurrency.lockutils [None req-f39c3939-2621-40b9-927d-ae5b7aafe54d tempest-ServerTagsTestJSON-19215696 tempest-ServerTagsTestJSON-19215696-project-member] Lock "48825592-dcdc-41d5-9fbf-500d1f31017b" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 845.660079] env[61923]: DEBUG oslo_concurrency.lockutils [None req-f39c3939-2621-40b9-927d-ae5b7aafe54d tempest-ServerTagsTestJSON-19215696 tempest-ServerTagsTestJSON-19215696-project-member] Acquiring lock "48825592-dcdc-41d5-9fbf-500d1f31017b-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 845.660265] env[61923]: DEBUG oslo_concurrency.lockutils [None req-f39c3939-2621-40b9-927d-ae5b7aafe54d tempest-ServerTagsTestJSON-19215696 tempest-ServerTagsTestJSON-19215696-project-member] Lock "48825592-dcdc-41d5-9fbf-500d1f31017b-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 845.660578] env[61923]: DEBUG oslo_concurrency.lockutils [None req-f39c3939-2621-40b9-927d-ae5b7aafe54d tempest-ServerTagsTestJSON-19215696 tempest-ServerTagsTestJSON-19215696-project-member] Lock "48825592-dcdc-41d5-9fbf-500d1f31017b-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 845.662939] env[61923]: DEBUG oslo_concurrency.lockutils [None req-62f42696-8467-4135-b6f9-a4440e44fead tempest-ServerShowV257Test-986496257 tempest-ServerShowV257Test-986496257-project-member] Releasing lock "refresh_cache-b8c60af7-3561-468c-a9a5-1c645080fc69" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 845.663361] env[61923]: DEBUG nova.compute.manager [None req-62f42696-8467-4135-b6f9-a4440e44fead tempest-ServerShowV257Test-986496257 tempest-ServerShowV257Test-986496257-project-member] [instance: b8c60af7-3561-468c-a9a5-1c645080fc69] Start destroying the instance on the hypervisor. {{(pid=61923) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 845.663540] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-62f42696-8467-4135-b6f9-a4440e44fead tempest-ServerShowV257Test-986496257 tempest-ServerShowV257Test-986496257-project-member] [instance: b8c60af7-3561-468c-a9a5-1c645080fc69] Destroying instance {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 845.664119] env[61923]: INFO nova.compute.manager [None req-f39c3939-2621-40b9-927d-ae5b7aafe54d tempest-ServerTagsTestJSON-19215696 tempest-ServerTagsTestJSON-19215696-project-member] [instance: 48825592-dcdc-41d5-9fbf-500d1f31017b] Terminating instance [ 845.666340] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67562a60-427e-41ed-a1a7-4e7d7952b73a {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.669719] env[61923]: DEBUG nova.compute.manager [None req-f39c3939-2621-40b9-927d-ae5b7aafe54d tempest-ServerTagsTestJSON-19215696 tempest-ServerTagsTestJSON-19215696-project-member] [instance: 48825592-dcdc-41d5-9fbf-500d1f31017b] Start destroying the instance on the hypervisor. {{(pid=61923) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 845.669950] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-f39c3939-2621-40b9-927d-ae5b7aafe54d tempest-ServerTagsTestJSON-19215696 tempest-ServerTagsTestJSON-19215696-project-member] [instance: 48825592-dcdc-41d5-9fbf-500d1f31017b] Destroying instance {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 845.670775] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28a2b2be-3b1f-4b76-a8dd-b929de04b23c {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.681054] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-f39c3939-2621-40b9-927d-ae5b7aafe54d tempest-ServerTagsTestJSON-19215696 tempest-ServerTagsTestJSON-19215696-project-member] [instance: 48825592-dcdc-41d5-9fbf-500d1f31017b] Powering off the VM {{(pid=61923) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 845.682023] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-62f42696-8467-4135-b6f9-a4440e44fead tempest-ServerShowV257Test-986496257 tempest-ServerShowV257Test-986496257-project-member] [instance: b8c60af7-3561-468c-a9a5-1c645080fc69] Powering off the VM {{(pid=61923) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 845.684665] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e17facf9-74e7-483f-b5e4-8291a7362077 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.685744] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2f193de4-1a36-4dd3-899e-8860abefbab5 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.693251] env[61923]: DEBUG oslo_vmware.api [None req-62f42696-8467-4135-b6f9-a4440e44fead tempest-ServerShowV257Test-986496257 tempest-ServerShowV257Test-986496257-project-member] Waiting for the task: (returnval){ [ 845.693251] env[61923]: value = "task-1377574" [ 845.693251] env[61923]: _type = "Task" [ 845.693251] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 845.694764] env[61923]: DEBUG oslo_vmware.api [None req-f39c3939-2621-40b9-927d-ae5b7aafe54d tempest-ServerTagsTestJSON-19215696 tempest-ServerTagsTestJSON-19215696-project-member] Waiting for the task: (returnval){ [ 845.694764] env[61923]: value = "task-1377573" [ 845.694764] env[61923]: _type = "Task" [ 845.694764] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 845.707493] env[61923]: DEBUG oslo_vmware.api [None req-f39c3939-2621-40b9-927d-ae5b7aafe54d tempest-ServerTagsTestJSON-19215696 tempest-ServerTagsTestJSON-19215696-project-member] Task: {'id': task-1377573, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.710960] env[61923]: DEBUG oslo_vmware.api [None req-62f42696-8467-4135-b6f9-a4440e44fead tempest-ServerShowV257Test-986496257 tempest-ServerShowV257Test-986496257-project-member] Task: {'id': task-1377574, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.770079] env[61923]: DEBUG oslo_vmware.api [None req-950e07c2-502c-43fa-8598-da5f189b07e3 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Task: {'id': task-1377568, 'name': PowerOnVM_Task, 'duration_secs': 0.560151} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 845.771355] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-950e07c2-502c-43fa-8598-da5f189b07e3 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] [instance: 422d06d6-6932-46c3-bb25-841e6f65c028] Powered on the VM {{(pid=61923) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 845.771355] env[61923]: INFO nova.compute.manager [None req-950e07c2-502c-43fa-8598-da5f189b07e3 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] [instance: 422d06d6-6932-46c3-bb25-841e6f65c028] Took 8.22 seconds to spawn the instance on the hypervisor. [ 845.771355] env[61923]: DEBUG nova.compute.manager [None req-950e07c2-502c-43fa-8598-da5f189b07e3 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] [instance: 422d06d6-6932-46c3-bb25-841e6f65c028] Checking state {{(pid=61923) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 845.773051] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5da2705a-0645-4085-aa0f-7ee9f48712c9 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.776655] env[61923]: DEBUG oslo_concurrency.lockutils [None req-13baf863-98e6-41c7-ae75-48fd1a9cef32 tempest-ServerAddressesTestJSON-1846886542 tempest-ServerAddressesTestJSON-1846886542-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 845.911749] env[61923]: DEBUG nova.network.neutron [req-f275f677-8863-4edb-838f-1cc42b444f5a req-8700db0e-e32a-4910-9080-1e0ff2c97ad8 service nova] [instance: 59198f4d-4dde-4eaf-9f6c-a962cbe53c6e] Updated VIF entry in instance network info cache for port cf99020c-d7cc-4912-a021-8861628b7540. {{(pid=61923) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 845.911749] env[61923]: DEBUG nova.network.neutron [req-f275f677-8863-4edb-838f-1cc42b444f5a req-8700db0e-e32a-4910-9080-1e0ff2c97ad8 service nova] [instance: 59198f4d-4dde-4eaf-9f6c-a962cbe53c6e] Updating instance_info_cache with network_info: [{"id": "cf99020c-d7cc-4912-a021-8861628b7540", "address": "fa:16:3e:89:5c:35", "network": {"id": "b8c2e837-0d64-4b07-9cc5-a66c8e8db1e9", "bridge": "br-int", "label": "tempest-ImagesTestJSON-133215580-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7abfe95660904cd4b8824257795d8e8a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f2c424c9-6446-4b2a-af8c-4d9c29117c39", "external-id": "nsx-vlan-transportzone-437", "segmentation_id": 437, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcf99020c-d7", "ovs_interfaceid": "cf99020c-d7cc-4912-a021-8861628b7540", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 845.933068] env[61923]: DEBUG nova.network.neutron [None req-1214e428-bf00-4cbe-9e1e-7c8be1612c36 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] [instance: af3c317a-4007-4cea-a060-1e7dde5ce49e] Successfully created port: e5c62d99-3250-4fca-9798-25780ae50b00 {{(pid=61923) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 846.048516] env[61923]: DEBUG nova.compute.manager [None req-1214e428-bf00-4cbe-9e1e-7c8be1612c36 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] [instance: af3c317a-4007-4cea-a060-1e7dde5ce49e] Start building block device mappings for instance. {{(pid=61923) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 846.084905] env[61923]: DEBUG oslo_vmware.api [None req-e62df7da-42b5-4263-bcf8-001a2d06427a tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Task: {'id': task-1377569, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.766965} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 846.087017] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-e62df7da-42b5-4263-bcf8-001a2d06427a tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk to [datastore2] 73692517-1816-4e9b-ab2d-8265c683e83d/73692517-1816-4e9b-ab2d-8265c683e83d.vmdk {{(pid=61923) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 846.087017] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-e62df7da-42b5-4263-bcf8-001a2d06427a tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: 73692517-1816-4e9b-ab2d-8265c683e83d] Extending root virtual disk to 1048576 {{(pid=61923) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 846.087017] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-98e1e7fb-ceea-4695-8675-9f277b8af633 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.096101] env[61923]: DEBUG oslo_vmware.api [None req-e62df7da-42b5-4263-bcf8-001a2d06427a tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Waiting for the task: (returnval){ [ 846.096101] env[61923]: value = "task-1377575" [ 846.096101] env[61923]: _type = "Task" [ 846.096101] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 846.102233] env[61923]: DEBUG oslo_vmware.api [-] Task: {'id': task-1377572, 'name': CreateVM_Task} progress is 25%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.109932] env[61923]: DEBUG oslo_vmware.api [None req-e62df7da-42b5-4263-bcf8-001a2d06427a tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Task: {'id': task-1377575, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.216249] env[61923]: DEBUG oslo_vmware.api [None req-62f42696-8467-4135-b6f9-a4440e44fead tempest-ServerShowV257Test-986496257 tempest-ServerShowV257Test-986496257-project-member] Task: {'id': task-1377574, 'name': PowerOffVM_Task, 'duration_secs': 0.390982} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 846.216530] env[61923]: DEBUG oslo_vmware.api [None req-f39c3939-2621-40b9-927d-ae5b7aafe54d tempest-ServerTagsTestJSON-19215696 tempest-ServerTagsTestJSON-19215696-project-member] Task: {'id': task-1377573, 'name': PowerOffVM_Task, 'duration_secs': 0.36529} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 846.216718] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-62f42696-8467-4135-b6f9-a4440e44fead tempest-ServerShowV257Test-986496257 tempest-ServerShowV257Test-986496257-project-member] [instance: b8c60af7-3561-468c-a9a5-1c645080fc69] Powered off the VM {{(pid=61923) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 846.216888] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-62f42696-8467-4135-b6f9-a4440e44fead tempest-ServerShowV257Test-986496257 tempest-ServerShowV257Test-986496257-project-member] [instance: b8c60af7-3561-468c-a9a5-1c645080fc69] Unregistering the VM {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 846.217170] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-f39c3939-2621-40b9-927d-ae5b7aafe54d tempest-ServerTagsTestJSON-19215696 tempest-ServerTagsTestJSON-19215696-project-member] [instance: 48825592-dcdc-41d5-9fbf-500d1f31017b] Powered off the VM {{(pid=61923) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 846.217338] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-f39c3939-2621-40b9-927d-ae5b7aafe54d tempest-ServerTagsTestJSON-19215696 tempest-ServerTagsTestJSON-19215696-project-member] [instance: 48825592-dcdc-41d5-9fbf-500d1f31017b] Unregistering the VM {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 846.217568] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a9b8a228-8460-4855-bdb2-812b9cf88037 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.219183] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f8cdcc52-cc9c-47a0-be68-f04b5b775cde {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.253718] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-62f42696-8467-4135-b6f9-a4440e44fead tempest-ServerShowV257Test-986496257 tempest-ServerShowV257Test-986496257-project-member] [instance: b8c60af7-3561-468c-a9a5-1c645080fc69] Unregistered the VM {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 846.253718] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-62f42696-8467-4135-b6f9-a4440e44fead tempest-ServerShowV257Test-986496257 tempest-ServerShowV257Test-986496257-project-member] [instance: b8c60af7-3561-468c-a9a5-1c645080fc69] Deleting contents of the VM from datastore datastore2 {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 846.253718] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-62f42696-8467-4135-b6f9-a4440e44fead tempest-ServerShowV257Test-986496257 tempest-ServerShowV257Test-986496257-project-member] Deleting the datastore file [datastore2] b8c60af7-3561-468c-a9a5-1c645080fc69 {{(pid=61923) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 846.253952] env[61923]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-cdb2e10f-4185-4d8e-8fa5-5b04f3e8205e {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.260746] env[61923]: DEBUG oslo_vmware.api [None req-62f42696-8467-4135-b6f9-a4440e44fead tempest-ServerShowV257Test-986496257 tempest-ServerShowV257Test-986496257-project-member] Waiting for the task: (returnval){ [ 846.260746] env[61923]: value = "task-1377578" [ 846.260746] env[61923]: _type = "Task" [ 846.260746] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 846.271391] env[61923]: DEBUG oslo_vmware.api [None req-62f42696-8467-4135-b6f9-a4440e44fead tempest-ServerShowV257Test-986496257 tempest-ServerShowV257Test-986496257-project-member] Task: {'id': task-1377578, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.277289] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a818424-23e4-4116-980d-1fe1adc50bc4 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.283676] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf2f202f-b204-4979-96ed-c1090a0e74d3 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.321226] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6b18881-076f-458b-9d44-5a088a6c3207 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.323725] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-f39c3939-2621-40b9-927d-ae5b7aafe54d tempest-ServerTagsTestJSON-19215696 tempest-ServerTagsTestJSON-19215696-project-member] [instance: 48825592-dcdc-41d5-9fbf-500d1f31017b] Unregistered the VM {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 846.323920] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-f39c3939-2621-40b9-927d-ae5b7aafe54d tempest-ServerTagsTestJSON-19215696 tempest-ServerTagsTestJSON-19215696-project-member] [instance: 48825592-dcdc-41d5-9fbf-500d1f31017b] Deleting contents of the VM from datastore datastore1 {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 846.324100] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-f39c3939-2621-40b9-927d-ae5b7aafe54d tempest-ServerTagsTestJSON-19215696 tempest-ServerTagsTestJSON-19215696-project-member] Deleting the datastore file [datastore1] 48825592-dcdc-41d5-9fbf-500d1f31017b {{(pid=61923) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 846.324564] env[61923]: INFO nova.compute.manager [None req-950e07c2-502c-43fa-8598-da5f189b07e3 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] [instance: 422d06d6-6932-46c3-bb25-841e6f65c028] Took 26.91 seconds to build instance. [ 846.325519] env[61923]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-db8ed75e-c8a0-4c79-aeac-776dfd5e546a {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.333405] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fef5f6f2-583a-4287-9ff8-d82b62376bd0 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.336941] env[61923]: DEBUG oslo_vmware.api [None req-f39c3939-2621-40b9-927d-ae5b7aafe54d tempest-ServerTagsTestJSON-19215696 tempest-ServerTagsTestJSON-19215696-project-member] Waiting for the task: (returnval){ [ 846.336941] env[61923]: value = "task-1377579" [ 846.336941] env[61923]: _type = "Task" [ 846.336941] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 846.348030] env[61923]: DEBUG nova.compute.provider_tree [None req-52eadf86-cc45-4cae-af2a-645ac3cb8658 tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 846.354213] env[61923]: DEBUG oslo_vmware.api [None req-f39c3939-2621-40b9-927d-ae5b7aafe54d tempest-ServerTagsTestJSON-19215696 tempest-ServerTagsTestJSON-19215696-project-member] Task: {'id': task-1377579, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.413722] env[61923]: DEBUG oslo_concurrency.lockutils [req-f275f677-8863-4edb-838f-1cc42b444f5a req-8700db0e-e32a-4910-9080-1e0ff2c97ad8 service nova] Releasing lock "refresh_cache-59198f4d-4dde-4eaf-9f6c-a962cbe53c6e" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 846.603240] env[61923]: DEBUG oslo_vmware.api [-] Task: {'id': task-1377572, 'name': CreateVM_Task, 'duration_secs': 0.727542} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 846.607234] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 59198f4d-4dde-4eaf-9f6c-a962cbe53c6e] Created VM on the ESX host {{(pid=61923) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 846.608730] env[61923]: DEBUG oslo_concurrency.lockutils [None req-513daa64-2b05-4c0d-94e9-55b95de63eec tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 846.609073] env[61923]: DEBUG oslo_concurrency.lockutils [None req-513daa64-2b05-4c0d-94e9-55b95de63eec tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 846.609475] env[61923]: DEBUG oslo_concurrency.lockutils [None req-513daa64-2b05-4c0d-94e9-55b95de63eec tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 846.610432] env[61923]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8dd4499e-23e9-4cd5-92c5-c933d1b7fa5a {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.615635] env[61923]: DEBUG oslo_vmware.api [None req-e62df7da-42b5-4263-bcf8-001a2d06427a tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Task: {'id': task-1377575, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.18247} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 846.616305] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-e62df7da-42b5-4263-bcf8-001a2d06427a tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: 73692517-1816-4e9b-ab2d-8265c683e83d] Extended root virtual disk {{(pid=61923) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 846.617288] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9bc8f03-d957-4ad6-bb6c-a8b2ac28dfe0 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.621381] env[61923]: DEBUG oslo_vmware.api [None req-513daa64-2b05-4c0d-94e9-55b95de63eec tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Waiting for the task: (returnval){ [ 846.621381] env[61923]: value = "session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52a86ae8-d027-6d58-b560-87b4da46ba5b" [ 846.621381] env[61923]: _type = "Task" [ 846.621381] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 846.642394] env[61923]: DEBUG nova.virt.vmwareapi.volumeops [None req-e62df7da-42b5-4263-bcf8-001a2d06427a tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: 73692517-1816-4e9b-ab2d-8265c683e83d] Reconfiguring VM instance instance-00000044 to attach disk [datastore2] 73692517-1816-4e9b-ab2d-8265c683e83d/73692517-1816-4e9b-ab2d-8265c683e83d.vmdk or device None with type sparse {{(pid=61923) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 846.643427] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-cad10c9f-e4eb-4f86-86b3-0284a0dcc169 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.661067] env[61923]: DEBUG oslo_vmware.api [None req-513daa64-2b05-4c0d-94e9-55b95de63eec tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52a86ae8-d027-6d58-b560-87b4da46ba5b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.669321] env[61923]: DEBUG oslo_vmware.api [None req-e62df7da-42b5-4263-bcf8-001a2d06427a tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Waiting for the task: (returnval){ [ 846.669321] env[61923]: value = "task-1377580" [ 846.669321] env[61923]: _type = "Task" [ 846.669321] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 846.677534] env[61923]: DEBUG oslo_vmware.api [None req-e62df7da-42b5-4263-bcf8-001a2d06427a tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Task: {'id': task-1377580, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.716165] env[61923]: DEBUG nova.compute.manager [req-6c5205d2-fbba-4e46-bc5f-0fef60a3693e req-64afe44a-5f89-4f81-982a-3e3e81406590 service nova] [instance: fc4fdf89-4615-4586-9dcd-ea90dc258361] Received event network-vif-plugged-8781f580-cc8d-4677-b6b3-2e71a0b907bc {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 846.716165] env[61923]: DEBUG oslo_concurrency.lockutils [req-6c5205d2-fbba-4e46-bc5f-0fef60a3693e req-64afe44a-5f89-4f81-982a-3e3e81406590 service nova] Acquiring lock "fc4fdf89-4615-4586-9dcd-ea90dc258361-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 846.716165] env[61923]: DEBUG oslo_concurrency.lockutils [req-6c5205d2-fbba-4e46-bc5f-0fef60a3693e req-64afe44a-5f89-4f81-982a-3e3e81406590 service nova] Lock "fc4fdf89-4615-4586-9dcd-ea90dc258361-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 846.716307] env[61923]: DEBUG oslo_concurrency.lockutils [req-6c5205d2-fbba-4e46-bc5f-0fef60a3693e req-64afe44a-5f89-4f81-982a-3e3e81406590 service nova] Lock "fc4fdf89-4615-4586-9dcd-ea90dc258361-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 846.716610] env[61923]: DEBUG nova.compute.manager [req-6c5205d2-fbba-4e46-bc5f-0fef60a3693e req-64afe44a-5f89-4f81-982a-3e3e81406590 service nova] [instance: fc4fdf89-4615-4586-9dcd-ea90dc258361] No waiting events found dispatching network-vif-plugged-8781f580-cc8d-4677-b6b3-2e71a0b907bc {{(pid=61923) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 846.716914] env[61923]: WARNING nova.compute.manager [req-6c5205d2-fbba-4e46-bc5f-0fef60a3693e req-64afe44a-5f89-4f81-982a-3e3e81406590 service nova] [instance: fc4fdf89-4615-4586-9dcd-ea90dc258361] Received unexpected event network-vif-plugged-8781f580-cc8d-4677-b6b3-2e71a0b907bc for instance with vm_state building and task_state spawning. [ 846.775758] env[61923]: DEBUG oslo_vmware.api [None req-62f42696-8467-4135-b6f9-a4440e44fead tempest-ServerShowV257Test-986496257 tempest-ServerShowV257Test-986496257-project-member] Task: {'id': task-1377578, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.438753} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 846.776295] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-62f42696-8467-4135-b6f9-a4440e44fead tempest-ServerShowV257Test-986496257 tempest-ServerShowV257Test-986496257-project-member] Deleted the datastore file {{(pid=61923) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 846.776625] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-62f42696-8467-4135-b6f9-a4440e44fead tempest-ServerShowV257Test-986496257 tempest-ServerShowV257Test-986496257-project-member] [instance: b8c60af7-3561-468c-a9a5-1c645080fc69] Deleted contents of the VM from datastore datastore2 {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 846.776922] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-62f42696-8467-4135-b6f9-a4440e44fead tempest-ServerShowV257Test-986496257 tempest-ServerShowV257Test-986496257-project-member] [instance: b8c60af7-3561-468c-a9a5-1c645080fc69] Instance destroyed {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 846.777219] env[61923]: INFO nova.compute.manager [None req-62f42696-8467-4135-b6f9-a4440e44fead tempest-ServerShowV257Test-986496257 tempest-ServerShowV257Test-986496257-project-member] [instance: b8c60af7-3561-468c-a9a5-1c645080fc69] Took 1.11 seconds to destroy the instance on the hypervisor. [ 846.777565] env[61923]: DEBUG oslo.service.loopingcall [None req-62f42696-8467-4135-b6f9-a4440e44fead tempest-ServerShowV257Test-986496257 tempest-ServerShowV257Test-986496257-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61923) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 846.777862] env[61923]: DEBUG nova.compute.manager [-] [instance: b8c60af7-3561-468c-a9a5-1c645080fc69] Deallocating network for instance {{(pid=61923) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 846.778083] env[61923]: DEBUG nova.network.neutron [-] [instance: b8c60af7-3561-468c-a9a5-1c645080fc69] deallocate_for_instance() {{(pid=61923) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 846.810638] env[61923]: DEBUG nova.network.neutron [-] [instance: b8c60af7-3561-468c-a9a5-1c645080fc69] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 846.828739] env[61923]: DEBUG oslo_concurrency.lockutils [None req-950e07c2-502c-43fa-8598-da5f189b07e3 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Lock "422d06d6-6932-46c3-bb25-841e6f65c028" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 85.478s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 846.847641] env[61923]: DEBUG oslo_vmware.api [None req-f39c3939-2621-40b9-927d-ae5b7aafe54d tempest-ServerTagsTestJSON-19215696 tempest-ServerTagsTestJSON-19215696-project-member] Task: {'id': task-1377579, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.854022] env[61923]: DEBUG nova.scheduler.client.report [None req-52eadf86-cc45-4cae-af2a-645ac3cb8658 tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 846.936912] env[61923]: DEBUG nova.network.neutron [None req-80740a2c-9779-4afc-be30-c4469cd3724d tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] [instance: fc4fdf89-4615-4586-9dcd-ea90dc258361] Successfully updated port: 8781f580-cc8d-4677-b6b3-2e71a0b907bc {{(pid=61923) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 847.065072] env[61923]: DEBUG nova.compute.manager [None req-1214e428-bf00-4cbe-9e1e-7c8be1612c36 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] [instance: af3c317a-4007-4cea-a060-1e7dde5ce49e] Start spawning the instance on the hypervisor. {{(pid=61923) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 847.091103] env[61923]: DEBUG nova.virt.hardware [None req-1214e428-bf00-4cbe-9e1e-7c8be1612c36 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-29T20:07:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-29T20:07:35Z,direct_url=,disk_format='vmdk',id=0f153f63-ae0a-45b1-b7f5-7f9b673c947f,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='4e7b5e3b3c3443c8bd5c70f8a15739f5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-29T20:07:36Z,virtual_size=,visibility=), allow threads: False {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 847.091103] env[61923]: DEBUG nova.virt.hardware [None req-1214e428-bf00-4cbe-9e1e-7c8be1612c36 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] Flavor limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 847.091103] env[61923]: DEBUG nova.virt.hardware [None req-1214e428-bf00-4cbe-9e1e-7c8be1612c36 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] Image limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 847.091251] env[61923]: DEBUG nova.virt.hardware [None req-1214e428-bf00-4cbe-9e1e-7c8be1612c36 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] Flavor pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 847.091251] env[61923]: DEBUG nova.virt.hardware [None req-1214e428-bf00-4cbe-9e1e-7c8be1612c36 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] Image pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 847.091251] env[61923]: DEBUG nova.virt.hardware [None req-1214e428-bf00-4cbe-9e1e-7c8be1612c36 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 847.091251] env[61923]: DEBUG nova.virt.hardware [None req-1214e428-bf00-4cbe-9e1e-7c8be1612c36 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 847.091251] env[61923]: DEBUG nova.virt.hardware [None req-1214e428-bf00-4cbe-9e1e-7c8be1612c36 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 847.091391] env[61923]: DEBUG nova.virt.hardware [None req-1214e428-bf00-4cbe-9e1e-7c8be1612c36 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] Got 1 possible topologies {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 847.091828] env[61923]: DEBUG nova.virt.hardware [None req-1214e428-bf00-4cbe-9e1e-7c8be1612c36 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 847.092223] env[61923]: DEBUG nova.virt.hardware [None req-1214e428-bf00-4cbe-9e1e-7c8be1612c36 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 847.093225] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af85be3e-f048-4777-af3a-6f3703f13557 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.101181] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea444916-2896-4068-a9cd-6af0e10fc24f {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.133019] env[61923]: DEBUG oslo_vmware.api [None req-513daa64-2b05-4c0d-94e9-55b95de63eec tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52a86ae8-d027-6d58-b560-87b4da46ba5b, 'name': SearchDatastore_Task, 'duration_secs': 0.039073} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 847.133019] env[61923]: DEBUG oslo_concurrency.lockutils [None req-513daa64-2b05-4c0d-94e9-55b95de63eec tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 847.133019] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-513daa64-2b05-4c0d-94e9-55b95de63eec tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] [instance: 59198f4d-4dde-4eaf-9f6c-a962cbe53c6e] Processing image 0f153f63-ae0a-45b1-b7f5-7f9b673c947f {{(pid=61923) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 847.133019] env[61923]: DEBUG oslo_concurrency.lockutils [None req-513daa64-2b05-4c0d-94e9-55b95de63eec tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 847.133196] env[61923]: DEBUG oslo_concurrency.lockutils [None req-513daa64-2b05-4c0d-94e9-55b95de63eec tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 847.133196] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-513daa64-2b05-4c0d-94e9-55b95de63eec tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61923) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 847.133196] env[61923]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-397575d9-1fc3-4b0b-8a03-86dfcf7a2940 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.139619] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-513daa64-2b05-4c0d-94e9-55b95de63eec tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61923) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 847.139977] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-513daa64-2b05-4c0d-94e9-55b95de63eec tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61923) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 847.140799] env[61923]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5f4bbffb-fe80-42e1-a880-b83f6c966e4a {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.145699] env[61923]: DEBUG oslo_vmware.api [None req-513daa64-2b05-4c0d-94e9-55b95de63eec tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Waiting for the task: (returnval){ [ 847.145699] env[61923]: value = "session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52b3e012-5e88-f961-a196-ea16afa455fe" [ 847.145699] env[61923]: _type = "Task" [ 847.145699] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 847.152915] env[61923]: DEBUG oslo_vmware.api [None req-513daa64-2b05-4c0d-94e9-55b95de63eec tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52b3e012-5e88-f961-a196-ea16afa455fe, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.176498] env[61923]: DEBUG oslo_vmware.api [None req-e62df7da-42b5-4263-bcf8-001a2d06427a tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Task: {'id': task-1377580, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.313444] env[61923]: DEBUG nova.network.neutron [-] [instance: b8c60af7-3561-468c-a9a5-1c645080fc69] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 847.332201] env[61923]: DEBUG nova.compute.manager [None req-00703b42-f309-4672-8dc7-10f4a69426aa tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] [instance: b779d183-89ae-4e4d-ae99-e514e145ed43] Starting instance... {{(pid=61923) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 847.348699] env[61923]: DEBUG oslo_vmware.api [None req-f39c3939-2621-40b9-927d-ae5b7aafe54d tempest-ServerTagsTestJSON-19215696 tempest-ServerTagsTestJSON-19215696-project-member] Task: {'id': task-1377579, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.546748} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 847.348699] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-f39c3939-2621-40b9-927d-ae5b7aafe54d tempest-ServerTagsTestJSON-19215696 tempest-ServerTagsTestJSON-19215696-project-member] Deleted the datastore file {{(pid=61923) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 847.348699] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-f39c3939-2621-40b9-927d-ae5b7aafe54d tempest-ServerTagsTestJSON-19215696 tempest-ServerTagsTestJSON-19215696-project-member] [instance: 48825592-dcdc-41d5-9fbf-500d1f31017b] Deleted contents of the VM from datastore datastore1 {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 847.348699] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-f39c3939-2621-40b9-927d-ae5b7aafe54d tempest-ServerTagsTestJSON-19215696 tempest-ServerTagsTestJSON-19215696-project-member] [instance: 48825592-dcdc-41d5-9fbf-500d1f31017b] Instance destroyed {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 847.348699] env[61923]: INFO nova.compute.manager [None req-f39c3939-2621-40b9-927d-ae5b7aafe54d tempest-ServerTagsTestJSON-19215696 tempest-ServerTagsTestJSON-19215696-project-member] [instance: 48825592-dcdc-41d5-9fbf-500d1f31017b] Took 1.68 seconds to destroy the instance on the hypervisor. [ 847.349203] env[61923]: DEBUG oslo.service.loopingcall [None req-f39c3939-2621-40b9-927d-ae5b7aafe54d tempest-ServerTagsTestJSON-19215696 tempest-ServerTagsTestJSON-19215696-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61923) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 847.349505] env[61923]: DEBUG nova.compute.manager [-] [instance: 48825592-dcdc-41d5-9fbf-500d1f31017b] Deallocating network for instance {{(pid=61923) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 847.349863] env[61923]: DEBUG nova.network.neutron [-] [instance: 48825592-dcdc-41d5-9fbf-500d1f31017b] deallocate_for_instance() {{(pid=61923) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 847.356161] env[61923]: DEBUG oslo_concurrency.lockutils [None req-52eadf86-cc45-4cae-af2a-645ac3cb8658 tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.321s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 847.357569] env[61923]: DEBUG nova.compute.manager [None req-52eadf86-cc45-4cae-af2a-645ac3cb8658 tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] [instance: 87ae37dc-d0d0-4b76-afdd-0ba3e8f6ce0b] Start building networks asynchronously for instance. {{(pid=61923) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 847.359216] env[61923]: DEBUG oslo_concurrency.lockutils [None req-5fbf0568-2a85-4e3c-9b63-7d91ecc77593 tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 11.793s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 847.360650] env[61923]: INFO nova.compute.claims [None req-5fbf0568-2a85-4e3c-9b63-7d91ecc77593 tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] [instance: 2719569b-8572-4199-8158-7bb367d17dc5] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 847.373181] env[61923]: DEBUG nova.compute.manager [req-25d38eb0-626b-4af2-9c64-c97c7f476eef req-e519a948-7fb1-46b5-8d26-70aa4b55f26d service nova] [instance: 422d06d6-6932-46c3-bb25-841e6f65c028] Received event network-changed-97caa1b8-427c-46b0-be7a-d59b932900c3 {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 847.373457] env[61923]: DEBUG nova.compute.manager [req-25d38eb0-626b-4af2-9c64-c97c7f476eef req-e519a948-7fb1-46b5-8d26-70aa4b55f26d service nova] [instance: 422d06d6-6932-46c3-bb25-841e6f65c028] Refreshing instance network info cache due to event network-changed-97caa1b8-427c-46b0-be7a-d59b932900c3. {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 847.373699] env[61923]: DEBUG oslo_concurrency.lockutils [req-25d38eb0-626b-4af2-9c64-c97c7f476eef req-e519a948-7fb1-46b5-8d26-70aa4b55f26d service nova] Acquiring lock "refresh_cache-422d06d6-6932-46c3-bb25-841e6f65c028" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 847.373920] env[61923]: DEBUG oslo_concurrency.lockutils [req-25d38eb0-626b-4af2-9c64-c97c7f476eef req-e519a948-7fb1-46b5-8d26-70aa4b55f26d service nova] Acquired lock "refresh_cache-422d06d6-6932-46c3-bb25-841e6f65c028" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 847.374168] env[61923]: DEBUG nova.network.neutron [req-25d38eb0-626b-4af2-9c64-c97c7f476eef req-e519a948-7fb1-46b5-8d26-70aa4b55f26d service nova] [instance: 422d06d6-6932-46c3-bb25-841e6f65c028] Refreshing network info cache for port 97caa1b8-427c-46b0-be7a-d59b932900c3 {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 847.410892] env[61923]: DEBUG nova.compute.manager [req-6e8a1a1a-d04f-4096-9ae2-fdac9fd31e14 req-5dcb842b-c6a0-4075-b8dc-7d4a73159095 service nova] [instance: af3c317a-4007-4cea-a060-1e7dde5ce49e] Received event network-vif-plugged-e5c62d99-3250-4fca-9798-25780ae50b00 {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 847.411331] env[61923]: DEBUG oslo_concurrency.lockutils [req-6e8a1a1a-d04f-4096-9ae2-fdac9fd31e14 req-5dcb842b-c6a0-4075-b8dc-7d4a73159095 service nova] Acquiring lock "af3c317a-4007-4cea-a060-1e7dde5ce49e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 847.411628] env[61923]: DEBUG oslo_concurrency.lockutils [req-6e8a1a1a-d04f-4096-9ae2-fdac9fd31e14 req-5dcb842b-c6a0-4075-b8dc-7d4a73159095 service nova] Lock "af3c317a-4007-4cea-a060-1e7dde5ce49e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 847.414019] env[61923]: DEBUG oslo_concurrency.lockutils [req-6e8a1a1a-d04f-4096-9ae2-fdac9fd31e14 req-5dcb842b-c6a0-4075-b8dc-7d4a73159095 service nova] Lock "af3c317a-4007-4cea-a060-1e7dde5ce49e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 847.414019] env[61923]: DEBUG nova.compute.manager [req-6e8a1a1a-d04f-4096-9ae2-fdac9fd31e14 req-5dcb842b-c6a0-4075-b8dc-7d4a73159095 service nova] [instance: af3c317a-4007-4cea-a060-1e7dde5ce49e] No waiting events found dispatching network-vif-plugged-e5c62d99-3250-4fca-9798-25780ae50b00 {{(pid=61923) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 847.414019] env[61923]: WARNING nova.compute.manager [req-6e8a1a1a-d04f-4096-9ae2-fdac9fd31e14 req-5dcb842b-c6a0-4075-b8dc-7d4a73159095 service nova] [instance: af3c317a-4007-4cea-a060-1e7dde5ce49e] Received unexpected event network-vif-plugged-e5c62d99-3250-4fca-9798-25780ae50b00 for instance with vm_state building and task_state spawning. [ 847.658650] env[61923]: DEBUG oslo_vmware.api [None req-513daa64-2b05-4c0d-94e9-55b95de63eec tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52b3e012-5e88-f961-a196-ea16afa455fe, 'name': SearchDatastore_Task, 'duration_secs': 0.009865} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 847.658650] env[61923]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cd7100bc-6e16-41ef-8c70-5937cdf2cba0 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.664018] env[61923]: DEBUG oslo_vmware.api [None req-513daa64-2b05-4c0d-94e9-55b95de63eec tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Waiting for the task: (returnval){ [ 847.664018] env[61923]: value = "session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]5226cd11-54bc-a7eb-12f9-9b76993072a2" [ 847.664018] env[61923]: _type = "Task" [ 847.664018] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 847.673448] env[61923]: DEBUG oslo_vmware.api [None req-513daa64-2b05-4c0d-94e9-55b95de63eec tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]5226cd11-54bc-a7eb-12f9-9b76993072a2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.681473] env[61923]: DEBUG oslo_vmware.api [None req-e62df7da-42b5-4263-bcf8-001a2d06427a tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Task: {'id': task-1377580, 'name': ReconfigVM_Task, 'duration_secs': 0.772788} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 847.681838] env[61923]: DEBUG nova.virt.vmwareapi.volumeops [None req-e62df7da-42b5-4263-bcf8-001a2d06427a tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: 73692517-1816-4e9b-ab2d-8265c683e83d] Reconfigured VM instance instance-00000044 to attach disk [datastore2] 73692517-1816-4e9b-ab2d-8265c683e83d/73692517-1816-4e9b-ab2d-8265c683e83d.vmdk or device None with type sparse {{(pid=61923) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 847.682493] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f890297e-6dd9-4f68-9540-1656ce9f3c0f {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.689329] env[61923]: DEBUG oslo_vmware.api [None req-e62df7da-42b5-4263-bcf8-001a2d06427a tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Waiting for the task: (returnval){ [ 847.689329] env[61923]: value = "task-1377581" [ 847.689329] env[61923]: _type = "Task" [ 847.689329] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 847.697761] env[61923]: DEBUG oslo_vmware.api [None req-e62df7da-42b5-4263-bcf8-001a2d06427a tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Task: {'id': task-1377581, 'name': Rename_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.816767] env[61923]: INFO nova.compute.manager [-] [instance: b8c60af7-3561-468c-a9a5-1c645080fc69] Took 1.04 seconds to deallocate network for instance. [ 847.863542] env[61923]: DEBUG oslo_concurrency.lockutils [None req-00703b42-f309-4672-8dc7-10f4a69426aa tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 847.866391] env[61923]: DEBUG nova.compute.utils [None req-52eadf86-cc45-4cae-af2a-645ac3cb8658 tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Using /dev/sd instead of None {{(pid=61923) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 847.867402] env[61923]: DEBUG nova.compute.manager [None req-52eadf86-cc45-4cae-af2a-645ac3cb8658 tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] [instance: 87ae37dc-d0d0-4b76-afdd-0ba3e8f6ce0b] Allocating IP information in the background. {{(pid=61923) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 847.867655] env[61923]: DEBUG nova.network.neutron [None req-52eadf86-cc45-4cae-af2a-645ac3cb8658 tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] [instance: 87ae37dc-d0d0-4b76-afdd-0ba3e8f6ce0b] allocate_for_instance() {{(pid=61923) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 848.028755] env[61923]: DEBUG nova.network.neutron [None req-1214e428-bf00-4cbe-9e1e-7c8be1612c36 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] [instance: af3c317a-4007-4cea-a060-1e7dde5ce49e] Successfully updated port: e5c62d99-3250-4fca-9798-25780ae50b00 {{(pid=61923) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 848.062768] env[61923]: DEBUG nova.policy [None req-52eadf86-cc45-4cae-af2a-645ac3cb8658 tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6e7ae11a571e4252ba52dae60db91b34', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd946a91d4cf9412a838fbcb74cb5da98', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61923) authorize /opt/stack/nova/nova/policy.py:201}} [ 848.174027] env[61923]: DEBUG oslo_vmware.api [None req-513daa64-2b05-4c0d-94e9-55b95de63eec tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]5226cd11-54bc-a7eb-12f9-9b76993072a2, 'name': SearchDatastore_Task, 'duration_secs': 0.011782} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 848.174645] env[61923]: DEBUG oslo_concurrency.lockutils [None req-513daa64-2b05-4c0d-94e9-55b95de63eec tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 848.174914] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-513daa64-2b05-4c0d-94e9-55b95de63eec tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk to [datastore2] 59198f4d-4dde-4eaf-9f6c-a962cbe53c6e/59198f4d-4dde-4eaf-9f6c-a962cbe53c6e.vmdk {{(pid=61923) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 848.175207] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-5c7eb3de-9acb-4dc7-ba8f-fa436acc8bb8 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.181832] env[61923]: DEBUG oslo_vmware.api [None req-513daa64-2b05-4c0d-94e9-55b95de63eec tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Waiting for the task: (returnval){ [ 848.181832] env[61923]: value = "task-1377582" [ 848.181832] env[61923]: _type = "Task" [ 848.181832] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 848.189502] env[61923]: DEBUG oslo_vmware.api [None req-513daa64-2b05-4c0d-94e9-55b95de63eec tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Task: {'id': task-1377582, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.197616] env[61923]: DEBUG oslo_vmware.api [None req-e62df7da-42b5-4263-bcf8-001a2d06427a tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Task: {'id': task-1377581, 'name': Rename_Task, 'duration_secs': 0.147158} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 848.197861] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-e62df7da-42b5-4263-bcf8-001a2d06427a tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: 73692517-1816-4e9b-ab2d-8265c683e83d] Powering on the VM {{(pid=61923) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 848.198110] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d27fc46e-f6e7-44dc-ab38-827bcc698eb8 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.203174] env[61923]: DEBUG oslo_vmware.api [None req-e62df7da-42b5-4263-bcf8-001a2d06427a tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Waiting for the task: (returnval){ [ 848.203174] env[61923]: value = "task-1377583" [ 848.203174] env[61923]: _type = "Task" [ 848.203174] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 848.210528] env[61923]: DEBUG oslo_vmware.api [None req-e62df7da-42b5-4263-bcf8-001a2d06427a tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Task: {'id': task-1377583, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.326112] env[61923]: DEBUG oslo_concurrency.lockutils [None req-62f42696-8467-4135-b6f9-a4440e44fead tempest-ServerShowV257Test-986496257 tempest-ServerShowV257Test-986496257-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 848.345691] env[61923]: DEBUG nova.network.neutron [-] [instance: 48825592-dcdc-41d5-9fbf-500d1f31017b] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 848.377396] env[61923]: DEBUG nova.compute.manager [None req-52eadf86-cc45-4cae-af2a-645ac3cb8658 tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] [instance: 87ae37dc-d0d0-4b76-afdd-0ba3e8f6ce0b] Start building block device mappings for instance. {{(pid=61923) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 848.536587] env[61923]: DEBUG oslo_concurrency.lockutils [None req-1214e428-bf00-4cbe-9e1e-7c8be1612c36 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] Acquiring lock "refresh_cache-af3c317a-4007-4cea-a060-1e7dde5ce49e" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 848.536587] env[61923]: DEBUG oslo_concurrency.lockutils [None req-1214e428-bf00-4cbe-9e1e-7c8be1612c36 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] Acquired lock "refresh_cache-af3c317a-4007-4cea-a060-1e7dde5ce49e" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 848.536587] env[61923]: DEBUG nova.network.neutron [None req-1214e428-bf00-4cbe-9e1e-7c8be1612c36 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] [instance: af3c317a-4007-4cea-a060-1e7dde5ce49e] Building network info cache for instance {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 848.629486] env[61923]: DEBUG oslo_concurrency.lockutils [None req-9887b108-13fe-4987-ad6f-1ef139e2a15d tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Acquiring lock "6012d3b5-739c-4762-9bb4-09c51171dcd7" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 848.630152] env[61923]: DEBUG oslo_concurrency.lockutils [None req-9887b108-13fe-4987-ad6f-1ef139e2a15d tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Lock "6012d3b5-739c-4762-9bb4-09c51171dcd7" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 848.692323] env[61923]: DEBUG oslo_vmware.api [None req-513daa64-2b05-4c0d-94e9-55b95de63eec tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Task: {'id': task-1377582, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.694457] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66d31f51-ad93-47f4-82fe-797009a15281 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.702888] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa2d926b-dd7f-40c8-bdc9-98236419eaad {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.746087] env[61923]: DEBUG oslo_vmware.api [None req-e62df7da-42b5-4263-bcf8-001a2d06427a tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Task: {'id': task-1377583, 'name': PowerOnVM_Task} progress is 78%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.749614] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b162a16-a23c-4dfe-af03-a0d8b761014b {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.759021] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a1f709f-e68b-481e-9ea6-86bce5c9e7fe {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.774262] env[61923]: DEBUG nova.compute.provider_tree [None req-5fbf0568-2a85-4e3c-9b63-7d91ecc77593 tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 848.848156] env[61923]: INFO nova.compute.manager [-] [instance: 48825592-dcdc-41d5-9fbf-500d1f31017b] Took 1.50 seconds to deallocate network for instance. [ 848.867894] env[61923]: DEBUG nova.network.neutron [req-25d38eb0-626b-4af2-9c64-c97c7f476eef req-e519a948-7fb1-46b5-8d26-70aa4b55f26d service nova] [instance: 422d06d6-6932-46c3-bb25-841e6f65c028] Updated VIF entry in instance network info cache for port 97caa1b8-427c-46b0-be7a-d59b932900c3. {{(pid=61923) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 848.868227] env[61923]: DEBUG nova.network.neutron [req-25d38eb0-626b-4af2-9c64-c97c7f476eef req-e519a948-7fb1-46b5-8d26-70aa4b55f26d service nova] [instance: 422d06d6-6932-46c3-bb25-841e6f65c028] Updating instance_info_cache with network_info: [{"id": "97caa1b8-427c-46b0-be7a-d59b932900c3", "address": "fa:16:3e:fa:f8:d1", "network": {"id": "a8b7e1b4-1ec6-4226-bdd6-535dbce450d2", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1928758125-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.146", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "912300a09b80452e85e3cd13c4a644cc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7806fe18-2b89-4386-87b1-f22876f82af2", "external-id": "nsx-vlan-transportzone-727", "segmentation_id": 727, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap97caa1b8-42", "ovs_interfaceid": "97caa1b8-427c-46b0-be7a-d59b932900c3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 848.906256] env[61923]: DEBUG nova.compute.manager [req-a2932dc2-1708-4570-88c9-5ea7176fcbeb req-ec160d3c-e143-45c2-81bc-0454045f558b service nova] [instance: fc4fdf89-4615-4586-9dcd-ea90dc258361] Received event network-changed-8781f580-cc8d-4677-b6b3-2e71a0b907bc {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 848.906469] env[61923]: DEBUG nova.compute.manager [req-a2932dc2-1708-4570-88c9-5ea7176fcbeb req-ec160d3c-e143-45c2-81bc-0454045f558b service nova] [instance: fc4fdf89-4615-4586-9dcd-ea90dc258361] Refreshing instance network info cache due to event network-changed-8781f580-cc8d-4677-b6b3-2e71a0b907bc. {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 848.906681] env[61923]: DEBUG oslo_concurrency.lockutils [req-a2932dc2-1708-4570-88c9-5ea7176fcbeb req-ec160d3c-e143-45c2-81bc-0454045f558b service nova] Acquiring lock "refresh_cache-fc4fdf89-4615-4586-9dcd-ea90dc258361" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 848.906865] env[61923]: DEBUG oslo_concurrency.lockutils [req-a2932dc2-1708-4570-88c9-5ea7176fcbeb req-ec160d3c-e143-45c2-81bc-0454045f558b service nova] Acquired lock "refresh_cache-fc4fdf89-4615-4586-9dcd-ea90dc258361" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 848.907075] env[61923]: DEBUG nova.network.neutron [req-a2932dc2-1708-4570-88c9-5ea7176fcbeb req-ec160d3c-e143-45c2-81bc-0454045f558b service nova] [instance: fc4fdf89-4615-4586-9dcd-ea90dc258361] Refreshing network info cache for port 8781f580-cc8d-4677-b6b3-2e71a0b907bc {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 849.042621] env[61923]: DEBUG nova.network.neutron [None req-52eadf86-cc45-4cae-af2a-645ac3cb8658 tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] [instance: 87ae37dc-d0d0-4b76-afdd-0ba3e8f6ce0b] Successfully created port: b0e2ee14-b051-47bc-9064-50f09916c0cb {{(pid=61923) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 849.081754] env[61923]: DEBUG nova.network.neutron [None req-1214e428-bf00-4cbe-9e1e-7c8be1612c36 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] [instance: af3c317a-4007-4cea-a060-1e7dde5ce49e] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 849.192625] env[61923]: DEBUG oslo_vmware.api [None req-513daa64-2b05-4c0d-94e9-55b95de63eec tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Task: {'id': task-1377582, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.575417} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 849.195141] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-513daa64-2b05-4c0d-94e9-55b95de63eec tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk to [datastore2] 59198f4d-4dde-4eaf-9f6c-a962cbe53c6e/59198f4d-4dde-4eaf-9f6c-a962cbe53c6e.vmdk {{(pid=61923) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 849.195366] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-513daa64-2b05-4c0d-94e9-55b95de63eec tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] [instance: 59198f4d-4dde-4eaf-9f6c-a962cbe53c6e] Extending root virtual disk to 1048576 {{(pid=61923) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 849.195724] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-012a60dc-afc1-42f4-8f9a-c2db45f17186 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.202786] env[61923]: DEBUG oslo_vmware.api [None req-513daa64-2b05-4c0d-94e9-55b95de63eec tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Waiting for the task: (returnval){ [ 849.202786] env[61923]: value = "task-1377584" [ 849.202786] env[61923]: _type = "Task" [ 849.202786] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 849.214312] env[61923]: DEBUG oslo_vmware.api [None req-513daa64-2b05-4c0d-94e9-55b95de63eec tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Task: {'id': task-1377584, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.217486] env[61923]: DEBUG oslo_vmware.api [None req-e62df7da-42b5-4263-bcf8-001a2d06427a tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Task: {'id': task-1377583, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.277743] env[61923]: DEBUG nova.scheduler.client.report [None req-5fbf0568-2a85-4e3c-9b63-7d91ecc77593 tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 849.323777] env[61923]: DEBUG nova.network.neutron [None req-1214e428-bf00-4cbe-9e1e-7c8be1612c36 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] [instance: af3c317a-4007-4cea-a060-1e7dde5ce49e] Updating instance_info_cache with network_info: [{"id": "e5c62d99-3250-4fca-9798-25780ae50b00", "address": "fa:16:3e:2c:0a:82", "network": {"id": "22a2f26f-6020-43a0-9e3e-32d28145dd2b", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1080215875-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "87f548d252244b759e3f157af9ab875d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "43798f54-0c07-4417-a23f-58bb6b7e204b", "external-id": "nsx-vlan-transportzone-571", "segmentation_id": 571, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape5c62d99-32", "ovs_interfaceid": "e5c62d99-3250-4fca-9798-25780ae50b00", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 849.354732] env[61923]: DEBUG oslo_concurrency.lockutils [None req-f39c3939-2621-40b9-927d-ae5b7aafe54d tempest-ServerTagsTestJSON-19215696 tempest-ServerTagsTestJSON-19215696-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 849.371092] env[61923]: DEBUG oslo_concurrency.lockutils [req-25d38eb0-626b-4af2-9c64-c97c7f476eef req-e519a948-7fb1-46b5-8d26-70aa4b55f26d service nova] Releasing lock "refresh_cache-422d06d6-6932-46c3-bb25-841e6f65c028" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 849.390265] env[61923]: DEBUG nova.compute.manager [None req-52eadf86-cc45-4cae-af2a-645ac3cb8658 tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] [instance: 87ae37dc-d0d0-4b76-afdd-0ba3e8f6ce0b] Start spawning the instance on the hypervisor. {{(pid=61923) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 849.406035] env[61923]: DEBUG nova.compute.manager [req-80af499b-ac12-4397-bd9b-716f21ff6987 req-f70ebdd4-67cf-43e1-b12d-7499e1f0fbeb service nova] [instance: af3c317a-4007-4cea-a060-1e7dde5ce49e] Received event network-changed-e5c62d99-3250-4fca-9798-25780ae50b00 {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 849.406035] env[61923]: DEBUG nova.compute.manager [req-80af499b-ac12-4397-bd9b-716f21ff6987 req-f70ebdd4-67cf-43e1-b12d-7499e1f0fbeb service nova] [instance: af3c317a-4007-4cea-a060-1e7dde5ce49e] Refreshing instance network info cache due to event network-changed-e5c62d99-3250-4fca-9798-25780ae50b00. {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 849.406035] env[61923]: DEBUG oslo_concurrency.lockutils [req-80af499b-ac12-4397-bd9b-716f21ff6987 req-f70ebdd4-67cf-43e1-b12d-7499e1f0fbeb service nova] Acquiring lock "refresh_cache-af3c317a-4007-4cea-a060-1e7dde5ce49e" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 849.422264] env[61923]: DEBUG nova.virt.hardware [None req-52eadf86-cc45-4cae-af2a-645ac3cb8658 tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-29T20:07:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-29T20:07:35Z,direct_url=,disk_format='vmdk',id=0f153f63-ae0a-45b1-b7f5-7f9b673c947f,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='4e7b5e3b3c3443c8bd5c70f8a15739f5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-29T20:07:36Z,virtual_size=,visibility=), allow threads: False {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 849.422587] env[61923]: DEBUG nova.virt.hardware [None req-52eadf86-cc45-4cae-af2a-645ac3cb8658 tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Flavor limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 849.422813] env[61923]: DEBUG nova.virt.hardware [None req-52eadf86-cc45-4cae-af2a-645ac3cb8658 tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Image limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 849.423099] env[61923]: DEBUG nova.virt.hardware [None req-52eadf86-cc45-4cae-af2a-645ac3cb8658 tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Flavor pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 849.423493] env[61923]: DEBUG nova.virt.hardware [None req-52eadf86-cc45-4cae-af2a-645ac3cb8658 tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Image pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 849.423731] env[61923]: DEBUG nova.virt.hardware [None req-52eadf86-cc45-4cae-af2a-645ac3cb8658 tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 849.424047] env[61923]: DEBUG nova.virt.hardware [None req-52eadf86-cc45-4cae-af2a-645ac3cb8658 tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 849.424288] env[61923]: DEBUG nova.virt.hardware [None req-52eadf86-cc45-4cae-af2a-645ac3cb8658 tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 849.424538] env[61923]: DEBUG nova.virt.hardware [None req-52eadf86-cc45-4cae-af2a-645ac3cb8658 tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Got 1 possible topologies {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 849.424775] env[61923]: DEBUG nova.virt.hardware [None req-52eadf86-cc45-4cae-af2a-645ac3cb8658 tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 849.425035] env[61923]: DEBUG nova.virt.hardware [None req-52eadf86-cc45-4cae-af2a-645ac3cb8658 tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 849.426839] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5ff8fd5-5dd6-475f-9cbf-f0eb66e291d5 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.436824] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c03c091c-a18d-4a11-a1d1-3fd51d1bb79f {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.464340] env[61923]: DEBUG nova.network.neutron [req-a2932dc2-1708-4570-88c9-5ea7176fcbeb req-ec160d3c-e143-45c2-81bc-0454045f558b service nova] [instance: fc4fdf89-4615-4586-9dcd-ea90dc258361] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 849.660980] env[61923]: DEBUG nova.network.neutron [req-a2932dc2-1708-4570-88c9-5ea7176fcbeb req-ec160d3c-e143-45c2-81bc-0454045f558b service nova] [instance: fc4fdf89-4615-4586-9dcd-ea90dc258361] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 849.716558] env[61923]: DEBUG oslo_vmware.api [None req-513daa64-2b05-4c0d-94e9-55b95de63eec tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Task: {'id': task-1377584, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.102806} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 849.719514] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-513daa64-2b05-4c0d-94e9-55b95de63eec tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] [instance: 59198f4d-4dde-4eaf-9f6c-a962cbe53c6e] Extended root virtual disk {{(pid=61923) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 849.719816] env[61923]: DEBUG oslo_vmware.api [None req-e62df7da-42b5-4263-bcf8-001a2d06427a tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Task: {'id': task-1377583, 'name': PowerOnVM_Task, 'duration_secs': 1.115345} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 849.721709] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-500ff9b2-b218-406b-98d5-cb37f9013d51 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.723284] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-e62df7da-42b5-4263-bcf8-001a2d06427a tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: 73692517-1816-4e9b-ab2d-8265c683e83d] Powered on the VM {{(pid=61923) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 849.723493] env[61923]: INFO nova.compute.manager [None req-e62df7da-42b5-4263-bcf8-001a2d06427a tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: 73692517-1816-4e9b-ab2d-8265c683e83d] Took 9.83 seconds to spawn the instance on the hypervisor. [ 849.723668] env[61923]: DEBUG nova.compute.manager [None req-e62df7da-42b5-4263-bcf8-001a2d06427a tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: 73692517-1816-4e9b-ab2d-8265c683e83d] Checking state {{(pid=61923) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 849.724463] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a220986-36a0-4fcf-93d2-6ef7a0806853 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.756175] env[61923]: DEBUG nova.virt.vmwareapi.volumeops [None req-513daa64-2b05-4c0d-94e9-55b95de63eec tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] [instance: 59198f4d-4dde-4eaf-9f6c-a962cbe53c6e] Reconfiguring VM instance instance-00000045 to attach disk [datastore2] 59198f4d-4dde-4eaf-9f6c-a962cbe53c6e/59198f4d-4dde-4eaf-9f6c-a962cbe53c6e.vmdk or device None with type sparse {{(pid=61923) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 849.756583] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d4e76a25-3d57-4d41-b1cc-f08372ebbee6 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.777410] env[61923]: DEBUG oslo_vmware.api [None req-513daa64-2b05-4c0d-94e9-55b95de63eec tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Waiting for the task: (returnval){ [ 849.777410] env[61923]: value = "task-1377585" [ 849.777410] env[61923]: _type = "Task" [ 849.777410] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 849.787472] env[61923]: DEBUG oslo_concurrency.lockutils [None req-5fbf0568-2a85-4e3c-9b63-7d91ecc77593 tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.428s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 849.787968] env[61923]: DEBUG nova.compute.manager [None req-5fbf0568-2a85-4e3c-9b63-7d91ecc77593 tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] [instance: 2719569b-8572-4199-8158-7bb367d17dc5] Start building networks asynchronously for instance. {{(pid=61923) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 849.790573] env[61923]: DEBUG oslo_vmware.api [None req-513daa64-2b05-4c0d-94e9-55b95de63eec tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Task: {'id': task-1377585, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.794022] env[61923]: DEBUG oslo_concurrency.lockutils [None req-cdbe0145-cd3c-4551-af87-0c7f8b53f26c tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 9.689s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 849.794022] env[61923]: INFO nova.compute.claims [None req-cdbe0145-cd3c-4551-af87-0c7f8b53f26c tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] [instance: 444affa5-a7ed-4a17-9015-9fd5724aab64] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 849.827255] env[61923]: DEBUG oslo_concurrency.lockutils [None req-1214e428-bf00-4cbe-9e1e-7c8be1612c36 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] Releasing lock "refresh_cache-af3c317a-4007-4cea-a060-1e7dde5ce49e" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 849.827577] env[61923]: DEBUG nova.compute.manager [None req-1214e428-bf00-4cbe-9e1e-7c8be1612c36 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] [instance: af3c317a-4007-4cea-a060-1e7dde5ce49e] Instance network_info: |[{"id": "e5c62d99-3250-4fca-9798-25780ae50b00", "address": "fa:16:3e:2c:0a:82", "network": {"id": "22a2f26f-6020-43a0-9e3e-32d28145dd2b", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1080215875-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "87f548d252244b759e3f157af9ab875d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "43798f54-0c07-4417-a23f-58bb6b7e204b", "external-id": "nsx-vlan-transportzone-571", "segmentation_id": 571, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape5c62d99-32", "ovs_interfaceid": "e5c62d99-3250-4fca-9798-25780ae50b00", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61923) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 849.827862] env[61923]: DEBUG oslo_concurrency.lockutils [req-80af499b-ac12-4397-bd9b-716f21ff6987 req-f70ebdd4-67cf-43e1-b12d-7499e1f0fbeb service nova] Acquired lock "refresh_cache-af3c317a-4007-4cea-a060-1e7dde5ce49e" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 849.828042] env[61923]: DEBUG nova.network.neutron [req-80af499b-ac12-4397-bd9b-716f21ff6987 req-f70ebdd4-67cf-43e1-b12d-7499e1f0fbeb service nova] [instance: af3c317a-4007-4cea-a060-1e7dde5ce49e] Refreshing network info cache for port e5c62d99-3250-4fca-9798-25780ae50b00 {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 849.829918] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-1214e428-bf00-4cbe-9e1e-7c8be1612c36 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] [instance: af3c317a-4007-4cea-a060-1e7dde5ce49e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:2c:0a:82', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '43798f54-0c07-4417-a23f-58bb6b7e204b', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'e5c62d99-3250-4fca-9798-25780ae50b00', 'vif_model': 'vmxnet3'}] {{(pid=61923) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 849.837283] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-1214e428-bf00-4cbe-9e1e-7c8be1612c36 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] Creating folder: Project (87f548d252244b759e3f157af9ab875d). Parent ref: group-v292629. {{(pid=61923) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 849.839125] env[61923]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-5aedaf51-312c-4e0e-bbee-47a3c2b05331 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.849029] env[61923]: INFO nova.virt.vmwareapi.vm_util [None req-1214e428-bf00-4cbe-9e1e-7c8be1612c36 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] Created folder: Project (87f548d252244b759e3f157af9ab875d) in parent group-v292629. [ 849.849285] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-1214e428-bf00-4cbe-9e1e-7c8be1612c36 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] Creating folder: Instances. Parent ref: group-v292672. {{(pid=61923) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 849.849565] env[61923]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-bcaa4bb1-a4fe-4b87-b9c4-65bf9dd35376 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.859220] env[61923]: INFO nova.virt.vmwareapi.vm_util [None req-1214e428-bf00-4cbe-9e1e-7c8be1612c36 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] Created folder: Instances in parent group-v292672. [ 849.859459] env[61923]: DEBUG oslo.service.loopingcall [None req-1214e428-bf00-4cbe-9e1e-7c8be1612c36 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61923) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 849.859643] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: af3c317a-4007-4cea-a060-1e7dde5ce49e] Creating VM on the ESX host {{(pid=61923) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 849.860115] env[61923]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-69f4562b-5a8b-46cb-bf58-a5ac3e973fab {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.879456] env[61923]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 849.879456] env[61923]: value = "task-1377588" [ 849.879456] env[61923]: _type = "Task" [ 849.879456] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 849.889345] env[61923]: DEBUG oslo_vmware.api [-] Task: {'id': task-1377588, 'name': CreateVM_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.020702] env[61923]: DEBUG nova.network.neutron [None req-80740a2c-9779-4afc-be30-c4469cd3724d tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] [instance: fc4fdf89-4615-4586-9dcd-ea90dc258361] Successfully updated port: a36809ea-2e4a-48ff-ac61-305da772b282 {{(pid=61923) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 850.164914] env[61923]: DEBUG oslo_concurrency.lockutils [req-a2932dc2-1708-4570-88c9-5ea7176fcbeb req-ec160d3c-e143-45c2-81bc-0454045f558b service nova] Releasing lock "refresh_cache-fc4fdf89-4615-4586-9dcd-ea90dc258361" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 850.165251] env[61923]: DEBUG nova.compute.manager [req-a2932dc2-1708-4570-88c9-5ea7176fcbeb req-ec160d3c-e143-45c2-81bc-0454045f558b service nova] [instance: 48825592-dcdc-41d5-9fbf-500d1f31017b] Received event network-vif-deleted-04aea882-6981-45ab-a2d9-685c86089980 {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 850.264818] env[61923]: INFO nova.compute.manager [None req-e62df7da-42b5-4263-bcf8-001a2d06427a tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: 73692517-1816-4e9b-ab2d-8265c683e83d] Took 29.04 seconds to build instance. [ 850.288628] env[61923]: DEBUG oslo_vmware.api [None req-513daa64-2b05-4c0d-94e9-55b95de63eec tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Task: {'id': task-1377585, 'name': ReconfigVM_Task, 'duration_secs': 0.301767} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 850.288919] env[61923]: DEBUG nova.virt.vmwareapi.volumeops [None req-513daa64-2b05-4c0d-94e9-55b95de63eec tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] [instance: 59198f4d-4dde-4eaf-9f6c-a962cbe53c6e] Reconfigured VM instance instance-00000045 to attach disk [datastore2] 59198f4d-4dde-4eaf-9f6c-a962cbe53c6e/59198f4d-4dde-4eaf-9f6c-a962cbe53c6e.vmdk or device None with type sparse {{(pid=61923) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 850.289557] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c3e22714-7aba-48fd-ae9f-9dde92f23150 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.291922] env[61923]: DEBUG nova.compute.utils [None req-5fbf0568-2a85-4e3c-9b63-7d91ecc77593 tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Using /dev/sd instead of None {{(pid=61923) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 850.294355] env[61923]: DEBUG nova.compute.manager [None req-5fbf0568-2a85-4e3c-9b63-7d91ecc77593 tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] [instance: 2719569b-8572-4199-8158-7bb367d17dc5] Allocating IP information in the background. {{(pid=61923) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 850.294355] env[61923]: DEBUG nova.network.neutron [None req-5fbf0568-2a85-4e3c-9b63-7d91ecc77593 tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] [instance: 2719569b-8572-4199-8158-7bb367d17dc5] allocate_for_instance() {{(pid=61923) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 850.301026] env[61923]: DEBUG oslo_vmware.api [None req-513daa64-2b05-4c0d-94e9-55b95de63eec tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Waiting for the task: (returnval){ [ 850.301026] env[61923]: value = "task-1377589" [ 850.301026] env[61923]: _type = "Task" [ 850.301026] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 850.310812] env[61923]: DEBUG oslo_vmware.api [None req-513daa64-2b05-4c0d-94e9-55b95de63eec tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Task: {'id': task-1377589, 'name': Rename_Task} progress is 5%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.344548] env[61923]: DEBUG nova.policy [None req-5fbf0568-2a85-4e3c-9b63-7d91ecc77593 tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6e7ae11a571e4252ba52dae60db91b34', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd946a91d4cf9412a838fbcb74cb5da98', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61923) authorize /opt/stack/nova/nova/policy.py:201}} [ 850.392416] env[61923]: DEBUG oslo_vmware.api [-] Task: {'id': task-1377588, 'name': CreateVM_Task, 'duration_secs': 0.372213} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 850.392734] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: af3c317a-4007-4cea-a060-1e7dde5ce49e] Created VM on the ESX host {{(pid=61923) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 850.393269] env[61923]: DEBUG oslo_concurrency.lockutils [None req-1214e428-bf00-4cbe-9e1e-7c8be1612c36 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 850.393437] env[61923]: DEBUG oslo_concurrency.lockutils [None req-1214e428-bf00-4cbe-9e1e-7c8be1612c36 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 850.393743] env[61923]: DEBUG oslo_concurrency.lockutils [None req-1214e428-bf00-4cbe-9e1e-7c8be1612c36 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 850.394010] env[61923]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ab9b553c-5791-4763-82ed-3921bec6436e {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.398365] env[61923]: DEBUG oslo_vmware.api [None req-1214e428-bf00-4cbe-9e1e-7c8be1612c36 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] Waiting for the task: (returnval){ [ 850.398365] env[61923]: value = "session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52dcae0b-0500-6858-8156-8f85c9d7e766" [ 850.398365] env[61923]: _type = "Task" [ 850.398365] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 850.406387] env[61923]: DEBUG oslo_vmware.api [None req-1214e428-bf00-4cbe-9e1e-7c8be1612c36 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52dcae0b-0500-6858-8156-8f85c9d7e766, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.767164] env[61923]: DEBUG oslo_concurrency.lockutils [None req-e62df7da-42b5-4263-bcf8-001a2d06427a tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Lock "73692517-1816-4e9b-ab2d-8265c683e83d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 89.191s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 850.794854] env[61923]: DEBUG nova.compute.manager [None req-5fbf0568-2a85-4e3c-9b63-7d91ecc77593 tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] [instance: 2719569b-8572-4199-8158-7bb367d17dc5] Start building block device mappings for instance. {{(pid=61923) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 850.812807] env[61923]: DEBUG oslo_vmware.api [None req-513daa64-2b05-4c0d-94e9-55b95de63eec tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Task: {'id': task-1377589, 'name': Rename_Task} progress is 99%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.820875] env[61923]: DEBUG nova.network.neutron [req-80af499b-ac12-4397-bd9b-716f21ff6987 req-f70ebdd4-67cf-43e1-b12d-7499e1f0fbeb service nova] [instance: af3c317a-4007-4cea-a060-1e7dde5ce49e] Updated VIF entry in instance network info cache for port e5c62d99-3250-4fca-9798-25780ae50b00. {{(pid=61923) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 850.821258] env[61923]: DEBUG nova.network.neutron [req-80af499b-ac12-4397-bd9b-716f21ff6987 req-f70ebdd4-67cf-43e1-b12d-7499e1f0fbeb service nova] [instance: af3c317a-4007-4cea-a060-1e7dde5ce49e] Updating instance_info_cache with network_info: [{"id": "e5c62d99-3250-4fca-9798-25780ae50b00", "address": "fa:16:3e:2c:0a:82", "network": {"id": "22a2f26f-6020-43a0-9e3e-32d28145dd2b", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1080215875-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "87f548d252244b759e3f157af9ab875d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "43798f54-0c07-4417-a23f-58bb6b7e204b", "external-id": "nsx-vlan-transportzone-571", "segmentation_id": 571, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape5c62d99-32", "ovs_interfaceid": "e5c62d99-3250-4fca-9798-25780ae50b00", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 850.907625] env[61923]: DEBUG oslo_vmware.api [None req-1214e428-bf00-4cbe-9e1e-7c8be1612c36 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52dcae0b-0500-6858-8156-8f85c9d7e766, 'name': SearchDatastore_Task, 'duration_secs': 0.011022} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 850.907927] env[61923]: DEBUG oslo_concurrency.lockutils [None req-1214e428-bf00-4cbe-9e1e-7c8be1612c36 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 850.908175] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-1214e428-bf00-4cbe-9e1e-7c8be1612c36 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] [instance: af3c317a-4007-4cea-a060-1e7dde5ce49e] Processing image 0f153f63-ae0a-45b1-b7f5-7f9b673c947f {{(pid=61923) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 850.908409] env[61923]: DEBUG oslo_concurrency.lockutils [None req-1214e428-bf00-4cbe-9e1e-7c8be1612c36 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 850.908556] env[61923]: DEBUG oslo_concurrency.lockutils [None req-1214e428-bf00-4cbe-9e1e-7c8be1612c36 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 850.908731] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-1214e428-bf00-4cbe-9e1e-7c8be1612c36 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61923) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 850.909026] env[61923]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-1e159d6c-82fc-4a89-9376-06114edae933 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.921495] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-1214e428-bf00-4cbe-9e1e-7c8be1612c36 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61923) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 850.921680] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-1214e428-bf00-4cbe-9e1e-7c8be1612c36 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61923) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 850.922419] env[61923]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-26adfdd8-98cf-483a-a9d4-2f76829e5e61 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.928257] env[61923]: DEBUG oslo_vmware.api [None req-1214e428-bf00-4cbe-9e1e-7c8be1612c36 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] Waiting for the task: (returnval){ [ 850.928257] env[61923]: value = "session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]522880f9-610e-83d1-2ca4-626909cdc1df" [ 850.928257] env[61923]: _type = "Task" [ 850.928257] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 850.939670] env[61923]: DEBUG oslo_vmware.api [None req-1214e428-bf00-4cbe-9e1e-7c8be1612c36 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]522880f9-610e-83d1-2ca4-626909cdc1df, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.036456] env[61923]: DEBUG nova.network.neutron [None req-5fbf0568-2a85-4e3c-9b63-7d91ecc77593 tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] [instance: 2719569b-8572-4199-8158-7bb367d17dc5] Successfully created port: e092aadb-6904-43bc-b14c-0b1e010c4b2f {{(pid=61923) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 851.057753] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db5f71c4-b1ce-4976-b2f7-d34cbed1d94e {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.065347] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8414186-0f94-44b5-bd40-8a6757f84f67 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.095906] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f002eae-c99c-4512-b3c2-25d54ce4ed50 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.102759] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f18efd3-7fce-4979-a1c2-3b7fbac346a5 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.115661] env[61923]: DEBUG nova.compute.provider_tree [None req-cdbe0145-cd3c-4551-af87-0c7f8b53f26c tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 851.229442] env[61923]: DEBUG nova.network.neutron [None req-52eadf86-cc45-4cae-af2a-645ac3cb8658 tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] [instance: 87ae37dc-d0d0-4b76-afdd-0ba3e8f6ce0b] Successfully updated port: b0e2ee14-b051-47bc-9064-50f09916c0cb {{(pid=61923) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 851.243604] env[61923]: DEBUG nova.compute.manager [req-54ad2863-eedc-4ce5-8870-134bc4bb2a57 req-33ba4597-a9dd-4d4d-8b0f-664bd69ea47e service nova] [instance: fc4fdf89-4615-4586-9dcd-ea90dc258361] Received event network-vif-plugged-a36809ea-2e4a-48ff-ac61-305da772b282 {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 851.243896] env[61923]: DEBUG oslo_concurrency.lockutils [req-54ad2863-eedc-4ce5-8870-134bc4bb2a57 req-33ba4597-a9dd-4d4d-8b0f-664bd69ea47e service nova] Acquiring lock "fc4fdf89-4615-4586-9dcd-ea90dc258361-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 851.244868] env[61923]: DEBUG oslo_concurrency.lockutils [req-54ad2863-eedc-4ce5-8870-134bc4bb2a57 req-33ba4597-a9dd-4d4d-8b0f-664bd69ea47e service nova] Lock "fc4fdf89-4615-4586-9dcd-ea90dc258361-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 851.245388] env[61923]: DEBUG oslo_concurrency.lockutils [req-54ad2863-eedc-4ce5-8870-134bc4bb2a57 req-33ba4597-a9dd-4d4d-8b0f-664bd69ea47e service nova] Lock "fc4fdf89-4615-4586-9dcd-ea90dc258361-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.001s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 851.245583] env[61923]: DEBUG nova.compute.manager [req-54ad2863-eedc-4ce5-8870-134bc4bb2a57 req-33ba4597-a9dd-4d4d-8b0f-664bd69ea47e service nova] [instance: fc4fdf89-4615-4586-9dcd-ea90dc258361] No waiting events found dispatching network-vif-plugged-a36809ea-2e4a-48ff-ac61-305da772b282 {{(pid=61923) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 851.245761] env[61923]: WARNING nova.compute.manager [req-54ad2863-eedc-4ce5-8870-134bc4bb2a57 req-33ba4597-a9dd-4d4d-8b0f-664bd69ea47e service nova] [instance: fc4fdf89-4615-4586-9dcd-ea90dc258361] Received unexpected event network-vif-plugged-a36809ea-2e4a-48ff-ac61-305da772b282 for instance with vm_state building and task_state spawning. [ 851.245925] env[61923]: DEBUG nova.compute.manager [req-54ad2863-eedc-4ce5-8870-134bc4bb2a57 req-33ba4597-a9dd-4d4d-8b0f-664bd69ea47e service nova] [instance: fc4fdf89-4615-4586-9dcd-ea90dc258361] Received event network-changed-a36809ea-2e4a-48ff-ac61-305da772b282 {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 851.246096] env[61923]: DEBUG nova.compute.manager [req-54ad2863-eedc-4ce5-8870-134bc4bb2a57 req-33ba4597-a9dd-4d4d-8b0f-664bd69ea47e service nova] [instance: fc4fdf89-4615-4586-9dcd-ea90dc258361] Refreshing instance network info cache due to event network-changed-a36809ea-2e4a-48ff-ac61-305da772b282. {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 851.246281] env[61923]: DEBUG oslo_concurrency.lockutils [req-54ad2863-eedc-4ce5-8870-134bc4bb2a57 req-33ba4597-a9dd-4d4d-8b0f-664bd69ea47e service nova] Acquiring lock "refresh_cache-fc4fdf89-4615-4586-9dcd-ea90dc258361" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 851.246696] env[61923]: DEBUG oslo_concurrency.lockutils [req-54ad2863-eedc-4ce5-8870-134bc4bb2a57 req-33ba4597-a9dd-4d4d-8b0f-664bd69ea47e service nova] Acquired lock "refresh_cache-fc4fdf89-4615-4586-9dcd-ea90dc258361" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 851.246696] env[61923]: DEBUG nova.network.neutron [req-54ad2863-eedc-4ce5-8870-134bc4bb2a57 req-33ba4597-a9dd-4d4d-8b0f-664bd69ea47e service nova] [instance: fc4fdf89-4615-4586-9dcd-ea90dc258361] Refreshing network info cache for port a36809ea-2e4a-48ff-ac61-305da772b282 {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 851.269597] env[61923]: DEBUG nova.compute.manager [None req-9887b108-13fe-4987-ad6f-1ef139e2a15d tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] [instance: 6012d3b5-739c-4762-9bb4-09c51171dcd7] Starting instance... {{(pid=61923) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 851.314416] env[61923]: DEBUG oslo_vmware.api [None req-513daa64-2b05-4c0d-94e9-55b95de63eec tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Task: {'id': task-1377589, 'name': Rename_Task} progress is 99%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.326938] env[61923]: DEBUG oslo_concurrency.lockutils [req-80af499b-ac12-4397-bd9b-716f21ff6987 req-f70ebdd4-67cf-43e1-b12d-7499e1f0fbeb service nova] Releasing lock "refresh_cache-af3c317a-4007-4cea-a060-1e7dde5ce49e" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 851.442437] env[61923]: DEBUG oslo_vmware.api [None req-1214e428-bf00-4cbe-9e1e-7c8be1612c36 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]522880f9-610e-83d1-2ca4-626909cdc1df, 'name': SearchDatastore_Task, 'duration_secs': 0.020823} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 851.443893] env[61923]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e4bab500-5688-42d9-a9d9-ef2e83deaa5b {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.451376] env[61923]: DEBUG oslo_vmware.api [None req-1214e428-bf00-4cbe-9e1e-7c8be1612c36 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] Waiting for the task: (returnval){ [ 851.451376] env[61923]: value = "session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52f50121-6032-89b4-2c57-d61fdfc23547" [ 851.451376] env[61923]: _type = "Task" [ 851.451376] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 851.460386] env[61923]: DEBUG oslo_vmware.api [None req-1214e428-bf00-4cbe-9e1e-7c8be1612c36 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52f50121-6032-89b4-2c57-d61fdfc23547, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.618691] env[61923]: DEBUG nova.scheduler.client.report [None req-cdbe0145-cd3c-4551-af87-0c7f8b53f26c tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 851.729288] env[61923]: DEBUG oslo_concurrency.lockutils [None req-52eadf86-cc45-4cae-af2a-645ac3cb8658 tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Acquiring lock "refresh_cache-87ae37dc-d0d0-4b76-afdd-0ba3e8f6ce0b" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 851.729411] env[61923]: DEBUG oslo_concurrency.lockutils [None req-52eadf86-cc45-4cae-af2a-645ac3cb8658 tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Acquired lock "refresh_cache-87ae37dc-d0d0-4b76-afdd-0ba3e8f6ce0b" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 851.729630] env[61923]: DEBUG nova.network.neutron [None req-52eadf86-cc45-4cae-af2a-645ac3cb8658 tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] [instance: 87ae37dc-d0d0-4b76-afdd-0ba3e8f6ce0b] Building network info cache for instance {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 851.795561] env[61923]: DEBUG nova.network.neutron [req-54ad2863-eedc-4ce5-8870-134bc4bb2a57 req-33ba4597-a9dd-4d4d-8b0f-664bd69ea47e service nova] [instance: fc4fdf89-4615-4586-9dcd-ea90dc258361] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 851.798356] env[61923]: DEBUG oslo_concurrency.lockutils [None req-9887b108-13fe-4987-ad6f-1ef139e2a15d tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 851.810816] env[61923]: DEBUG nova.compute.manager [None req-5fbf0568-2a85-4e3c-9b63-7d91ecc77593 tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] [instance: 2719569b-8572-4199-8158-7bb367d17dc5] Start spawning the instance on the hypervisor. {{(pid=61923) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 851.818223] env[61923]: DEBUG oslo_vmware.api [None req-513daa64-2b05-4c0d-94e9-55b95de63eec tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Task: {'id': task-1377589, 'name': Rename_Task, 'duration_secs': 1.141237} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 851.818507] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-513daa64-2b05-4c0d-94e9-55b95de63eec tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] [instance: 59198f4d-4dde-4eaf-9f6c-a962cbe53c6e] Powering on the VM {{(pid=61923) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 851.818771] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b34920da-2c5c-4dd4-abb8-47d937e57f7a {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.826417] env[61923]: DEBUG oslo_vmware.api [None req-513daa64-2b05-4c0d-94e9-55b95de63eec tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Waiting for the task: (returnval){ [ 851.826417] env[61923]: value = "task-1377590" [ 851.826417] env[61923]: _type = "Task" [ 851.826417] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 851.836147] env[61923]: DEBUG oslo_vmware.api [None req-513daa64-2b05-4c0d-94e9-55b95de63eec tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Task: {'id': task-1377590, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.840371] env[61923]: DEBUG nova.virt.hardware [None req-5fbf0568-2a85-4e3c-9b63-7d91ecc77593 tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-29T20:07:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-29T20:07:35Z,direct_url=,disk_format='vmdk',id=0f153f63-ae0a-45b1-b7f5-7f9b673c947f,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='4e7b5e3b3c3443c8bd5c70f8a15739f5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-29T20:07:36Z,virtual_size=,visibility=), allow threads: False {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 851.840630] env[61923]: DEBUG nova.virt.hardware [None req-5fbf0568-2a85-4e3c-9b63-7d91ecc77593 tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Flavor limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 851.840741] env[61923]: DEBUG nova.virt.hardware [None req-5fbf0568-2a85-4e3c-9b63-7d91ecc77593 tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Image limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 851.840917] env[61923]: DEBUG nova.virt.hardware [None req-5fbf0568-2a85-4e3c-9b63-7d91ecc77593 tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Flavor pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 851.841105] env[61923]: DEBUG nova.virt.hardware [None req-5fbf0568-2a85-4e3c-9b63-7d91ecc77593 tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Image pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 851.841270] env[61923]: DEBUG nova.virt.hardware [None req-5fbf0568-2a85-4e3c-9b63-7d91ecc77593 tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 851.841494] env[61923]: DEBUG nova.virt.hardware [None req-5fbf0568-2a85-4e3c-9b63-7d91ecc77593 tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 851.841636] env[61923]: DEBUG nova.virt.hardware [None req-5fbf0568-2a85-4e3c-9b63-7d91ecc77593 tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 851.841796] env[61923]: DEBUG nova.virt.hardware [None req-5fbf0568-2a85-4e3c-9b63-7d91ecc77593 tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Got 1 possible topologies {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 851.841954] env[61923]: DEBUG nova.virt.hardware [None req-5fbf0568-2a85-4e3c-9b63-7d91ecc77593 tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 851.842135] env[61923]: DEBUG nova.virt.hardware [None req-5fbf0568-2a85-4e3c-9b63-7d91ecc77593 tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 851.842910] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7906e817-3295-49b9-9d5d-aa01a85de4d6 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.850096] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a18450e-91b1-42fc-a69c-34aa51d72434 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.893124] env[61923]: DEBUG nova.network.neutron [req-54ad2863-eedc-4ce5-8870-134bc4bb2a57 req-33ba4597-a9dd-4d4d-8b0f-664bd69ea47e service nova] [instance: fc4fdf89-4615-4586-9dcd-ea90dc258361] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 851.962461] env[61923]: DEBUG oslo_vmware.api [None req-1214e428-bf00-4cbe-9e1e-7c8be1612c36 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52f50121-6032-89b4-2c57-d61fdfc23547, 'name': SearchDatastore_Task, 'duration_secs': 0.009296} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 851.962736] env[61923]: DEBUG oslo_concurrency.lockutils [None req-1214e428-bf00-4cbe-9e1e-7c8be1612c36 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 851.962997] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-1214e428-bf00-4cbe-9e1e-7c8be1612c36 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk to [datastore1] af3c317a-4007-4cea-a060-1e7dde5ce49e/af3c317a-4007-4cea-a060-1e7dde5ce49e.vmdk {{(pid=61923) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 851.963276] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-8d351b99-1b63-487f-b8e0-917c47f4e013 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.970090] env[61923]: DEBUG oslo_vmware.api [None req-1214e428-bf00-4cbe-9e1e-7c8be1612c36 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] Waiting for the task: (returnval){ [ 851.970090] env[61923]: value = "task-1377591" [ 851.970090] env[61923]: _type = "Task" [ 851.970090] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 851.978307] env[61923]: DEBUG oslo_vmware.api [None req-1214e428-bf00-4cbe-9e1e-7c8be1612c36 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] Task: {'id': task-1377591, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 852.036487] env[61923]: DEBUG nova.network.neutron [None req-80740a2c-9779-4afc-be30-c4469cd3724d tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] [instance: fc4fdf89-4615-4586-9dcd-ea90dc258361] Successfully updated port: 0738a978-9439-4919-8894-4b264480593e {{(pid=61923) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 852.069009] env[61923]: DEBUG nova.compute.manager [req-7bef4356-16f6-4ac8-be8b-2830e908edf3 req-4dccd123-0004-4f5b-9f80-255d49150881 service nova] [instance: fc4fdf89-4615-4586-9dcd-ea90dc258361] Received event network-vif-plugged-0738a978-9439-4919-8894-4b264480593e {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 852.069255] env[61923]: DEBUG oslo_concurrency.lockutils [req-7bef4356-16f6-4ac8-be8b-2830e908edf3 req-4dccd123-0004-4f5b-9f80-255d49150881 service nova] Acquiring lock "fc4fdf89-4615-4586-9dcd-ea90dc258361-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 852.069452] env[61923]: DEBUG oslo_concurrency.lockutils [req-7bef4356-16f6-4ac8-be8b-2830e908edf3 req-4dccd123-0004-4f5b-9f80-255d49150881 service nova] Lock "fc4fdf89-4615-4586-9dcd-ea90dc258361-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 852.069590] env[61923]: DEBUG oslo_concurrency.lockutils [req-7bef4356-16f6-4ac8-be8b-2830e908edf3 req-4dccd123-0004-4f5b-9f80-255d49150881 service nova] Lock "fc4fdf89-4615-4586-9dcd-ea90dc258361-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 852.069761] env[61923]: DEBUG nova.compute.manager [req-7bef4356-16f6-4ac8-be8b-2830e908edf3 req-4dccd123-0004-4f5b-9f80-255d49150881 service nova] [instance: fc4fdf89-4615-4586-9dcd-ea90dc258361] No waiting events found dispatching network-vif-plugged-0738a978-9439-4919-8894-4b264480593e {{(pid=61923) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 852.069927] env[61923]: WARNING nova.compute.manager [req-7bef4356-16f6-4ac8-be8b-2830e908edf3 req-4dccd123-0004-4f5b-9f80-255d49150881 service nova] [instance: fc4fdf89-4615-4586-9dcd-ea90dc258361] Received unexpected event network-vif-plugged-0738a978-9439-4919-8894-4b264480593e for instance with vm_state building and task_state spawning. [ 852.123369] env[61923]: DEBUG oslo_concurrency.lockutils [None req-cdbe0145-cd3c-4551-af87-0c7f8b53f26c tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.332s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 852.124644] env[61923]: DEBUG nova.compute.manager [None req-cdbe0145-cd3c-4551-af87-0c7f8b53f26c tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] [instance: 444affa5-a7ed-4a17-9015-9fd5724aab64] Start building networks asynchronously for instance. {{(pid=61923) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 852.126613] env[61923]: DEBUG oslo_concurrency.lockutils [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 9.677s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 852.127654] env[61923]: DEBUG oslo_concurrency.lockutils [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 852.127888] env[61923]: DEBUG nova.compute.resource_tracker [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61923) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 852.128219] env[61923]: DEBUG oslo_concurrency.lockutils [None req-7d4ecc17-51d2-4753-8cb6-fd83e220b388 tempest-ServersNegativeTestMultiTenantJSON-1497120864 tempest-ServersNegativeTestMultiTenantJSON-1497120864-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 8.770s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 852.128424] env[61923]: DEBUG nova.objects.instance [None req-7d4ecc17-51d2-4753-8cb6-fd83e220b388 tempest-ServersNegativeTestMultiTenantJSON-1497120864 tempest-ServersNegativeTestMultiTenantJSON-1497120864-project-member] Lazy-loading 'resources' on Instance uuid 1d1b1dc4-c701-42b2-beba-344d622aef41 {{(pid=61923) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 852.131418] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b68eec3-e8b6-42b7-946c-02edaf7b7f78 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.141938] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de26a941-5d96-4cbc-b041-b8d136d24c9b {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.158463] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-103ccdbe-aba6-44b8-8342-f693aeb6c49f {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.165834] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0fb8afd8-835e-4532-972d-bad0288876d0 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.198560] env[61923]: DEBUG oslo_concurrency.lockutils [None req-b8ed9576-aa00-419c-8225-19ac3c654fd7 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Acquiring lock "73692517-1816-4e9b-ab2d-8265c683e83d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 852.198882] env[61923]: DEBUG oslo_concurrency.lockutils [None req-b8ed9576-aa00-419c-8225-19ac3c654fd7 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Lock "73692517-1816-4e9b-ab2d-8265c683e83d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 852.199098] env[61923]: DEBUG oslo_concurrency.lockutils [None req-b8ed9576-aa00-419c-8225-19ac3c654fd7 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Acquiring lock "73692517-1816-4e9b-ab2d-8265c683e83d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 852.199304] env[61923]: DEBUG oslo_concurrency.lockutils [None req-b8ed9576-aa00-419c-8225-19ac3c654fd7 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Lock "73692517-1816-4e9b-ab2d-8265c683e83d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 852.199576] env[61923]: DEBUG oslo_concurrency.lockutils [None req-b8ed9576-aa00-419c-8225-19ac3c654fd7 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Lock "73692517-1816-4e9b-ab2d-8265c683e83d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 852.201639] env[61923]: INFO nova.compute.manager [None req-b8ed9576-aa00-419c-8225-19ac3c654fd7 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: 73692517-1816-4e9b-ab2d-8265c683e83d] Terminating instance [ 852.203418] env[61923]: DEBUG nova.compute.manager [None req-b8ed9576-aa00-419c-8225-19ac3c654fd7 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: 73692517-1816-4e9b-ab2d-8265c683e83d] Start destroying the instance on the hypervisor. {{(pid=61923) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 852.203615] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-b8ed9576-aa00-419c-8225-19ac3c654fd7 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: 73692517-1816-4e9b-ab2d-8265c683e83d] Destroying instance {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 852.203993] env[61923]: DEBUG nova.compute.resource_tracker [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181418MB free_disk=178GB free_vcpus=48 pci_devices=None {{(pid=61923) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 852.204141] env[61923]: DEBUG oslo_concurrency.lockutils [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 852.205571] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae197bde-5fa9-4da3-bbfe-9fd082c10908 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.213069] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-b8ed9576-aa00-419c-8225-19ac3c654fd7 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: 73692517-1816-4e9b-ab2d-8265c683e83d] Powering off the VM {{(pid=61923) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 852.213720] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-17c251ad-ebb7-4cc2-bed1-4d5aa6bea29a {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.220647] env[61923]: DEBUG oslo_vmware.api [None req-b8ed9576-aa00-419c-8225-19ac3c654fd7 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Waiting for the task: (returnval){ [ 852.220647] env[61923]: value = "task-1377592" [ 852.220647] env[61923]: _type = "Task" [ 852.220647] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 852.265491] env[61923]: DEBUG nova.network.neutron [None req-52eadf86-cc45-4cae-af2a-645ac3cb8658 tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] [instance: 87ae37dc-d0d0-4b76-afdd-0ba3e8f6ce0b] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 852.337777] env[61923]: DEBUG oslo_vmware.api [None req-513daa64-2b05-4c0d-94e9-55b95de63eec tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Task: {'id': task-1377590, 'name': PowerOnVM_Task, 'duration_secs': 0.504073} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 852.340034] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-513daa64-2b05-4c0d-94e9-55b95de63eec tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] [instance: 59198f4d-4dde-4eaf-9f6c-a962cbe53c6e] Powered on the VM {{(pid=61923) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 852.340034] env[61923]: INFO nova.compute.manager [None req-513daa64-2b05-4c0d-94e9-55b95de63eec tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] [instance: 59198f4d-4dde-4eaf-9f6c-a962cbe53c6e] Took 10.12 seconds to spawn the instance on the hypervisor. [ 852.340034] env[61923]: DEBUG nova.compute.manager [None req-513daa64-2b05-4c0d-94e9-55b95de63eec tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] [instance: 59198f4d-4dde-4eaf-9f6c-a962cbe53c6e] Checking state {{(pid=61923) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 852.340034] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7497995-0745-41f2-a85d-98064c590b0d {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.395833] env[61923]: DEBUG oslo_concurrency.lockutils [req-54ad2863-eedc-4ce5-8870-134bc4bb2a57 req-33ba4597-a9dd-4d4d-8b0f-664bd69ea47e service nova] Releasing lock "refresh_cache-fc4fdf89-4615-4586-9dcd-ea90dc258361" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 852.396187] env[61923]: DEBUG nova.compute.manager [req-54ad2863-eedc-4ce5-8870-134bc4bb2a57 req-33ba4597-a9dd-4d4d-8b0f-664bd69ea47e service nova] [instance: 87ae37dc-d0d0-4b76-afdd-0ba3e8f6ce0b] Received event network-vif-plugged-b0e2ee14-b051-47bc-9064-50f09916c0cb {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 852.396433] env[61923]: DEBUG oslo_concurrency.lockutils [req-54ad2863-eedc-4ce5-8870-134bc4bb2a57 req-33ba4597-a9dd-4d4d-8b0f-664bd69ea47e service nova] Acquiring lock "87ae37dc-d0d0-4b76-afdd-0ba3e8f6ce0b-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 852.396670] env[61923]: DEBUG oslo_concurrency.lockutils [req-54ad2863-eedc-4ce5-8870-134bc4bb2a57 req-33ba4597-a9dd-4d4d-8b0f-664bd69ea47e service nova] Lock "87ae37dc-d0d0-4b76-afdd-0ba3e8f6ce0b-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 852.396912] env[61923]: DEBUG oslo_concurrency.lockutils [req-54ad2863-eedc-4ce5-8870-134bc4bb2a57 req-33ba4597-a9dd-4d4d-8b0f-664bd69ea47e service nova] Lock "87ae37dc-d0d0-4b76-afdd-0ba3e8f6ce0b-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 852.397165] env[61923]: DEBUG nova.compute.manager [req-54ad2863-eedc-4ce5-8870-134bc4bb2a57 req-33ba4597-a9dd-4d4d-8b0f-664bd69ea47e service nova] [instance: 87ae37dc-d0d0-4b76-afdd-0ba3e8f6ce0b] No waiting events found dispatching network-vif-plugged-b0e2ee14-b051-47bc-9064-50f09916c0cb {{(pid=61923) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 852.397373] env[61923]: WARNING nova.compute.manager [req-54ad2863-eedc-4ce5-8870-134bc4bb2a57 req-33ba4597-a9dd-4d4d-8b0f-664bd69ea47e service nova] [instance: 87ae37dc-d0d0-4b76-afdd-0ba3e8f6ce0b] Received unexpected event network-vif-plugged-b0e2ee14-b051-47bc-9064-50f09916c0cb for instance with vm_state building and task_state spawning. [ 852.428825] env[61923]: DEBUG nova.network.neutron [None req-52eadf86-cc45-4cae-af2a-645ac3cb8658 tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] [instance: 87ae37dc-d0d0-4b76-afdd-0ba3e8f6ce0b] Updating instance_info_cache with network_info: [{"id": "b0e2ee14-b051-47bc-9064-50f09916c0cb", "address": "fa:16:3e:26:d6:71", "network": {"id": "1f010d1c-1c6c-4ce2-88c8-57af627c8e07", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.147", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "4e7b5e3b3c3443c8bd5c70f8a15739f5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "db68bd64-5b56-49af-a075-13dcf85cb2e0", "external-id": "nsx-vlan-transportzone-590", "segmentation_id": 590, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb0e2ee14-b0", "ovs_interfaceid": "b0e2ee14-b051-47bc-9064-50f09916c0cb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 852.483370] env[61923]: DEBUG oslo_vmware.api [None req-1214e428-bf00-4cbe-9e1e-7c8be1612c36 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] Task: {'id': task-1377591, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 852.538738] env[61923]: DEBUG oslo_concurrency.lockutils [None req-80740a2c-9779-4afc-be30-c4469cd3724d tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] Acquiring lock "refresh_cache-fc4fdf89-4615-4586-9dcd-ea90dc258361" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 852.538978] env[61923]: DEBUG oslo_concurrency.lockutils [None req-80740a2c-9779-4afc-be30-c4469cd3724d tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] Acquired lock "refresh_cache-fc4fdf89-4615-4586-9dcd-ea90dc258361" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 852.539129] env[61923]: DEBUG nova.network.neutron [None req-80740a2c-9779-4afc-be30-c4469cd3724d tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] [instance: fc4fdf89-4615-4586-9dcd-ea90dc258361] Building network info cache for instance {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 852.635474] env[61923]: DEBUG nova.compute.utils [None req-cdbe0145-cd3c-4551-af87-0c7f8b53f26c tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Using /dev/sd instead of None {{(pid=61923) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 852.637257] env[61923]: DEBUG nova.compute.manager [None req-cdbe0145-cd3c-4551-af87-0c7f8b53f26c tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] [instance: 444affa5-a7ed-4a17-9015-9fd5724aab64] Allocating IP information in the background. {{(pid=61923) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 852.637508] env[61923]: DEBUG nova.network.neutron [None req-cdbe0145-cd3c-4551-af87-0c7f8b53f26c tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] [instance: 444affa5-a7ed-4a17-9015-9fd5724aab64] allocate_for_instance() {{(pid=61923) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 852.678303] env[61923]: DEBUG nova.policy [None req-cdbe0145-cd3c-4551-af87-0c7f8b53f26c tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2631257787d44fcbac010b2bf1fd4b0e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '6e260342cd1b41ef8f0ceb15b6438daa', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61923) authorize /opt/stack/nova/nova/policy.py:201}} [ 852.732939] env[61923]: DEBUG oslo_vmware.api [None req-b8ed9576-aa00-419c-8225-19ac3c654fd7 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Task: {'id': task-1377592, 'name': PowerOffVM_Task, 'duration_secs': 0.267542} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 852.733385] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-b8ed9576-aa00-419c-8225-19ac3c654fd7 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: 73692517-1816-4e9b-ab2d-8265c683e83d] Powered off the VM {{(pid=61923) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 852.733522] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-b8ed9576-aa00-419c-8225-19ac3c654fd7 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: 73692517-1816-4e9b-ab2d-8265c683e83d] Unregistering the VM {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 852.733840] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f5c72ad9-806c-4f23-8ec9-8cc9fa48b403 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.803918] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-b8ed9576-aa00-419c-8225-19ac3c654fd7 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: 73692517-1816-4e9b-ab2d-8265c683e83d] Unregistered the VM {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 852.803918] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-b8ed9576-aa00-419c-8225-19ac3c654fd7 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: 73692517-1816-4e9b-ab2d-8265c683e83d] Deleting contents of the VM from datastore datastore2 {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 852.803918] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-b8ed9576-aa00-419c-8225-19ac3c654fd7 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Deleting the datastore file [datastore2] 73692517-1816-4e9b-ab2d-8265c683e83d {{(pid=61923) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 852.803918] env[61923]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b71c6684-abbf-4a04-b0e9-6765fdbd698d {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.809028] env[61923]: DEBUG oslo_vmware.api [None req-b8ed9576-aa00-419c-8225-19ac3c654fd7 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Waiting for the task: (returnval){ [ 852.809028] env[61923]: value = "task-1377594" [ 852.809028] env[61923]: _type = "Task" [ 852.809028] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 852.819457] env[61923]: DEBUG oslo_vmware.api [None req-b8ed9576-aa00-419c-8225-19ac3c654fd7 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Task: {'id': task-1377594, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 852.831604] env[61923]: DEBUG nova.network.neutron [None req-5fbf0568-2a85-4e3c-9b63-7d91ecc77593 tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] [instance: 2719569b-8572-4199-8158-7bb367d17dc5] Successfully updated port: e092aadb-6904-43bc-b14c-0b1e010c4b2f {{(pid=61923) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 852.838018] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ea91979-a9fd-4919-8c4d-c6e09ef0652b {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.844454] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37711102-00e3-453c-9dbe-3d806c7ebcfd {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.883890] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ad4d8ec-1b9f-433f-866e-b478cc0f2056 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.890270] env[61923]: INFO nova.compute.manager [None req-513daa64-2b05-4c0d-94e9-55b95de63eec tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] [instance: 59198f4d-4dde-4eaf-9f6c-a962cbe53c6e] Took 29.88 seconds to build instance. [ 852.895212] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc53cb7b-c938-4298-b7c3-c9e81201a010 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.911363] env[61923]: DEBUG nova.compute.provider_tree [None req-7d4ecc17-51d2-4753-8cb6-fd83e220b388 tempest-ServersNegativeTestMultiTenantJSON-1497120864 tempest-ServersNegativeTestMultiTenantJSON-1497120864-project-member] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 852.932018] env[61923]: DEBUG oslo_concurrency.lockutils [None req-52eadf86-cc45-4cae-af2a-645ac3cb8658 tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Releasing lock "refresh_cache-87ae37dc-d0d0-4b76-afdd-0ba3e8f6ce0b" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 852.932615] env[61923]: DEBUG nova.compute.manager [None req-52eadf86-cc45-4cae-af2a-645ac3cb8658 tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] [instance: 87ae37dc-d0d0-4b76-afdd-0ba3e8f6ce0b] Instance network_info: |[{"id": "b0e2ee14-b051-47bc-9064-50f09916c0cb", "address": "fa:16:3e:26:d6:71", "network": {"id": "1f010d1c-1c6c-4ce2-88c8-57af627c8e07", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.147", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "4e7b5e3b3c3443c8bd5c70f8a15739f5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "db68bd64-5b56-49af-a075-13dcf85cb2e0", "external-id": "nsx-vlan-transportzone-590", "segmentation_id": 590, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb0e2ee14-b0", "ovs_interfaceid": "b0e2ee14-b051-47bc-9064-50f09916c0cb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61923) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 852.933055] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-52eadf86-cc45-4cae-af2a-645ac3cb8658 tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] [instance: 87ae37dc-d0d0-4b76-afdd-0ba3e8f6ce0b] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:26:d6:71', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'db68bd64-5b56-49af-a075-13dcf85cb2e0', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'b0e2ee14-b051-47bc-9064-50f09916c0cb', 'vif_model': 'vmxnet3'}] {{(pid=61923) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 852.940807] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-52eadf86-cc45-4cae-af2a-645ac3cb8658 tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Creating folder: Project (d946a91d4cf9412a838fbcb74cb5da98). Parent ref: group-v292629. {{(pid=61923) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 852.941444] env[61923]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-77c4238b-d70b-4f9b-9e10-b713fad29204 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.946410] env[61923]: DEBUG nova.network.neutron [None req-cdbe0145-cd3c-4551-af87-0c7f8b53f26c tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] [instance: 444affa5-a7ed-4a17-9015-9fd5724aab64] Successfully created port: 7b131b11-927e-4c0c-98f2-714b63683d77 {{(pid=61923) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 852.951760] env[61923]: INFO nova.virt.vmwareapi.vm_util [None req-52eadf86-cc45-4cae-af2a-645ac3cb8658 tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Created folder: Project (d946a91d4cf9412a838fbcb74cb5da98) in parent group-v292629. [ 852.951942] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-52eadf86-cc45-4cae-af2a-645ac3cb8658 tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Creating folder: Instances. Parent ref: group-v292675. {{(pid=61923) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 852.952188] env[61923]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-8d53a4bf-7da7-463d-932e-ead9888b2d41 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.960678] env[61923]: INFO nova.virt.vmwareapi.vm_util [None req-52eadf86-cc45-4cae-af2a-645ac3cb8658 tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Created folder: Instances in parent group-v292675. [ 852.961102] env[61923]: DEBUG oslo.service.loopingcall [None req-52eadf86-cc45-4cae-af2a-645ac3cb8658 tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61923) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 852.961102] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 87ae37dc-d0d0-4b76-afdd-0ba3e8f6ce0b] Creating VM on the ESX host {{(pid=61923) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 852.961304] env[61923]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-9123a89d-003f-43d4-8049-43a9b78fb9ca {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.984090] env[61923]: DEBUG oslo_vmware.api [None req-1214e428-bf00-4cbe-9e1e-7c8be1612c36 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] Task: {'id': task-1377591, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.542558} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 852.985258] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-1214e428-bf00-4cbe-9e1e-7c8be1612c36 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk to [datastore1] af3c317a-4007-4cea-a060-1e7dde5ce49e/af3c317a-4007-4cea-a060-1e7dde5ce49e.vmdk {{(pid=61923) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 852.985473] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-1214e428-bf00-4cbe-9e1e-7c8be1612c36 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] [instance: af3c317a-4007-4cea-a060-1e7dde5ce49e] Extending root virtual disk to 1048576 {{(pid=61923) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 852.985693] env[61923]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 852.985693] env[61923]: value = "task-1377597" [ 852.985693] env[61923]: _type = "Task" [ 852.985693] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 852.985870] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-5e4d6680-5edb-44e6-b3c7-7b4574453122 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.995267] env[61923]: DEBUG oslo_vmware.api [-] Task: {'id': task-1377597, 'name': CreateVM_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 852.996421] env[61923]: DEBUG oslo_vmware.api [None req-1214e428-bf00-4cbe-9e1e-7c8be1612c36 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] Waiting for the task: (returnval){ [ 852.996421] env[61923]: value = "task-1377598" [ 852.996421] env[61923]: _type = "Task" [ 852.996421] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 853.003753] env[61923]: DEBUG oslo_vmware.api [None req-1214e428-bf00-4cbe-9e1e-7c8be1612c36 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] Task: {'id': task-1377598, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.084594] env[61923]: DEBUG nova.network.neutron [None req-80740a2c-9779-4afc-be30-c4469cd3724d tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] [instance: fc4fdf89-4615-4586-9dcd-ea90dc258361] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 853.141956] env[61923]: DEBUG nova.compute.manager [None req-cdbe0145-cd3c-4551-af87-0c7f8b53f26c tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] [instance: 444affa5-a7ed-4a17-9015-9fd5724aab64] Start building block device mappings for instance. {{(pid=61923) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 853.321557] env[61923]: DEBUG oslo_vmware.api [None req-b8ed9576-aa00-419c-8225-19ac3c654fd7 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Task: {'id': task-1377594, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.169412} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 853.323976] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-b8ed9576-aa00-419c-8225-19ac3c654fd7 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Deleted the datastore file {{(pid=61923) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 853.324198] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-b8ed9576-aa00-419c-8225-19ac3c654fd7 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: 73692517-1816-4e9b-ab2d-8265c683e83d] Deleted contents of the VM from datastore datastore2 {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 853.324380] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-b8ed9576-aa00-419c-8225-19ac3c654fd7 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: 73692517-1816-4e9b-ab2d-8265c683e83d] Instance destroyed {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 853.324553] env[61923]: INFO nova.compute.manager [None req-b8ed9576-aa00-419c-8225-19ac3c654fd7 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: 73692517-1816-4e9b-ab2d-8265c683e83d] Took 1.12 seconds to destroy the instance on the hypervisor. [ 853.324791] env[61923]: DEBUG oslo.service.loopingcall [None req-b8ed9576-aa00-419c-8225-19ac3c654fd7 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61923) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 853.324995] env[61923]: DEBUG nova.compute.manager [-] [instance: 73692517-1816-4e9b-ab2d-8265c683e83d] Deallocating network for instance {{(pid=61923) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 853.325104] env[61923]: DEBUG nova.network.neutron [-] [instance: 73692517-1816-4e9b-ab2d-8265c683e83d] deallocate_for_instance() {{(pid=61923) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 853.333734] env[61923]: DEBUG oslo_concurrency.lockutils [None req-5fbf0568-2a85-4e3c-9b63-7d91ecc77593 tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Acquiring lock "refresh_cache-2719569b-8572-4199-8158-7bb367d17dc5" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 853.333868] env[61923]: DEBUG oslo_concurrency.lockutils [None req-5fbf0568-2a85-4e3c-9b63-7d91ecc77593 tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Acquired lock "refresh_cache-2719569b-8572-4199-8158-7bb367d17dc5" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 853.334015] env[61923]: DEBUG nova.network.neutron [None req-5fbf0568-2a85-4e3c-9b63-7d91ecc77593 tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] [instance: 2719569b-8572-4199-8158-7bb367d17dc5] Building network info cache for instance {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 853.392614] env[61923]: DEBUG oslo_concurrency.lockutils [None req-513daa64-2b05-4c0d-94e9-55b95de63eec tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Lock "59198f4d-4dde-4eaf-9f6c-a962cbe53c6e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 87.549s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 853.414341] env[61923]: DEBUG nova.scheduler.client.report [None req-7d4ecc17-51d2-4753-8cb6-fd83e220b388 tempest-ServersNegativeTestMultiTenantJSON-1497120864 tempest-ServersNegativeTestMultiTenantJSON-1497120864-project-member] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 853.497634] env[61923]: DEBUG oslo_vmware.api [-] Task: {'id': task-1377597, 'name': CreateVM_Task} progress is 25%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.505309] env[61923]: DEBUG oslo_vmware.api [None req-1214e428-bf00-4cbe-9e1e-7c8be1612c36 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] Task: {'id': task-1377598, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.062761} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 853.505601] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-1214e428-bf00-4cbe-9e1e-7c8be1612c36 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] [instance: af3c317a-4007-4cea-a060-1e7dde5ce49e] Extended root virtual disk {{(pid=61923) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 853.506349] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0cba590a-eeb5-4371-8f24-eb3aae1db226 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.530593] env[61923]: DEBUG nova.virt.vmwareapi.volumeops [None req-1214e428-bf00-4cbe-9e1e-7c8be1612c36 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] [instance: af3c317a-4007-4cea-a060-1e7dde5ce49e] Reconfiguring VM instance instance-00000047 to attach disk [datastore1] af3c317a-4007-4cea-a060-1e7dde5ce49e/af3c317a-4007-4cea-a060-1e7dde5ce49e.vmdk or device None with type sparse {{(pid=61923) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 853.533161] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b4968998-c055-480e-906b-fa7b03a9a056 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.554073] env[61923]: DEBUG oslo_vmware.api [None req-1214e428-bf00-4cbe-9e1e-7c8be1612c36 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] Waiting for the task: (returnval){ [ 853.554073] env[61923]: value = "task-1377599" [ 853.554073] env[61923]: _type = "Task" [ 853.554073] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 853.563978] env[61923]: DEBUG oslo_vmware.api [None req-1214e428-bf00-4cbe-9e1e-7c8be1612c36 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] Task: {'id': task-1377599, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.638042] env[61923]: DEBUG nova.compute.manager [req-f26f629d-d194-4189-a701-987ac4d71395 req-293a8171-4f20-4725-b080-35b16191a57e service nova] [instance: 87ae37dc-d0d0-4b76-afdd-0ba3e8f6ce0b] Received event network-changed-b0e2ee14-b051-47bc-9064-50f09916c0cb {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 853.638315] env[61923]: DEBUG nova.compute.manager [req-f26f629d-d194-4189-a701-987ac4d71395 req-293a8171-4f20-4725-b080-35b16191a57e service nova] [instance: 87ae37dc-d0d0-4b76-afdd-0ba3e8f6ce0b] Refreshing instance network info cache due to event network-changed-b0e2ee14-b051-47bc-9064-50f09916c0cb. {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 853.638597] env[61923]: DEBUG oslo_concurrency.lockutils [req-f26f629d-d194-4189-a701-987ac4d71395 req-293a8171-4f20-4725-b080-35b16191a57e service nova] Acquiring lock "refresh_cache-87ae37dc-d0d0-4b76-afdd-0ba3e8f6ce0b" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 853.638796] env[61923]: DEBUG oslo_concurrency.lockutils [req-f26f629d-d194-4189-a701-987ac4d71395 req-293a8171-4f20-4725-b080-35b16191a57e service nova] Acquired lock "refresh_cache-87ae37dc-d0d0-4b76-afdd-0ba3e8f6ce0b" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 853.639065] env[61923]: DEBUG nova.network.neutron [req-f26f629d-d194-4189-a701-987ac4d71395 req-293a8171-4f20-4725-b080-35b16191a57e service nova] [instance: 87ae37dc-d0d0-4b76-afdd-0ba3e8f6ce0b] Refreshing network info cache for port b0e2ee14-b051-47bc-9064-50f09916c0cb {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 853.920185] env[61923]: DEBUG oslo_concurrency.lockutils [None req-7d4ecc17-51d2-4753-8cb6-fd83e220b388 tempest-ServersNegativeTestMultiTenantJSON-1497120864 tempest-ServersNegativeTestMultiTenantJSON-1497120864-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.791s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 853.923663] env[61923]: DEBUG nova.network.neutron [None req-5fbf0568-2a85-4e3c-9b63-7d91ecc77593 tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] [instance: 2719569b-8572-4199-8158-7bb367d17dc5] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 853.925868] env[61923]: DEBUG oslo_concurrency.lockutils [None req-fc756c76-ebec-403f-b8ba-39ff93058631 tempest-ServerShowV257Test-986496257 tempest-ServerShowV257Test-986496257-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 9.939s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 853.926089] env[61923]: DEBUG nova.objects.instance [None req-fc756c76-ebec-403f-b8ba-39ff93058631 tempest-ServerShowV257Test-986496257 tempest-ServerShowV257Test-986496257-project-member] [instance: b8c60af7-3561-468c-a9a5-1c645080fc69] Trying to apply a migration context that does not seem to be set for this instance {{(pid=61923) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 853.958185] env[61923]: INFO nova.scheduler.client.report [None req-7d4ecc17-51d2-4753-8cb6-fd83e220b388 tempest-ServersNegativeTestMultiTenantJSON-1497120864 tempest-ServersNegativeTestMultiTenantJSON-1497120864-project-member] Deleted allocations for instance 1d1b1dc4-c701-42b2-beba-344d622aef41 [ 853.998795] env[61923]: DEBUG oslo_vmware.api [-] Task: {'id': task-1377597, 'name': CreateVM_Task, 'duration_secs': 0.763166} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 854.001621] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 87ae37dc-d0d0-4b76-afdd-0ba3e8f6ce0b] Created VM on the ESX host {{(pid=61923) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 854.004525] env[61923]: DEBUG oslo_concurrency.lockutils [None req-52eadf86-cc45-4cae-af2a-645ac3cb8658 tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 854.004691] env[61923]: DEBUG oslo_concurrency.lockutils [None req-52eadf86-cc45-4cae-af2a-645ac3cb8658 tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 854.005021] env[61923]: DEBUG oslo_concurrency.lockutils [None req-52eadf86-cc45-4cae-af2a-645ac3cb8658 tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 854.005561] env[61923]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6962f903-2765-43fb-b4b6-5e6c49ed1ece {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.010243] env[61923]: DEBUG oslo_vmware.api [None req-52eadf86-cc45-4cae-af2a-645ac3cb8658 tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Waiting for the task: (returnval){ [ 854.010243] env[61923]: value = "session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52308212-5d76-7667-bc45-a7102eeea318" [ 854.010243] env[61923]: _type = "Task" [ 854.010243] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 854.020266] env[61923]: DEBUG oslo_vmware.api [None req-52eadf86-cc45-4cae-af2a-645ac3cb8658 tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52308212-5d76-7667-bc45-a7102eeea318, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.066037] env[61923]: DEBUG oslo_vmware.api [None req-1214e428-bf00-4cbe-9e1e-7c8be1612c36 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] Task: {'id': task-1377599, 'name': ReconfigVM_Task, 'duration_secs': 0.262358} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 854.069967] env[61923]: DEBUG nova.virt.vmwareapi.volumeops [None req-1214e428-bf00-4cbe-9e1e-7c8be1612c36 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] [instance: af3c317a-4007-4cea-a060-1e7dde5ce49e] Reconfigured VM instance instance-00000047 to attach disk [datastore1] af3c317a-4007-4cea-a060-1e7dde5ce49e/af3c317a-4007-4cea-a060-1e7dde5ce49e.vmdk or device None with type sparse {{(pid=61923) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 854.073420] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-d7632fb2-8077-482e-af9d-830d71df24eb {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.083150] env[61923]: DEBUG oslo_vmware.api [None req-1214e428-bf00-4cbe-9e1e-7c8be1612c36 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] Waiting for the task: (returnval){ [ 854.083150] env[61923]: value = "task-1377600" [ 854.083150] env[61923]: _type = "Task" [ 854.083150] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 854.093539] env[61923]: DEBUG oslo_vmware.api [None req-1214e428-bf00-4cbe-9e1e-7c8be1612c36 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] Task: {'id': task-1377600, 'name': Rename_Task} progress is 5%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.150369] env[61923]: DEBUG nova.compute.manager [None req-cdbe0145-cd3c-4551-af87-0c7f8b53f26c tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] [instance: 444affa5-a7ed-4a17-9015-9fd5724aab64] Start spawning the instance on the hypervisor. {{(pid=61923) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 854.185932] env[61923]: DEBUG nova.compute.manager [req-7b949edc-987b-4162-b29e-f0bebfafb328 req-453a5b90-b8a8-42a5-ac5a-e452fb20d33b service nova] [instance: fc4fdf89-4615-4586-9dcd-ea90dc258361] Received event network-changed-0738a978-9439-4919-8894-4b264480593e {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 854.186180] env[61923]: DEBUG nova.compute.manager [req-7b949edc-987b-4162-b29e-f0bebfafb328 req-453a5b90-b8a8-42a5-ac5a-e452fb20d33b service nova] [instance: fc4fdf89-4615-4586-9dcd-ea90dc258361] Refreshing instance network info cache due to event network-changed-0738a978-9439-4919-8894-4b264480593e. {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 854.186332] env[61923]: DEBUG oslo_concurrency.lockutils [req-7b949edc-987b-4162-b29e-f0bebfafb328 req-453a5b90-b8a8-42a5-ac5a-e452fb20d33b service nova] Acquiring lock "refresh_cache-fc4fdf89-4615-4586-9dcd-ea90dc258361" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 854.188914] env[61923]: DEBUG nova.virt.hardware [None req-cdbe0145-cd3c-4551-af87-0c7f8b53f26c tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-29T20:07:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-29T20:07:35Z,direct_url=,disk_format='vmdk',id=0f153f63-ae0a-45b1-b7f5-7f9b673c947f,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='4e7b5e3b3c3443c8bd5c70f8a15739f5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-29T20:07:36Z,virtual_size=,visibility=), allow threads: False {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 854.189651] env[61923]: DEBUG nova.virt.hardware [None req-cdbe0145-cd3c-4551-af87-0c7f8b53f26c tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Flavor limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 854.189830] env[61923]: DEBUG nova.virt.hardware [None req-cdbe0145-cd3c-4551-af87-0c7f8b53f26c tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Image limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 854.190075] env[61923]: DEBUG nova.virt.hardware [None req-cdbe0145-cd3c-4551-af87-0c7f8b53f26c tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Flavor pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 854.190233] env[61923]: DEBUG nova.virt.hardware [None req-cdbe0145-cd3c-4551-af87-0c7f8b53f26c tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Image pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 854.190377] env[61923]: DEBUG nova.virt.hardware [None req-cdbe0145-cd3c-4551-af87-0c7f8b53f26c tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 854.190583] env[61923]: DEBUG nova.virt.hardware [None req-cdbe0145-cd3c-4551-af87-0c7f8b53f26c tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 854.190739] env[61923]: DEBUG nova.virt.hardware [None req-cdbe0145-cd3c-4551-af87-0c7f8b53f26c tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 854.190927] env[61923]: DEBUG nova.virt.hardware [None req-cdbe0145-cd3c-4551-af87-0c7f8b53f26c tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Got 1 possible topologies {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 854.191100] env[61923]: DEBUG nova.virt.hardware [None req-cdbe0145-cd3c-4551-af87-0c7f8b53f26c tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 854.191273] env[61923]: DEBUG nova.virt.hardware [None req-cdbe0145-cd3c-4551-af87-0c7f8b53f26c tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 854.192731] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-abe1f66e-389b-493a-be25-b2e1fbc14d8e {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.201836] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d558aa5-9f6a-4b3b-add3-4b30d11231b9 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.252271] env[61923]: DEBUG nova.network.neutron [None req-5fbf0568-2a85-4e3c-9b63-7d91ecc77593 tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] [instance: 2719569b-8572-4199-8158-7bb367d17dc5] Updating instance_info_cache with network_info: [{"id": "e092aadb-6904-43bc-b14c-0b1e010c4b2f", "address": "fa:16:3e:d5:76:0f", "network": {"id": "1f010d1c-1c6c-4ce2-88c8-57af627c8e07", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.234", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "4e7b5e3b3c3443c8bd5c70f8a15739f5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "db68bd64-5b56-49af-a075-13dcf85cb2e0", "external-id": "nsx-vlan-transportzone-590", "segmentation_id": 590, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape092aadb-69", "ovs_interfaceid": "e092aadb-6904-43bc-b14c-0b1e010c4b2f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 854.285244] env[61923]: DEBUG nova.network.neutron [None req-80740a2c-9779-4afc-be30-c4469cd3724d tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] [instance: fc4fdf89-4615-4586-9dcd-ea90dc258361] Updating instance_info_cache with network_info: [{"id": "8781f580-cc8d-4677-b6b3-2e71a0b907bc", "address": "fa:16:3e:74:53:67", "network": {"id": "982ad408-258d-4c40-a80e-fb1c8b012c16", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1379323635", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.85", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c304b5b5d4df485fbda55a570494a62f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "32faf59b-014c-4f1f-8331-40df95bf741f", "external-id": "nsx-vlan-transportzone-996", "segmentation_id": 996, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8781f580-cc", "ovs_interfaceid": "8781f580-cc8d-4677-b6b3-2e71a0b907bc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "a36809ea-2e4a-48ff-ac61-305da772b282", "address": "fa:16:3e:78:7e:ef", "network": {"id": "3adb8073-ba83-4101-a1fa-aaef57320dbf", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-563610619", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.15", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "c304b5b5d4df485fbda55a570494a62f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a485857d-7086-4dcf-9d65-d0dcd177fcb0", "external-id": "nsx-vlan-transportzone-232", "segmentation_id": 232, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa36809ea-2e", "ovs_interfaceid": "a36809ea-2e4a-48ff-ac61-305da772b282", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "0738a978-9439-4919-8894-4b264480593e", "address": "fa:16:3e:37:4a:a2", "network": {"id": "982ad408-258d-4c40-a80e-fb1c8b012c16", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1379323635", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.167", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c304b5b5d4df485fbda55a570494a62f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "32faf59b-014c-4f1f-8331-40df95bf741f", "external-id": "nsx-vlan-transportzone-996", "segmentation_id": 996, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0738a978-94", "ovs_interfaceid": "0738a978-9439-4919-8894-4b264480593e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 854.415390] env[61923]: DEBUG nova.network.neutron [-] [instance: 73692517-1816-4e9b-ab2d-8265c683e83d] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 854.425535] env[61923]: DEBUG nova.network.neutron [req-f26f629d-d194-4189-a701-987ac4d71395 req-293a8171-4f20-4725-b080-35b16191a57e service nova] [instance: 87ae37dc-d0d0-4b76-afdd-0ba3e8f6ce0b] Updated VIF entry in instance network info cache for port b0e2ee14-b051-47bc-9064-50f09916c0cb. {{(pid=61923) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 854.426464] env[61923]: DEBUG nova.network.neutron [req-f26f629d-d194-4189-a701-987ac4d71395 req-293a8171-4f20-4725-b080-35b16191a57e service nova] [instance: 87ae37dc-d0d0-4b76-afdd-0ba3e8f6ce0b] Updating instance_info_cache with network_info: [{"id": "b0e2ee14-b051-47bc-9064-50f09916c0cb", "address": "fa:16:3e:26:d6:71", "network": {"id": "1f010d1c-1c6c-4ce2-88c8-57af627c8e07", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.147", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "4e7b5e3b3c3443c8bd5c70f8a15739f5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "db68bd64-5b56-49af-a075-13dcf85cb2e0", "external-id": "nsx-vlan-transportzone-590", "segmentation_id": 590, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb0e2ee14-b0", "ovs_interfaceid": "b0e2ee14-b051-47bc-9064-50f09916c0cb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 854.472227] env[61923]: DEBUG oslo_concurrency.lockutils [None req-7d4ecc17-51d2-4753-8cb6-fd83e220b388 tempest-ServersNegativeTestMultiTenantJSON-1497120864 tempest-ServersNegativeTestMultiTenantJSON-1497120864-project-member] Lock "1d1b1dc4-c701-42b2-beba-344d622aef41" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 14.304s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 854.481076] env[61923]: DEBUG nova.network.neutron [None req-cdbe0145-cd3c-4551-af87-0c7f8b53f26c tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] [instance: 444affa5-a7ed-4a17-9015-9fd5724aab64] Successfully updated port: 7b131b11-927e-4c0c-98f2-714b63683d77 {{(pid=61923) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 854.521979] env[61923]: DEBUG oslo_vmware.api [None req-52eadf86-cc45-4cae-af2a-645ac3cb8658 tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52308212-5d76-7667-bc45-a7102eeea318, 'name': SearchDatastore_Task, 'duration_secs': 0.010023} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 854.521979] env[61923]: DEBUG oslo_concurrency.lockutils [None req-52eadf86-cc45-4cae-af2a-645ac3cb8658 tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 854.521979] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-52eadf86-cc45-4cae-af2a-645ac3cb8658 tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] [instance: 87ae37dc-d0d0-4b76-afdd-0ba3e8f6ce0b] Processing image 0f153f63-ae0a-45b1-b7f5-7f9b673c947f {{(pid=61923) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 854.522378] env[61923]: DEBUG oslo_concurrency.lockutils [None req-52eadf86-cc45-4cae-af2a-645ac3cb8658 tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 854.522378] env[61923]: DEBUG oslo_concurrency.lockutils [None req-52eadf86-cc45-4cae-af2a-645ac3cb8658 tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 854.522378] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-52eadf86-cc45-4cae-af2a-645ac3cb8658 tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61923) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 854.522636] env[61923]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c78b2c0d-1402-4bac-9926-b63b6457df4d {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.533021] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-52eadf86-cc45-4cae-af2a-645ac3cb8658 tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61923) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 854.533021] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-52eadf86-cc45-4cae-af2a-645ac3cb8658 tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61923) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 854.533354] env[61923]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3a89c195-bcd8-464b-82d8-d2d4043cec0c {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.538639] env[61923]: DEBUG oslo_vmware.api [None req-52eadf86-cc45-4cae-af2a-645ac3cb8658 tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Waiting for the task: (returnval){ [ 854.538639] env[61923]: value = "session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52a7061d-8504-2204-f731-662dd7c60a9d" [ 854.538639] env[61923]: _type = "Task" [ 854.538639] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 854.546086] env[61923]: DEBUG oslo_vmware.api [None req-52eadf86-cc45-4cae-af2a-645ac3cb8658 tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52a7061d-8504-2204-f731-662dd7c60a9d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.592681] env[61923]: DEBUG oslo_vmware.api [None req-1214e428-bf00-4cbe-9e1e-7c8be1612c36 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] Task: {'id': task-1377600, 'name': Rename_Task} progress is 99%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.755341] env[61923]: DEBUG oslo_concurrency.lockutils [None req-5fbf0568-2a85-4e3c-9b63-7d91ecc77593 tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Releasing lock "refresh_cache-2719569b-8572-4199-8158-7bb367d17dc5" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 854.755684] env[61923]: DEBUG nova.compute.manager [None req-5fbf0568-2a85-4e3c-9b63-7d91ecc77593 tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] [instance: 2719569b-8572-4199-8158-7bb367d17dc5] Instance network_info: |[{"id": "e092aadb-6904-43bc-b14c-0b1e010c4b2f", "address": "fa:16:3e:d5:76:0f", "network": {"id": "1f010d1c-1c6c-4ce2-88c8-57af627c8e07", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.234", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "4e7b5e3b3c3443c8bd5c70f8a15739f5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "db68bd64-5b56-49af-a075-13dcf85cb2e0", "external-id": "nsx-vlan-transportzone-590", "segmentation_id": 590, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape092aadb-69", "ovs_interfaceid": "e092aadb-6904-43bc-b14c-0b1e010c4b2f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61923) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 854.756149] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-5fbf0568-2a85-4e3c-9b63-7d91ecc77593 tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] [instance: 2719569b-8572-4199-8158-7bb367d17dc5] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:d5:76:0f', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'db68bd64-5b56-49af-a075-13dcf85cb2e0', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'e092aadb-6904-43bc-b14c-0b1e010c4b2f', 'vif_model': 'vmxnet3'}] {{(pid=61923) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 854.764449] env[61923]: DEBUG oslo.service.loopingcall [None req-5fbf0568-2a85-4e3c-9b63-7d91ecc77593 tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61923) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 854.764723] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2719569b-8572-4199-8158-7bb367d17dc5] Creating VM on the ESX host {{(pid=61923) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 854.765364] env[61923]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-de1e0e43-baf0-49b5-a431-eef91222247b {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.788565] env[61923]: DEBUG oslo_concurrency.lockutils [None req-80740a2c-9779-4afc-be30-c4469cd3724d tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] Releasing lock "refresh_cache-fc4fdf89-4615-4586-9dcd-ea90dc258361" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 854.788953] env[61923]: DEBUG nova.compute.manager [None req-80740a2c-9779-4afc-be30-c4469cd3724d tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] [instance: fc4fdf89-4615-4586-9dcd-ea90dc258361] Instance network_info: |[{"id": "8781f580-cc8d-4677-b6b3-2e71a0b907bc", "address": "fa:16:3e:74:53:67", "network": {"id": "982ad408-258d-4c40-a80e-fb1c8b012c16", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1379323635", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.85", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c304b5b5d4df485fbda55a570494a62f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "32faf59b-014c-4f1f-8331-40df95bf741f", "external-id": "nsx-vlan-transportzone-996", "segmentation_id": 996, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8781f580-cc", "ovs_interfaceid": "8781f580-cc8d-4677-b6b3-2e71a0b907bc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "a36809ea-2e4a-48ff-ac61-305da772b282", "address": "fa:16:3e:78:7e:ef", "network": {"id": "3adb8073-ba83-4101-a1fa-aaef57320dbf", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-563610619", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.15", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "c304b5b5d4df485fbda55a570494a62f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a485857d-7086-4dcf-9d65-d0dcd177fcb0", "external-id": "nsx-vlan-transportzone-232", "segmentation_id": 232, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa36809ea-2e", "ovs_interfaceid": "a36809ea-2e4a-48ff-ac61-305da772b282", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "0738a978-9439-4919-8894-4b264480593e", "address": "fa:16:3e:37:4a:a2", "network": {"id": "982ad408-258d-4c40-a80e-fb1c8b012c16", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1379323635", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.167", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c304b5b5d4df485fbda55a570494a62f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "32faf59b-014c-4f1f-8331-40df95bf741f", "external-id": "nsx-vlan-transportzone-996", "segmentation_id": 996, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0738a978-94", "ovs_interfaceid": "0738a978-9439-4919-8894-4b264480593e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61923) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 854.789312] env[61923]: DEBUG oslo_concurrency.lockutils [req-7b949edc-987b-4162-b29e-f0bebfafb328 req-453a5b90-b8a8-42a5-ac5a-e452fb20d33b service nova] Acquired lock "refresh_cache-fc4fdf89-4615-4586-9dcd-ea90dc258361" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 854.789507] env[61923]: DEBUG nova.network.neutron [req-7b949edc-987b-4162-b29e-f0bebfafb328 req-453a5b90-b8a8-42a5-ac5a-e452fb20d33b service nova] [instance: fc4fdf89-4615-4586-9dcd-ea90dc258361] Refreshing network info cache for port 0738a978-9439-4919-8894-4b264480593e {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 854.790736] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-80740a2c-9779-4afc-be30-c4469cd3724d tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] [instance: fc4fdf89-4615-4586-9dcd-ea90dc258361] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:74:53:67', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '32faf59b-014c-4f1f-8331-40df95bf741f', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '8781f580-cc8d-4677-b6b3-2e71a0b907bc', 'vif_model': 'vmxnet3'}, {'network_name': 'br-int', 'mac_address': 'fa:16:3e:78:7e:ef', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'a485857d-7086-4dcf-9d65-d0dcd177fcb0', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'a36809ea-2e4a-48ff-ac61-305da772b282', 'vif_model': 'vmxnet3'}, {'network_name': 'br-int', 'mac_address': 'fa:16:3e:37:4a:a2', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '32faf59b-014c-4f1f-8331-40df95bf741f', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '0738a978-9439-4919-8894-4b264480593e', 'vif_model': 'vmxnet3'}] {{(pid=61923) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 854.801103] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-80740a2c-9779-4afc-be30-c4469cd3724d tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] Creating folder: Project (c304b5b5d4df485fbda55a570494a62f). Parent ref: group-v292629. {{(pid=61923) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 854.805258] env[61923]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-81f29df2-b997-45f2-9a8d-abe9e7a7de4c {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.806970] env[61923]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 854.806970] env[61923]: value = "task-1377601" [ 854.806970] env[61923]: _type = "Task" [ 854.806970] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 854.818101] env[61923]: DEBUG oslo_vmware.api [-] Task: {'id': task-1377601, 'name': CreateVM_Task} progress is 10%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.819369] env[61923]: INFO nova.virt.vmwareapi.vm_util [None req-80740a2c-9779-4afc-be30-c4469cd3724d tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] Created folder: Project (c304b5b5d4df485fbda55a570494a62f) in parent group-v292629. [ 854.819546] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-80740a2c-9779-4afc-be30-c4469cd3724d tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] Creating folder: Instances. Parent ref: group-v292679. {{(pid=61923) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 854.819761] env[61923]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-506ebf9b-91fe-4838-9232-4e409640ef34 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.830551] env[61923]: INFO nova.virt.vmwareapi.vm_util [None req-80740a2c-9779-4afc-be30-c4469cd3724d tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] Created folder: Instances in parent group-v292679. [ 854.830803] env[61923]: DEBUG oslo.service.loopingcall [None req-80740a2c-9779-4afc-be30-c4469cd3724d tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61923) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 854.831427] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fc4fdf89-4615-4586-9dcd-ea90dc258361] Creating VM on the ESX host {{(pid=61923) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 854.831427] env[61923]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-367c7f4f-efac-4111-9324-026003ad99fd {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.857097] env[61923]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 854.857097] env[61923]: value = "task-1377604" [ 854.857097] env[61923]: _type = "Task" [ 854.857097] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 854.866914] env[61923]: DEBUG oslo_vmware.api [-] Task: {'id': task-1377604, 'name': CreateVM_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.918100] env[61923]: INFO nova.compute.manager [-] [instance: 73692517-1816-4e9b-ab2d-8265c683e83d] Took 1.59 seconds to deallocate network for instance. [ 854.928212] env[61923]: DEBUG oslo_concurrency.lockutils [req-f26f629d-d194-4189-a701-987ac4d71395 req-293a8171-4f20-4725-b080-35b16191a57e service nova] Releasing lock "refresh_cache-87ae37dc-d0d0-4b76-afdd-0ba3e8f6ce0b" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 854.928502] env[61923]: DEBUG nova.compute.manager [req-f26f629d-d194-4189-a701-987ac4d71395 req-293a8171-4f20-4725-b080-35b16191a57e service nova] [instance: 2719569b-8572-4199-8158-7bb367d17dc5] Received event network-vif-plugged-e092aadb-6904-43bc-b14c-0b1e010c4b2f {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 854.928702] env[61923]: DEBUG oslo_concurrency.lockutils [req-f26f629d-d194-4189-a701-987ac4d71395 req-293a8171-4f20-4725-b080-35b16191a57e service nova] Acquiring lock "2719569b-8572-4199-8158-7bb367d17dc5-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 854.929169] env[61923]: DEBUG oslo_concurrency.lockutils [req-f26f629d-d194-4189-a701-987ac4d71395 req-293a8171-4f20-4725-b080-35b16191a57e service nova] Lock "2719569b-8572-4199-8158-7bb367d17dc5-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 854.929169] env[61923]: DEBUG oslo_concurrency.lockutils [req-f26f629d-d194-4189-a701-987ac4d71395 req-293a8171-4f20-4725-b080-35b16191a57e service nova] Lock "2719569b-8572-4199-8158-7bb367d17dc5-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 854.929367] env[61923]: DEBUG nova.compute.manager [req-f26f629d-d194-4189-a701-987ac4d71395 req-293a8171-4f20-4725-b080-35b16191a57e service nova] [instance: 2719569b-8572-4199-8158-7bb367d17dc5] No waiting events found dispatching network-vif-plugged-e092aadb-6904-43bc-b14c-0b1e010c4b2f {{(pid=61923) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 854.929569] env[61923]: WARNING nova.compute.manager [req-f26f629d-d194-4189-a701-987ac4d71395 req-293a8171-4f20-4725-b080-35b16191a57e service nova] [instance: 2719569b-8572-4199-8158-7bb367d17dc5] Received unexpected event network-vif-plugged-e092aadb-6904-43bc-b14c-0b1e010c4b2f for instance with vm_state building and task_state spawning. [ 854.929738] env[61923]: DEBUG nova.compute.manager [req-f26f629d-d194-4189-a701-987ac4d71395 req-293a8171-4f20-4725-b080-35b16191a57e service nova] [instance: 2719569b-8572-4199-8158-7bb367d17dc5] Received event network-changed-e092aadb-6904-43bc-b14c-0b1e010c4b2f {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 854.929896] env[61923]: DEBUG nova.compute.manager [req-f26f629d-d194-4189-a701-987ac4d71395 req-293a8171-4f20-4725-b080-35b16191a57e service nova] [instance: 2719569b-8572-4199-8158-7bb367d17dc5] Refreshing instance network info cache due to event network-changed-e092aadb-6904-43bc-b14c-0b1e010c4b2f. {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 854.930437] env[61923]: DEBUG oslo_concurrency.lockutils [req-f26f629d-d194-4189-a701-987ac4d71395 req-293a8171-4f20-4725-b080-35b16191a57e service nova] Acquiring lock "refresh_cache-2719569b-8572-4199-8158-7bb367d17dc5" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 854.930437] env[61923]: DEBUG oslo_concurrency.lockutils [req-f26f629d-d194-4189-a701-987ac4d71395 req-293a8171-4f20-4725-b080-35b16191a57e service nova] Acquired lock "refresh_cache-2719569b-8572-4199-8158-7bb367d17dc5" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 854.930437] env[61923]: DEBUG nova.network.neutron [req-f26f629d-d194-4189-a701-987ac4d71395 req-293a8171-4f20-4725-b080-35b16191a57e service nova] [instance: 2719569b-8572-4199-8158-7bb367d17dc5] Refreshing network info cache for port e092aadb-6904-43bc-b14c-0b1e010c4b2f {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 854.935619] env[61923]: DEBUG oslo_concurrency.lockutils [None req-fc756c76-ebec-403f-b8ba-39ff93058631 tempest-ServerShowV257Test-986496257 tempest-ServerShowV257Test-986496257-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.010s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 854.940975] env[61923]: DEBUG oslo_concurrency.lockutils [None req-cdbe0145-cd3c-4551-af87-0c7f8b53f26c tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 10.453s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 854.942488] env[61923]: INFO nova.compute.claims [None req-cdbe0145-cd3c-4551-af87-0c7f8b53f26c tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] [instance: e413c45d-cd89-44d4-9102-3d188907e7bb] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 854.984336] env[61923]: DEBUG oslo_concurrency.lockutils [None req-cdbe0145-cd3c-4551-af87-0c7f8b53f26c tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Acquiring lock "refresh_cache-444affa5-a7ed-4a17-9015-9fd5724aab64" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 854.984784] env[61923]: DEBUG oslo_concurrency.lockutils [None req-cdbe0145-cd3c-4551-af87-0c7f8b53f26c tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Acquired lock "refresh_cache-444affa5-a7ed-4a17-9015-9fd5724aab64" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 854.984986] env[61923]: DEBUG nova.network.neutron [None req-cdbe0145-cd3c-4551-af87-0c7f8b53f26c tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] [instance: 444affa5-a7ed-4a17-9015-9fd5724aab64] Building network info cache for instance {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 855.049620] env[61923]: DEBUG oslo_vmware.api [None req-52eadf86-cc45-4cae-af2a-645ac3cb8658 tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52a7061d-8504-2204-f731-662dd7c60a9d, 'name': SearchDatastore_Task, 'duration_secs': 0.022185} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 855.050493] env[61923]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9e430189-b0a9-486d-bf61-0445c409b140 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.056247] env[61923]: DEBUG oslo_vmware.api [None req-52eadf86-cc45-4cae-af2a-645ac3cb8658 tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Waiting for the task: (returnval){ [ 855.056247] env[61923]: value = "session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52ae9144-f517-4e75-a963-67643080ed3a" [ 855.056247] env[61923]: _type = "Task" [ 855.056247] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 855.066288] env[61923]: DEBUG oslo_vmware.api [None req-52eadf86-cc45-4cae-af2a-645ac3cb8658 tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52ae9144-f517-4e75-a963-67643080ed3a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.067298] env[61923]: DEBUG nova.network.neutron [req-7b949edc-987b-4162-b29e-f0bebfafb328 req-453a5b90-b8a8-42a5-ac5a-e452fb20d33b service nova] [instance: fc4fdf89-4615-4586-9dcd-ea90dc258361] Updated VIF entry in instance network info cache for port 0738a978-9439-4919-8894-4b264480593e. {{(pid=61923) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 855.067805] env[61923]: DEBUG nova.network.neutron [req-7b949edc-987b-4162-b29e-f0bebfafb328 req-453a5b90-b8a8-42a5-ac5a-e452fb20d33b service nova] [instance: fc4fdf89-4615-4586-9dcd-ea90dc258361] Updating instance_info_cache with network_info: [{"id": "8781f580-cc8d-4677-b6b3-2e71a0b907bc", "address": "fa:16:3e:74:53:67", "network": {"id": "982ad408-258d-4c40-a80e-fb1c8b012c16", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1379323635", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.85", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c304b5b5d4df485fbda55a570494a62f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "32faf59b-014c-4f1f-8331-40df95bf741f", "external-id": "nsx-vlan-transportzone-996", "segmentation_id": 996, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8781f580-cc", "ovs_interfaceid": "8781f580-cc8d-4677-b6b3-2e71a0b907bc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "a36809ea-2e4a-48ff-ac61-305da772b282", "address": "fa:16:3e:78:7e:ef", "network": {"id": "3adb8073-ba83-4101-a1fa-aaef57320dbf", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-563610619", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.15", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "c304b5b5d4df485fbda55a570494a62f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a485857d-7086-4dcf-9d65-d0dcd177fcb0", "external-id": "nsx-vlan-transportzone-232", "segmentation_id": 232, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa36809ea-2e", "ovs_interfaceid": "a36809ea-2e4a-48ff-ac61-305da772b282", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "0738a978-9439-4919-8894-4b264480593e", "address": "fa:16:3e:37:4a:a2", "network": {"id": "982ad408-258d-4c40-a80e-fb1c8b012c16", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1379323635", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.167", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c304b5b5d4df485fbda55a570494a62f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "32faf59b-014c-4f1f-8331-40df95bf741f", "external-id": "nsx-vlan-transportzone-996", "segmentation_id": 996, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0738a978-94", "ovs_interfaceid": "0738a978-9439-4919-8894-4b264480593e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 855.095247] env[61923]: DEBUG oslo_vmware.api [None req-1214e428-bf00-4cbe-9e1e-7c8be1612c36 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] Task: {'id': task-1377600, 'name': Rename_Task} progress is 99%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.319811] env[61923]: DEBUG oslo_vmware.api [-] Task: {'id': task-1377601, 'name': CreateVM_Task, 'duration_secs': 0.32335} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 855.319991] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2719569b-8572-4199-8158-7bb367d17dc5] Created VM on the ESX host {{(pid=61923) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 855.320713] env[61923]: DEBUG oslo_concurrency.lockutils [None req-5fbf0568-2a85-4e3c-9b63-7d91ecc77593 tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 855.320884] env[61923]: DEBUG oslo_concurrency.lockutils [None req-5fbf0568-2a85-4e3c-9b63-7d91ecc77593 tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 855.321236] env[61923]: DEBUG oslo_concurrency.lockutils [None req-5fbf0568-2a85-4e3c-9b63-7d91ecc77593 tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 855.321492] env[61923]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3f223b9c-f1a0-4b29-9dd5-e4c1d492979e {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.326028] env[61923]: DEBUG oslo_vmware.api [None req-5fbf0568-2a85-4e3c-9b63-7d91ecc77593 tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Waiting for the task: (returnval){ [ 855.326028] env[61923]: value = "session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]5278a488-675d-c208-5c64-80242edeca10" [ 855.326028] env[61923]: _type = "Task" [ 855.326028] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 855.333754] env[61923]: DEBUG oslo_vmware.api [None req-5fbf0568-2a85-4e3c-9b63-7d91ecc77593 tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]5278a488-675d-c208-5c64-80242edeca10, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.365974] env[61923]: DEBUG oslo_vmware.api [-] Task: {'id': task-1377604, 'name': CreateVM_Task, 'duration_secs': 0.463026} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 855.366147] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: fc4fdf89-4615-4586-9dcd-ea90dc258361] Created VM on the ESX host {{(pid=61923) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 855.366955] env[61923]: DEBUG oslo_concurrency.lockutils [None req-80740a2c-9779-4afc-be30-c4469cd3724d tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 855.384614] env[61923]: DEBUG nova.compute.manager [None req-aa1aa995-4e31-45e6-b824-29d545be859e tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] [instance: 59198f4d-4dde-4eaf-9f6c-a962cbe53c6e] Checking state {{(pid=61923) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 855.385458] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c58c7af-a975-42b9-b5de-269792cd80f6 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.425111] env[61923]: DEBUG oslo_concurrency.lockutils [None req-b8ed9576-aa00-419c-8225-19ac3c654fd7 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 855.530935] env[61923]: DEBUG nova.network.neutron [None req-cdbe0145-cd3c-4551-af87-0c7f8b53f26c tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] [instance: 444affa5-a7ed-4a17-9015-9fd5724aab64] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 855.567981] env[61923]: DEBUG oslo_vmware.api [None req-52eadf86-cc45-4cae-af2a-645ac3cb8658 tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52ae9144-f517-4e75-a963-67643080ed3a, 'name': SearchDatastore_Task, 'duration_secs': 0.009719} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 855.570324] env[61923]: DEBUG oslo_concurrency.lockutils [None req-52eadf86-cc45-4cae-af2a-645ac3cb8658 tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 855.570545] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-52eadf86-cc45-4cae-af2a-645ac3cb8658 tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk to [datastore1] 87ae37dc-d0d0-4b76-afdd-0ba3e8f6ce0b/87ae37dc-d0d0-4b76-afdd-0ba3e8f6ce0b.vmdk {{(pid=61923) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 855.571925] env[61923]: DEBUG oslo_concurrency.lockutils [req-7b949edc-987b-4162-b29e-f0bebfafb328 req-453a5b90-b8a8-42a5-ac5a-e452fb20d33b service nova] Releasing lock "refresh_cache-fc4fdf89-4615-4586-9dcd-ea90dc258361" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 855.571925] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-2bd81983-88fd-42c2-835b-697c1616a5e5 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.577435] env[61923]: DEBUG oslo_vmware.api [None req-52eadf86-cc45-4cae-af2a-645ac3cb8658 tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Waiting for the task: (returnval){ [ 855.577435] env[61923]: value = "task-1377605" [ 855.577435] env[61923]: _type = "Task" [ 855.577435] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 855.584853] env[61923]: DEBUG oslo_vmware.api [None req-52eadf86-cc45-4cae-af2a-645ac3cb8658 tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Task: {'id': task-1377605, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.594263] env[61923]: DEBUG oslo_vmware.api [None req-1214e428-bf00-4cbe-9e1e-7c8be1612c36 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] Task: {'id': task-1377600, 'name': Rename_Task, 'duration_secs': 1.269259} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 855.594263] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-1214e428-bf00-4cbe-9e1e-7c8be1612c36 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] [instance: af3c317a-4007-4cea-a060-1e7dde5ce49e] Powering on the VM {{(pid=61923) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 855.594263] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1c293b8a-12f2-4155-a2bd-2d42e186494c {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.600144] env[61923]: DEBUG oslo_vmware.api [None req-1214e428-bf00-4cbe-9e1e-7c8be1612c36 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] Waiting for the task: (returnval){ [ 855.600144] env[61923]: value = "task-1377606" [ 855.600144] env[61923]: _type = "Task" [ 855.600144] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 855.609350] env[61923]: DEBUG oslo_vmware.api [None req-1214e428-bf00-4cbe-9e1e-7c8be1612c36 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] Task: {'id': task-1377606, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.647221] env[61923]: DEBUG nova.network.neutron [req-f26f629d-d194-4189-a701-987ac4d71395 req-293a8171-4f20-4725-b080-35b16191a57e service nova] [instance: 2719569b-8572-4199-8158-7bb367d17dc5] Updated VIF entry in instance network info cache for port e092aadb-6904-43bc-b14c-0b1e010c4b2f. {{(pid=61923) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 855.647606] env[61923]: DEBUG nova.network.neutron [req-f26f629d-d194-4189-a701-987ac4d71395 req-293a8171-4f20-4725-b080-35b16191a57e service nova] [instance: 2719569b-8572-4199-8158-7bb367d17dc5] Updating instance_info_cache with network_info: [{"id": "e092aadb-6904-43bc-b14c-0b1e010c4b2f", "address": "fa:16:3e:d5:76:0f", "network": {"id": "1f010d1c-1c6c-4ce2-88c8-57af627c8e07", "bridge": "br-int", "label": "shared", "subnets": [{"cidr": "192.168.233.0/24", "dns": [], "gateway": {"address": "192.168.233.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.233.234", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.233.2"}}], "meta": {"injected": false, "tenant_id": "4e7b5e3b3c3443c8bd5c70f8a15739f5", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "db68bd64-5b56-49af-a075-13dcf85cb2e0", "external-id": "nsx-vlan-transportzone-590", "segmentation_id": 590, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape092aadb-69", "ovs_interfaceid": "e092aadb-6904-43bc-b14c-0b1e010c4b2f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 855.701470] env[61923]: DEBUG nova.compute.manager [req-59e10802-3e9c-4dd4-a1fb-61f2ffd1605d req-bb989924-bd62-48e6-9d51-ec9783994436 service nova] [instance: 73692517-1816-4e9b-ab2d-8265c683e83d] Received event network-vif-deleted-dbed34c7-9716-4c24-8f0e-2298de6cf4b9 {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 855.843195] env[61923]: DEBUG oslo_vmware.api [None req-5fbf0568-2a85-4e3c-9b63-7d91ecc77593 tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]5278a488-675d-c208-5c64-80242edeca10, 'name': SearchDatastore_Task, 'duration_secs': 0.008058} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 855.843584] env[61923]: DEBUG oslo_concurrency.lockutils [None req-5fbf0568-2a85-4e3c-9b63-7d91ecc77593 tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 855.843853] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-5fbf0568-2a85-4e3c-9b63-7d91ecc77593 tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] [instance: 2719569b-8572-4199-8158-7bb367d17dc5] Processing image 0f153f63-ae0a-45b1-b7f5-7f9b673c947f {{(pid=61923) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 855.844143] env[61923]: DEBUG oslo_concurrency.lockutils [None req-5fbf0568-2a85-4e3c-9b63-7d91ecc77593 tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 855.844303] env[61923]: DEBUG oslo_concurrency.lockutils [None req-5fbf0568-2a85-4e3c-9b63-7d91ecc77593 tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 855.844489] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-5fbf0568-2a85-4e3c-9b63-7d91ecc77593 tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61923) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 855.844792] env[61923]: DEBUG oslo_concurrency.lockutils [None req-80740a2c-9779-4afc-be30-c4469cd3724d tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 855.845143] env[61923]: DEBUG oslo_concurrency.lockutils [None req-80740a2c-9779-4afc-be30-c4469cd3724d tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 855.845411] env[61923]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-130a43a8-11d6-486c-a8e3-2370a6c4c486 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.849923] env[61923]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7cdea08e-22ce-47bc-b260-349a16649e1b {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.855250] env[61923]: DEBUG oslo_vmware.api [None req-80740a2c-9779-4afc-be30-c4469cd3724d tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] Waiting for the task: (returnval){ [ 855.855250] env[61923]: value = "session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]526cb45c-861b-45cc-2b23-9498eff2af7f" [ 855.855250] env[61923]: _type = "Task" [ 855.855250] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 855.864577] env[61923]: DEBUG oslo_vmware.api [None req-80740a2c-9779-4afc-be30-c4469cd3724d tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]526cb45c-861b-45cc-2b23-9498eff2af7f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.866209] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-5fbf0568-2a85-4e3c-9b63-7d91ecc77593 tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61923) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 855.866590] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-5fbf0568-2a85-4e3c-9b63-7d91ecc77593 tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61923) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 855.869114] env[61923]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-beed7abf-4fdf-4c43-bb87-d59e7188544e {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.873456] env[61923]: DEBUG oslo_vmware.api [None req-5fbf0568-2a85-4e3c-9b63-7d91ecc77593 tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Waiting for the task: (returnval){ [ 855.873456] env[61923]: value = "session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52ca0bc7-f06a-e35d-c73a-beb8dc043464" [ 855.873456] env[61923]: _type = "Task" [ 855.873456] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 855.875595] env[61923]: DEBUG nova.network.neutron [None req-cdbe0145-cd3c-4551-af87-0c7f8b53f26c tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] [instance: 444affa5-a7ed-4a17-9015-9fd5724aab64] Updating instance_info_cache with network_info: [{"id": "7b131b11-927e-4c0c-98f2-714b63683d77", "address": "fa:16:3e:70:f8:a3", "network": {"id": "545a170d-946b-4e0b-b9f5-b21c2dcda19d", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-40231173-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6e260342cd1b41ef8f0ceb15b6438daa", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3753f451-fa23-4988-9361-074fb0bd3fd4", "external-id": "nsx-vlan-transportzone-440", "segmentation_id": 440, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7b131b11-92", "ovs_interfaceid": "7b131b11-927e-4c0c-98f2-714b63683d77", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 855.887642] env[61923]: DEBUG oslo_vmware.api [None req-5fbf0568-2a85-4e3c-9b63-7d91ecc77593 tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52ca0bc7-f06a-e35d-c73a-beb8dc043464, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.895974] env[61923]: INFO nova.compute.manager [None req-aa1aa995-4e31-45e6-b824-29d545be859e tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] [instance: 59198f4d-4dde-4eaf-9f6c-a962cbe53c6e] instance snapshotting [ 855.899456] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9432f02-6a6d-41d5-a21f-854ab7c8bf09 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.919858] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f38498f9-2261-48d5-9cf4-3b9e0c2d2a56 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.096281] env[61923]: DEBUG oslo_vmware.api [None req-52eadf86-cc45-4cae-af2a-645ac3cb8658 tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Task: {'id': task-1377605, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.493847} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 856.096611] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-52eadf86-cc45-4cae-af2a-645ac3cb8658 tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk to [datastore1] 87ae37dc-d0d0-4b76-afdd-0ba3e8f6ce0b/87ae37dc-d0d0-4b76-afdd-0ba3e8f6ce0b.vmdk {{(pid=61923) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 856.096919] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-52eadf86-cc45-4cae-af2a-645ac3cb8658 tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] [instance: 87ae37dc-d0d0-4b76-afdd-0ba3e8f6ce0b] Extending root virtual disk to 1048576 {{(pid=61923) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 856.097854] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-0378dca5-4bda-4ce7-b70e-14e6a65c7ed1 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.110165] env[61923]: DEBUG oslo_vmware.api [None req-52eadf86-cc45-4cae-af2a-645ac3cb8658 tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Waiting for the task: (returnval){ [ 856.110165] env[61923]: value = "task-1377607" [ 856.110165] env[61923]: _type = "Task" [ 856.110165] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 856.117915] env[61923]: DEBUG oslo_vmware.api [None req-1214e428-bf00-4cbe-9e1e-7c8be1612c36 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] Task: {'id': task-1377606, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.122714] env[61923]: DEBUG oslo_vmware.api [None req-52eadf86-cc45-4cae-af2a-645ac3cb8658 tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Task: {'id': task-1377607, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.150834] env[61923]: DEBUG oslo_concurrency.lockutils [req-f26f629d-d194-4189-a701-987ac4d71395 req-293a8171-4f20-4725-b080-35b16191a57e service nova] Releasing lock "refresh_cache-2719569b-8572-4199-8158-7bb367d17dc5" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 856.215687] env[61923]: DEBUG nova.compute.manager [req-59ae460b-8f16-4abc-b4bc-e22bc9cf173e req-1291c6e8-dfa4-42af-94c4-01622cac4a2d service nova] [instance: 444affa5-a7ed-4a17-9015-9fd5724aab64] Received event network-vif-plugged-7b131b11-927e-4c0c-98f2-714b63683d77 {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 856.216024] env[61923]: DEBUG oslo_concurrency.lockutils [req-59ae460b-8f16-4abc-b4bc-e22bc9cf173e req-1291c6e8-dfa4-42af-94c4-01622cac4a2d service nova] Acquiring lock "444affa5-a7ed-4a17-9015-9fd5724aab64-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 856.216150] env[61923]: DEBUG oslo_concurrency.lockutils [req-59ae460b-8f16-4abc-b4bc-e22bc9cf173e req-1291c6e8-dfa4-42af-94c4-01622cac4a2d service nova] Lock "444affa5-a7ed-4a17-9015-9fd5724aab64-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 856.216284] env[61923]: DEBUG oslo_concurrency.lockutils [req-59ae460b-8f16-4abc-b4bc-e22bc9cf173e req-1291c6e8-dfa4-42af-94c4-01622cac4a2d service nova] Lock "444affa5-a7ed-4a17-9015-9fd5724aab64-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 856.216496] env[61923]: DEBUG nova.compute.manager [req-59ae460b-8f16-4abc-b4bc-e22bc9cf173e req-1291c6e8-dfa4-42af-94c4-01622cac4a2d service nova] [instance: 444affa5-a7ed-4a17-9015-9fd5724aab64] No waiting events found dispatching network-vif-plugged-7b131b11-927e-4c0c-98f2-714b63683d77 {{(pid=61923) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 856.216630] env[61923]: WARNING nova.compute.manager [req-59ae460b-8f16-4abc-b4bc-e22bc9cf173e req-1291c6e8-dfa4-42af-94c4-01622cac4a2d service nova] [instance: 444affa5-a7ed-4a17-9015-9fd5724aab64] Received unexpected event network-vif-plugged-7b131b11-927e-4c0c-98f2-714b63683d77 for instance with vm_state building and task_state spawning. [ 856.216824] env[61923]: DEBUG nova.compute.manager [req-59ae460b-8f16-4abc-b4bc-e22bc9cf173e req-1291c6e8-dfa4-42af-94c4-01622cac4a2d service nova] [instance: 444affa5-a7ed-4a17-9015-9fd5724aab64] Received event network-changed-7b131b11-927e-4c0c-98f2-714b63683d77 {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 856.217039] env[61923]: DEBUG nova.compute.manager [req-59ae460b-8f16-4abc-b4bc-e22bc9cf173e req-1291c6e8-dfa4-42af-94c4-01622cac4a2d service nova] [instance: 444affa5-a7ed-4a17-9015-9fd5724aab64] Refreshing instance network info cache due to event network-changed-7b131b11-927e-4c0c-98f2-714b63683d77. {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 856.217215] env[61923]: DEBUG oslo_concurrency.lockutils [req-59ae460b-8f16-4abc-b4bc-e22bc9cf173e req-1291c6e8-dfa4-42af-94c4-01622cac4a2d service nova] Acquiring lock "refresh_cache-444affa5-a7ed-4a17-9015-9fd5724aab64" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 856.219363] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-131c7f81-c829-4b1f-8855-3b042499d54a {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.228286] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7044d721-2408-41e0-976a-8251c3508bea {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.258440] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d99f4d3-023a-4bd8-bdfc-ab2af0bb79a7 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.266115] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b17f17d-5b71-4c0b-91ab-a049322a9c0c {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.279687] env[61923]: DEBUG nova.compute.provider_tree [None req-cdbe0145-cd3c-4551-af87-0c7f8b53f26c tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 856.367817] env[61923]: DEBUG oslo_vmware.api [None req-80740a2c-9779-4afc-be30-c4469cd3724d tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]526cb45c-861b-45cc-2b23-9498eff2af7f, 'name': SearchDatastore_Task, 'duration_secs': 0.064267} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 856.368242] env[61923]: DEBUG oslo_concurrency.lockutils [None req-80740a2c-9779-4afc-be30-c4469cd3724d tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 856.368560] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-80740a2c-9779-4afc-be30-c4469cd3724d tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] [instance: fc4fdf89-4615-4586-9dcd-ea90dc258361] Processing image 0f153f63-ae0a-45b1-b7f5-7f9b673c947f {{(pid=61923) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 856.368839] env[61923]: DEBUG oslo_concurrency.lockutils [None req-80740a2c-9779-4afc-be30-c4469cd3724d tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 856.382495] env[61923]: DEBUG oslo_concurrency.lockutils [None req-cdbe0145-cd3c-4551-af87-0c7f8b53f26c tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Releasing lock "refresh_cache-444affa5-a7ed-4a17-9015-9fd5724aab64" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 856.382906] env[61923]: DEBUG nova.compute.manager [None req-cdbe0145-cd3c-4551-af87-0c7f8b53f26c tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] [instance: 444affa5-a7ed-4a17-9015-9fd5724aab64] Instance network_info: |[{"id": "7b131b11-927e-4c0c-98f2-714b63683d77", "address": "fa:16:3e:70:f8:a3", "network": {"id": "545a170d-946b-4e0b-b9f5-b21c2dcda19d", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-40231173-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6e260342cd1b41ef8f0ceb15b6438daa", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3753f451-fa23-4988-9361-074fb0bd3fd4", "external-id": "nsx-vlan-transportzone-440", "segmentation_id": 440, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7b131b11-92", "ovs_interfaceid": "7b131b11-927e-4c0c-98f2-714b63683d77", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61923) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 856.383307] env[61923]: DEBUG oslo_concurrency.lockutils [req-59ae460b-8f16-4abc-b4bc-e22bc9cf173e req-1291c6e8-dfa4-42af-94c4-01622cac4a2d service nova] Acquired lock "refresh_cache-444affa5-a7ed-4a17-9015-9fd5724aab64" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 856.383562] env[61923]: DEBUG nova.network.neutron [req-59ae460b-8f16-4abc-b4bc-e22bc9cf173e req-1291c6e8-dfa4-42af-94c4-01622cac4a2d service nova] [instance: 444affa5-a7ed-4a17-9015-9fd5724aab64] Refreshing network info cache for port 7b131b11-927e-4c0c-98f2-714b63683d77 {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 856.385237] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-cdbe0145-cd3c-4551-af87-0c7f8b53f26c tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] [instance: 444affa5-a7ed-4a17-9015-9fd5724aab64] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:70:f8:a3', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '3753f451-fa23-4988-9361-074fb0bd3fd4', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '7b131b11-927e-4c0c-98f2-714b63683d77', 'vif_model': 'vmxnet3'}] {{(pid=61923) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 856.397468] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-cdbe0145-cd3c-4551-af87-0c7f8b53f26c tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Creating folder: Project (6e260342cd1b41ef8f0ceb15b6438daa). Parent ref: group-v292629. {{(pid=61923) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 856.399455] env[61923]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-210938c4-9463-4465-a91a-11e47264075e {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.407159] env[61923]: DEBUG oslo_vmware.api [None req-5fbf0568-2a85-4e3c-9b63-7d91ecc77593 tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52ca0bc7-f06a-e35d-c73a-beb8dc043464, 'name': SearchDatastore_Task, 'duration_secs': 0.065177} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 856.408685] env[61923]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-29ec15c7-03a1-4fb3-8e56-cb01fb053dc9 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.416333] env[61923]: INFO nova.virt.vmwareapi.vm_util [None req-cdbe0145-cd3c-4551-af87-0c7f8b53f26c tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Created folder: Project (6e260342cd1b41ef8f0ceb15b6438daa) in parent group-v292629. [ 856.416623] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-cdbe0145-cd3c-4551-af87-0c7f8b53f26c tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Creating folder: Instances. Parent ref: group-v292682. {{(pid=61923) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 856.416918] env[61923]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-4b3f753f-81c9-47a6-b2f4-07e648be412a {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.421443] env[61923]: DEBUG oslo_vmware.api [None req-5fbf0568-2a85-4e3c-9b63-7d91ecc77593 tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Waiting for the task: (returnval){ [ 856.421443] env[61923]: value = "session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52ca82e8-eb97-9660-f8b9-88da2fdefd30" [ 856.421443] env[61923]: _type = "Task" [ 856.421443] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 856.427537] env[61923]: INFO nova.virt.vmwareapi.vm_util [None req-cdbe0145-cd3c-4551-af87-0c7f8b53f26c tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Created folder: Instances in parent group-v292682. [ 856.429054] env[61923]: DEBUG oslo.service.loopingcall [None req-cdbe0145-cd3c-4551-af87-0c7f8b53f26c tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61923) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 856.432244] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 444affa5-a7ed-4a17-9015-9fd5724aab64] Creating VM on the ESX host {{(pid=61923) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 856.433805] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-aa1aa995-4e31-45e6-b824-29d545be859e tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] [instance: 59198f4d-4dde-4eaf-9f6c-a962cbe53c6e] Creating Snapshot of the VM instance {{(pid=61923) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 856.434212] env[61923]: DEBUG oslo_vmware.api [None req-5fbf0568-2a85-4e3c-9b63-7d91ecc77593 tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52ca82e8-eb97-9660-f8b9-88da2fdefd30, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.434438] env[61923]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-7b823d08-114a-4fc3-a46f-05ec1f8fc815 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.461118] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-979c7ec3-b380-4a5b-9d42-62a5cc8a3941 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.469156] env[61923]: DEBUG oslo_vmware.api [None req-aa1aa995-4e31-45e6-b824-29d545be859e tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Waiting for the task: (returnval){ [ 856.469156] env[61923]: value = "task-1377610" [ 856.469156] env[61923]: _type = "Task" [ 856.469156] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 856.470753] env[61923]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 856.470753] env[61923]: value = "task-1377611" [ 856.470753] env[61923]: _type = "Task" [ 856.470753] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 856.485287] env[61923]: DEBUG oslo_vmware.api [None req-aa1aa995-4e31-45e6-b824-29d545be859e tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Task: {'id': task-1377610, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.488740] env[61923]: DEBUG oslo_vmware.api [-] Task: {'id': task-1377611, 'name': CreateVM_Task} progress is 6%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.610854] env[61923]: DEBUG oslo_vmware.api [None req-1214e428-bf00-4cbe-9e1e-7c8be1612c36 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] Task: {'id': task-1377606, 'name': PowerOnVM_Task, 'duration_secs': 0.786549} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 856.611240] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-1214e428-bf00-4cbe-9e1e-7c8be1612c36 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] [instance: af3c317a-4007-4cea-a060-1e7dde5ce49e] Powered on the VM {{(pid=61923) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 856.611435] env[61923]: INFO nova.compute.manager [None req-1214e428-bf00-4cbe-9e1e-7c8be1612c36 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] [instance: af3c317a-4007-4cea-a060-1e7dde5ce49e] Took 9.55 seconds to spawn the instance on the hypervisor. [ 856.611610] env[61923]: DEBUG nova.compute.manager [None req-1214e428-bf00-4cbe-9e1e-7c8be1612c36 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] [instance: af3c317a-4007-4cea-a060-1e7dde5ce49e] Checking state {{(pid=61923) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 856.615154] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb6b1548-a429-498d-a768-de4279b6b993 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.622043] env[61923]: DEBUG oslo_vmware.api [None req-52eadf86-cc45-4cae-af2a-645ac3cb8658 tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Task: {'id': task-1377607, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.078599} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 856.623512] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-52eadf86-cc45-4cae-af2a-645ac3cb8658 tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] [instance: 87ae37dc-d0d0-4b76-afdd-0ba3e8f6ce0b] Extended root virtual disk {{(pid=61923) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 856.628038] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91bedeca-a403-4ccf-964d-7f8e21ec5397 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.648428] env[61923]: DEBUG nova.virt.vmwareapi.volumeops [None req-52eadf86-cc45-4cae-af2a-645ac3cb8658 tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] [instance: 87ae37dc-d0d0-4b76-afdd-0ba3e8f6ce0b] Reconfiguring VM instance instance-00000048 to attach disk [datastore1] 87ae37dc-d0d0-4b76-afdd-0ba3e8f6ce0b/87ae37dc-d0d0-4b76-afdd-0ba3e8f6ce0b.vmdk or device None with type sparse {{(pid=61923) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 856.648777] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6e1d0c92-f189-4ce1-9508-b2ac9a6e6458 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.670309] env[61923]: DEBUG oslo_vmware.api [None req-52eadf86-cc45-4cae-af2a-645ac3cb8658 tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Waiting for the task: (returnval){ [ 856.670309] env[61923]: value = "task-1377612" [ 856.670309] env[61923]: _type = "Task" [ 856.670309] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 856.681652] env[61923]: DEBUG oslo_vmware.api [None req-52eadf86-cc45-4cae-af2a-645ac3cb8658 tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Task: {'id': task-1377612, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.782914] env[61923]: DEBUG nova.scheduler.client.report [None req-cdbe0145-cd3c-4551-af87-0c7f8b53f26c tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 856.931544] env[61923]: DEBUG oslo_vmware.api [None req-5fbf0568-2a85-4e3c-9b63-7d91ecc77593 tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52ca82e8-eb97-9660-f8b9-88da2fdefd30, 'name': SearchDatastore_Task, 'duration_secs': 0.032394} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 856.931813] env[61923]: DEBUG oslo_concurrency.lockutils [None req-5fbf0568-2a85-4e3c-9b63-7d91ecc77593 tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 856.932083] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-5fbf0568-2a85-4e3c-9b63-7d91ecc77593 tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk to [datastore1] 2719569b-8572-4199-8158-7bb367d17dc5/2719569b-8572-4199-8158-7bb367d17dc5.vmdk {{(pid=61923) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 856.932684] env[61923]: DEBUG oslo_concurrency.lockutils [None req-80740a2c-9779-4afc-be30-c4469cd3724d tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 856.932874] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-80740a2c-9779-4afc-be30-c4469cd3724d tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61923) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 856.933111] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-2ae5da39-936d-4e8e-942d-45e04120d45a {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.935049] env[61923]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-fc346be4-a7bc-46f9-99c2-594c04909cfa {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.942401] env[61923]: DEBUG oslo_vmware.api [None req-5fbf0568-2a85-4e3c-9b63-7d91ecc77593 tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Waiting for the task: (returnval){ [ 856.942401] env[61923]: value = "task-1377613" [ 856.942401] env[61923]: _type = "Task" [ 856.942401] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 856.946398] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-80740a2c-9779-4afc-be30-c4469cd3724d tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61923) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 856.946564] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-80740a2c-9779-4afc-be30-c4469cd3724d tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61923) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 856.947587] env[61923]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7947330e-08ce-48b7-85bb-e0ad69ebe521 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.954447] env[61923]: DEBUG oslo_vmware.api [None req-5fbf0568-2a85-4e3c-9b63-7d91ecc77593 tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Task: {'id': task-1377613, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.957193] env[61923]: DEBUG oslo_vmware.api [None req-80740a2c-9779-4afc-be30-c4469cd3724d tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] Waiting for the task: (returnval){ [ 856.957193] env[61923]: value = "session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]523c1134-fdb2-ef24-5a9f-aefc82cba757" [ 856.957193] env[61923]: _type = "Task" [ 856.957193] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 856.964355] env[61923]: DEBUG oslo_vmware.api [None req-80740a2c-9779-4afc-be30-c4469cd3724d tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]523c1134-fdb2-ef24-5a9f-aefc82cba757, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.980726] env[61923]: DEBUG oslo_vmware.api [None req-aa1aa995-4e31-45e6-b824-29d545be859e tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Task: {'id': task-1377610, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.985354] env[61923]: DEBUG oslo_vmware.api [-] Task: {'id': task-1377611, 'name': CreateVM_Task, 'duration_secs': 0.345071} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 856.985516] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 444affa5-a7ed-4a17-9015-9fd5724aab64] Created VM on the ESX host {{(pid=61923) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 856.990963] env[61923]: DEBUG oslo_concurrency.lockutils [None req-cdbe0145-cd3c-4551-af87-0c7f8b53f26c tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 856.990963] env[61923]: DEBUG oslo_concurrency.lockutils [None req-cdbe0145-cd3c-4551-af87-0c7f8b53f26c tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 856.990963] env[61923]: DEBUG oslo_concurrency.lockutils [None req-cdbe0145-cd3c-4551-af87-0c7f8b53f26c tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 856.990963] env[61923]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-463578cc-a2fe-44ca-9f16-04276bc88481 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.993891] env[61923]: DEBUG oslo_vmware.api [None req-cdbe0145-cd3c-4551-af87-0c7f8b53f26c tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Waiting for the task: (returnval){ [ 856.993891] env[61923]: value = "session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]529455a2-b56b-f24b-d51a-c1b07e65d8f4" [ 856.993891] env[61923]: _type = "Task" [ 856.993891] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 857.005856] env[61923]: DEBUG oslo_vmware.api [None req-cdbe0145-cd3c-4551-af87-0c7f8b53f26c tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]529455a2-b56b-f24b-d51a-c1b07e65d8f4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.138455] env[61923]: INFO nova.compute.manager [None req-1214e428-bf00-4cbe-9e1e-7c8be1612c36 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] [instance: af3c317a-4007-4cea-a060-1e7dde5ce49e] Took 30.42 seconds to build instance. [ 857.180302] env[61923]: DEBUG oslo_vmware.api [None req-52eadf86-cc45-4cae-af2a-645ac3cb8658 tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Task: {'id': task-1377612, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.291021] env[61923]: DEBUG oslo_concurrency.lockutils [None req-cdbe0145-cd3c-4551-af87-0c7f8b53f26c tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.347s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 857.291021] env[61923]: DEBUG nova.compute.manager [None req-cdbe0145-cd3c-4551-af87-0c7f8b53f26c tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] [instance: e413c45d-cd89-44d4-9102-3d188907e7bb] Start building networks asynchronously for instance. {{(pid=61923) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 857.305162] env[61923]: DEBUG oslo_concurrency.lockutils [None req-13baf863-98e6-41c7-ae75-48fd1a9cef32 tempest-ServerAddressesTestJSON-1846886542 tempest-ServerAddressesTestJSON-1846886542-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 11.521s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 857.305162] env[61923]: INFO nova.compute.claims [None req-13baf863-98e6-41c7-ae75-48fd1a9cef32 tempest-ServerAddressesTestJSON-1846886542 tempest-ServerAddressesTestJSON-1846886542-project-member] [instance: 5b73e07a-d6ef-4dcf-bdae-eea91d2aeb6f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 857.452384] env[61923]: DEBUG oslo_vmware.api [None req-5fbf0568-2a85-4e3c-9b63-7d91ecc77593 tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Task: {'id': task-1377613, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.467145] env[61923]: DEBUG oslo_vmware.api [None req-80740a2c-9779-4afc-be30-c4469cd3724d tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]523c1134-fdb2-ef24-5a9f-aefc82cba757, 'name': SearchDatastore_Task, 'duration_secs': 0.026946} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 857.469088] env[61923]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9f05c4f5-8672-4d1f-b095-f0ce73c1d29f {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.483369] env[61923]: DEBUG oslo_vmware.api [None req-80740a2c-9779-4afc-be30-c4469cd3724d tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] Waiting for the task: (returnval){ [ 857.483369] env[61923]: value = "session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]5241f29f-f2e2-06b5-787e-fee81d7fef4e" [ 857.483369] env[61923]: _type = "Task" [ 857.483369] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 857.490945] env[61923]: DEBUG oslo_vmware.api [None req-aa1aa995-4e31-45e6-b824-29d545be859e tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Task: {'id': task-1377610, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.496797] env[61923]: DEBUG oslo_vmware.api [None req-80740a2c-9779-4afc-be30-c4469cd3724d tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]5241f29f-f2e2-06b5-787e-fee81d7fef4e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.508830] env[61923]: DEBUG oslo_vmware.api [None req-cdbe0145-cd3c-4551-af87-0c7f8b53f26c tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]529455a2-b56b-f24b-d51a-c1b07e65d8f4, 'name': SearchDatastore_Task, 'duration_secs': 0.039916} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 857.508830] env[61923]: DEBUG oslo_concurrency.lockutils [None req-cdbe0145-cd3c-4551-af87-0c7f8b53f26c tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 857.509047] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-cdbe0145-cd3c-4551-af87-0c7f8b53f26c tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] [instance: 444affa5-a7ed-4a17-9015-9fd5724aab64] Processing image 0f153f63-ae0a-45b1-b7f5-7f9b673c947f {{(pid=61923) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 857.511150] env[61923]: DEBUG oslo_concurrency.lockutils [None req-cdbe0145-cd3c-4551-af87-0c7f8b53f26c tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 857.589153] env[61923]: DEBUG nova.network.neutron [req-59ae460b-8f16-4abc-b4bc-e22bc9cf173e req-1291c6e8-dfa4-42af-94c4-01622cac4a2d service nova] [instance: 444affa5-a7ed-4a17-9015-9fd5724aab64] Updated VIF entry in instance network info cache for port 7b131b11-927e-4c0c-98f2-714b63683d77. {{(pid=61923) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 857.589551] env[61923]: DEBUG nova.network.neutron [req-59ae460b-8f16-4abc-b4bc-e22bc9cf173e req-1291c6e8-dfa4-42af-94c4-01622cac4a2d service nova] [instance: 444affa5-a7ed-4a17-9015-9fd5724aab64] Updating instance_info_cache with network_info: [{"id": "7b131b11-927e-4c0c-98f2-714b63683d77", "address": "fa:16:3e:70:f8:a3", "network": {"id": "545a170d-946b-4e0b-b9f5-b21c2dcda19d", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-40231173-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6e260342cd1b41ef8f0ceb15b6438daa", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3753f451-fa23-4988-9361-074fb0bd3fd4", "external-id": "nsx-vlan-transportzone-440", "segmentation_id": 440, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7b131b11-92", "ovs_interfaceid": "7b131b11-927e-4c0c-98f2-714b63683d77", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 857.640371] env[61923]: DEBUG oslo_concurrency.lockutils [None req-1214e428-bf00-4cbe-9e1e-7c8be1612c36 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] Lock "af3c317a-4007-4cea-a060-1e7dde5ce49e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 44.219s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 857.682143] env[61923]: DEBUG oslo_vmware.api [None req-52eadf86-cc45-4cae-af2a-645ac3cb8658 tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Task: {'id': task-1377612, 'name': ReconfigVM_Task, 'duration_secs': 0.875954} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 857.682527] env[61923]: DEBUG nova.virt.vmwareapi.volumeops [None req-52eadf86-cc45-4cae-af2a-645ac3cb8658 tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] [instance: 87ae37dc-d0d0-4b76-afdd-0ba3e8f6ce0b] Reconfigured VM instance instance-00000048 to attach disk [datastore1] 87ae37dc-d0d0-4b76-afdd-0ba3e8f6ce0b/87ae37dc-d0d0-4b76-afdd-0ba3e8f6ce0b.vmdk or device None with type sparse {{(pid=61923) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 857.683246] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f4d9c69d-453f-4a75-ab80-aec0b6f1a607 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.689723] env[61923]: DEBUG oslo_vmware.api [None req-52eadf86-cc45-4cae-af2a-645ac3cb8658 tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Waiting for the task: (returnval){ [ 857.689723] env[61923]: value = "task-1377614" [ 857.689723] env[61923]: _type = "Task" [ 857.689723] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 857.700575] env[61923]: DEBUG oslo_vmware.api [None req-52eadf86-cc45-4cae-af2a-645ac3cb8658 tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Task: {'id': task-1377614, 'name': Rename_Task} progress is 5%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.808960] env[61923]: DEBUG nova.compute.utils [None req-cdbe0145-cd3c-4551-af87-0c7f8b53f26c tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Using /dev/sd instead of None {{(pid=61923) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 857.821023] env[61923]: DEBUG nova.compute.manager [None req-cdbe0145-cd3c-4551-af87-0c7f8b53f26c tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] [instance: e413c45d-cd89-44d4-9102-3d188907e7bb] Allocating IP information in the background. {{(pid=61923) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 857.821023] env[61923]: DEBUG nova.network.neutron [None req-cdbe0145-cd3c-4551-af87-0c7f8b53f26c tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] [instance: e413c45d-cd89-44d4-9102-3d188907e7bb] allocate_for_instance() {{(pid=61923) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 857.911260] env[61923]: DEBUG nova.policy [None req-cdbe0145-cd3c-4551-af87-0c7f8b53f26c tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2631257787d44fcbac010b2bf1fd4b0e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '6e260342cd1b41ef8f0ceb15b6438daa', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61923) authorize /opt/stack/nova/nova/policy.py:201}} [ 857.954377] env[61923]: DEBUG oslo_vmware.api [None req-5fbf0568-2a85-4e3c-9b63-7d91ecc77593 tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Task: {'id': task-1377613, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.990079] env[61923]: DEBUG oslo_vmware.api [None req-aa1aa995-4e31-45e6-b824-29d545be859e tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Task: {'id': task-1377610, 'name': CreateSnapshot_Task, 'duration_secs': 1.347723} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 857.996155] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-aa1aa995-4e31-45e6-b824-29d545be859e tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] [instance: 59198f4d-4dde-4eaf-9f6c-a962cbe53c6e] Created Snapshot of the VM instance {{(pid=61923) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 857.997868] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d12531b-4c6f-4300-a0c3-37d7a1499398 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.007981] env[61923]: DEBUG oslo_vmware.api [None req-80740a2c-9779-4afc-be30-c4469cd3724d tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]5241f29f-f2e2-06b5-787e-fee81d7fef4e, 'name': SearchDatastore_Task, 'duration_secs': 0.01814} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 858.011786] env[61923]: DEBUG oslo_concurrency.lockutils [None req-80740a2c-9779-4afc-be30-c4469cd3724d tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 858.012118] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-80740a2c-9779-4afc-be30-c4469cd3724d tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk to [datastore1] fc4fdf89-4615-4586-9dcd-ea90dc258361/fc4fdf89-4615-4586-9dcd-ea90dc258361.vmdk {{(pid=61923) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 858.015015] env[61923]: DEBUG oslo_concurrency.lockutils [None req-cdbe0145-cd3c-4551-af87-0c7f8b53f26c tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 858.015263] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-cdbe0145-cd3c-4551-af87-0c7f8b53f26c tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61923) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 858.015696] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-67fd7b9e-093a-497f-8bb8-65cfdfc1f46c {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.017731] env[61923]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ccb36184-f555-4cd1-b1b1-33ce4cbd01d7 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.024335] env[61923]: DEBUG oslo_vmware.api [None req-80740a2c-9779-4afc-be30-c4469cd3724d tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] Waiting for the task: (returnval){ [ 858.024335] env[61923]: value = "task-1377615" [ 858.024335] env[61923]: _type = "Task" [ 858.024335] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 858.025622] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-cdbe0145-cd3c-4551-af87-0c7f8b53f26c tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61923) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 858.025922] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-cdbe0145-cd3c-4551-af87-0c7f8b53f26c tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61923) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 858.029297] env[61923]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cde6829e-6022-47f7-8a81-6ecc495cb2bb {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.034710] env[61923]: DEBUG oslo_vmware.api [None req-cdbe0145-cd3c-4551-af87-0c7f8b53f26c tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Waiting for the task: (returnval){ [ 858.034710] env[61923]: value = "session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]5235b651-94cb-f2cd-5e13-2d0d42407942" [ 858.034710] env[61923]: _type = "Task" [ 858.034710] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 858.037873] env[61923]: DEBUG oslo_vmware.api [None req-80740a2c-9779-4afc-be30-c4469cd3724d tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] Task: {'id': task-1377615, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.045341] env[61923]: DEBUG oslo_vmware.api [None req-cdbe0145-cd3c-4551-af87-0c7f8b53f26c tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]5235b651-94cb-f2cd-5e13-2d0d42407942, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.092567] env[61923]: DEBUG oslo_concurrency.lockutils [req-59ae460b-8f16-4abc-b4bc-e22bc9cf173e req-1291c6e8-dfa4-42af-94c4-01622cac4a2d service nova] Releasing lock "refresh_cache-444affa5-a7ed-4a17-9015-9fd5724aab64" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 858.200183] env[61923]: DEBUG oslo_vmware.api [None req-52eadf86-cc45-4cae-af2a-645ac3cb8658 tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Task: {'id': task-1377614, 'name': Rename_Task, 'duration_secs': 0.170312} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 858.200478] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-52eadf86-cc45-4cae-af2a-645ac3cb8658 tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] [instance: 87ae37dc-d0d0-4b76-afdd-0ba3e8f6ce0b] Powering on the VM {{(pid=61923) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 858.200788] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-af3b004c-18ae-4828-ab46-37e0468fc1dd {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.207803] env[61923]: DEBUG oslo_vmware.api [None req-52eadf86-cc45-4cae-af2a-645ac3cb8658 tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Waiting for the task: (returnval){ [ 858.207803] env[61923]: value = "task-1377616" [ 858.207803] env[61923]: _type = "Task" [ 858.207803] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 858.215692] env[61923]: DEBUG oslo_vmware.api [None req-52eadf86-cc45-4cae-af2a-645ac3cb8658 tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Task: {'id': task-1377616, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.242958] env[61923]: DEBUG nova.compute.manager [req-9e09373d-bccb-4db7-9ce3-bf87e67d23f6 req-22ece412-5d3d-4114-b010-122828f5b4bb service nova] [instance: af3c317a-4007-4cea-a060-1e7dde5ce49e] Received event network-changed-e5c62d99-3250-4fca-9798-25780ae50b00 {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 858.243331] env[61923]: DEBUG nova.compute.manager [req-9e09373d-bccb-4db7-9ce3-bf87e67d23f6 req-22ece412-5d3d-4114-b010-122828f5b4bb service nova] [instance: af3c317a-4007-4cea-a060-1e7dde5ce49e] Refreshing instance network info cache due to event network-changed-e5c62d99-3250-4fca-9798-25780ae50b00. {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 858.243639] env[61923]: DEBUG oslo_concurrency.lockutils [req-9e09373d-bccb-4db7-9ce3-bf87e67d23f6 req-22ece412-5d3d-4114-b010-122828f5b4bb service nova] Acquiring lock "refresh_cache-af3c317a-4007-4cea-a060-1e7dde5ce49e" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 858.243827] env[61923]: DEBUG oslo_concurrency.lockutils [req-9e09373d-bccb-4db7-9ce3-bf87e67d23f6 req-22ece412-5d3d-4114-b010-122828f5b4bb service nova] Acquired lock "refresh_cache-af3c317a-4007-4cea-a060-1e7dde5ce49e" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 858.244150] env[61923]: DEBUG nova.network.neutron [req-9e09373d-bccb-4db7-9ce3-bf87e67d23f6 req-22ece412-5d3d-4114-b010-122828f5b4bb service nova] [instance: af3c317a-4007-4cea-a060-1e7dde5ce49e] Refreshing network info cache for port e5c62d99-3250-4fca-9798-25780ae50b00 {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 858.313016] env[61923]: DEBUG nova.compute.manager [None req-cdbe0145-cd3c-4551-af87-0c7f8b53f26c tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] [instance: e413c45d-cd89-44d4-9102-3d188907e7bb] Start building block device mappings for instance. {{(pid=61923) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 858.453446] env[61923]: DEBUG nova.network.neutron [None req-cdbe0145-cd3c-4551-af87-0c7f8b53f26c tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] [instance: e413c45d-cd89-44d4-9102-3d188907e7bb] Successfully created port: 4ed56dd5-beb3-4254-932b-7290c7b85700 {{(pid=61923) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 858.465422] env[61923]: DEBUG oslo_vmware.api [None req-5fbf0568-2a85-4e3c-9b63-7d91ecc77593 tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Task: {'id': task-1377613, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.107373} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 858.465683] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-5fbf0568-2a85-4e3c-9b63-7d91ecc77593 tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk to [datastore1] 2719569b-8572-4199-8158-7bb367d17dc5/2719569b-8572-4199-8158-7bb367d17dc5.vmdk {{(pid=61923) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 858.465892] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-5fbf0568-2a85-4e3c-9b63-7d91ecc77593 tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] [instance: 2719569b-8572-4199-8158-7bb367d17dc5] Extending root virtual disk to 1048576 {{(pid=61923) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 858.466188] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-9abb4918-abc4-4010-a03d-97ba101e50a7 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.472931] env[61923]: DEBUG oslo_vmware.api [None req-5fbf0568-2a85-4e3c-9b63-7d91ecc77593 tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Waiting for the task: (returnval){ [ 858.472931] env[61923]: value = "task-1377617" [ 858.472931] env[61923]: _type = "Task" [ 858.472931] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 858.484594] env[61923]: DEBUG oslo_vmware.api [None req-5fbf0568-2a85-4e3c-9b63-7d91ecc77593 tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Task: {'id': task-1377617, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.527330] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-aa1aa995-4e31-45e6-b824-29d545be859e tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] [instance: 59198f4d-4dde-4eaf-9f6c-a962cbe53c6e] Creating linked-clone VM from snapshot {{(pid=61923) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 858.531074] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-7922ee96-1a85-4bf3-ad5a-27843cd13dd8 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.546470] env[61923]: DEBUG oslo_vmware.api [None req-aa1aa995-4e31-45e6-b824-29d545be859e tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Waiting for the task: (returnval){ [ 858.546470] env[61923]: value = "task-1377618" [ 858.546470] env[61923]: _type = "Task" [ 858.546470] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 858.557765] env[61923]: DEBUG oslo_vmware.api [None req-80740a2c-9779-4afc-be30-c4469cd3724d tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] Task: {'id': task-1377615, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.565227] env[61923]: DEBUG oslo_vmware.api [None req-cdbe0145-cd3c-4551-af87-0c7f8b53f26c tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]5235b651-94cb-f2cd-5e13-2d0d42407942, 'name': SearchDatastore_Task, 'duration_secs': 0.037083} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 858.569380] env[61923]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5a397968-dcd6-47cd-a2f1-eb9c6464e3ab {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.580593] env[61923]: DEBUG oslo_vmware.api [None req-aa1aa995-4e31-45e6-b824-29d545be859e tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Task: {'id': task-1377618, 'name': CloneVM_Task} progress is 11%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.587602] env[61923]: DEBUG oslo_vmware.api [None req-cdbe0145-cd3c-4551-af87-0c7f8b53f26c tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Waiting for the task: (returnval){ [ 858.587602] env[61923]: value = "session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52bef4e9-eb2c-c09f-e01d-c378128fb2dd" [ 858.587602] env[61923]: _type = "Task" [ 858.587602] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 858.599591] env[61923]: DEBUG oslo_vmware.api [None req-cdbe0145-cd3c-4551-af87-0c7f8b53f26c tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52bef4e9-eb2c-c09f-e01d-c378128fb2dd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.630601] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97e97c76-932a-45f6-b4e5-aa66dd2f8a9c {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.640204] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ce2e892-edbf-44eb-a60a-fbcf5415f288 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.673670] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3a0174e-cf34-43d6-833f-2176e0586ee7 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.681503] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c804da7-00fb-4973-bdf3-79b7440ac684 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.697988] env[61923]: DEBUG nova.compute.provider_tree [None req-13baf863-98e6-41c7-ae75-48fd1a9cef32 tempest-ServerAddressesTestJSON-1846886542 tempest-ServerAddressesTestJSON-1846886542-project-member] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 858.719179] env[61923]: DEBUG oslo_vmware.api [None req-52eadf86-cc45-4cae-af2a-645ac3cb8658 tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Task: {'id': task-1377616, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.986546] env[61923]: DEBUG oslo_vmware.api [None req-5fbf0568-2a85-4e3c-9b63-7d91ecc77593 tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Task: {'id': task-1377617, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.064945} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 858.986826] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-5fbf0568-2a85-4e3c-9b63-7d91ecc77593 tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] [instance: 2719569b-8572-4199-8158-7bb367d17dc5] Extended root virtual disk {{(pid=61923) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 858.987694] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf7702df-ffe6-4569-a46e-87825e4d39f7 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.011259] env[61923]: DEBUG nova.virt.vmwareapi.volumeops [None req-5fbf0568-2a85-4e3c-9b63-7d91ecc77593 tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] [instance: 2719569b-8572-4199-8158-7bb367d17dc5] Reconfiguring VM instance instance-00000049 to attach disk [datastore1] 2719569b-8572-4199-8158-7bb367d17dc5/2719569b-8572-4199-8158-7bb367d17dc5.vmdk or device None with type sparse {{(pid=61923) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 859.013901] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-61ccd0b2-0cae-4a9b-a4ab-fd6610b83caf {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.035857] env[61923]: DEBUG oslo_vmware.api [None req-5fbf0568-2a85-4e3c-9b63-7d91ecc77593 tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Waiting for the task: (returnval){ [ 859.035857] env[61923]: value = "task-1377619" [ 859.035857] env[61923]: _type = "Task" [ 859.035857] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 859.044030] env[61923]: DEBUG oslo_vmware.api [None req-80740a2c-9779-4afc-be30-c4469cd3724d tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] Task: {'id': task-1377615, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.922213} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 859.044927] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-80740a2c-9779-4afc-be30-c4469cd3724d tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk to [datastore1] fc4fdf89-4615-4586-9dcd-ea90dc258361/fc4fdf89-4615-4586-9dcd-ea90dc258361.vmdk {{(pid=61923) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 859.045275] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-80740a2c-9779-4afc-be30-c4469cd3724d tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] [instance: fc4fdf89-4615-4586-9dcd-ea90dc258361] Extending root virtual disk to 1048576 {{(pid=61923) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 859.045564] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-f61e9563-71f0-47fe-8358-94a2c93b0def {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.051011] env[61923]: DEBUG oslo_vmware.api [None req-5fbf0568-2a85-4e3c-9b63-7d91ecc77593 tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Task: {'id': task-1377619, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.055810] env[61923]: DEBUG oslo_vmware.api [None req-80740a2c-9779-4afc-be30-c4469cd3724d tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] Waiting for the task: (returnval){ [ 859.055810] env[61923]: value = "task-1377620" [ 859.055810] env[61923]: _type = "Task" [ 859.055810] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 859.068587] env[61923]: DEBUG oslo_vmware.api [None req-80740a2c-9779-4afc-be30-c4469cd3724d tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] Task: {'id': task-1377620, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.072511] env[61923]: DEBUG oslo_vmware.api [None req-aa1aa995-4e31-45e6-b824-29d545be859e tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Task: {'id': task-1377618, 'name': CloneVM_Task} progress is 94%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.088933] env[61923]: DEBUG nova.network.neutron [req-9e09373d-bccb-4db7-9ce3-bf87e67d23f6 req-22ece412-5d3d-4114-b010-122828f5b4bb service nova] [instance: af3c317a-4007-4cea-a060-1e7dde5ce49e] Updated VIF entry in instance network info cache for port e5c62d99-3250-4fca-9798-25780ae50b00. {{(pid=61923) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 859.088933] env[61923]: DEBUG nova.network.neutron [req-9e09373d-bccb-4db7-9ce3-bf87e67d23f6 req-22ece412-5d3d-4114-b010-122828f5b4bb service nova] [instance: af3c317a-4007-4cea-a060-1e7dde5ce49e] Updating instance_info_cache with network_info: [{"id": "e5c62d99-3250-4fca-9798-25780ae50b00", "address": "fa:16:3e:2c:0a:82", "network": {"id": "22a2f26f-6020-43a0-9e3e-32d28145dd2b", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-1080215875-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.249", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "87f548d252244b759e3f157af9ab875d", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "43798f54-0c07-4417-a23f-58bb6b7e204b", "external-id": "nsx-vlan-transportzone-571", "segmentation_id": 571, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape5c62d99-32", "ovs_interfaceid": "e5c62d99-3250-4fca-9798-25780ae50b00", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 859.099618] env[61923]: DEBUG oslo_vmware.api [None req-cdbe0145-cd3c-4551-af87-0c7f8b53f26c tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52bef4e9-eb2c-c09f-e01d-c378128fb2dd, 'name': SearchDatastore_Task, 'duration_secs': 0.063036} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 859.100196] env[61923]: DEBUG oslo_concurrency.lockutils [None req-cdbe0145-cd3c-4551-af87-0c7f8b53f26c tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 859.100980] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-cdbe0145-cd3c-4551-af87-0c7f8b53f26c tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk to [datastore1] 444affa5-a7ed-4a17-9015-9fd5724aab64/444affa5-a7ed-4a17-9015-9fd5724aab64.vmdk {{(pid=61923) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 859.100980] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f5081f48-b5bc-4650-a4e7-bf9845ca587f {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.107776] env[61923]: DEBUG oslo_vmware.api [None req-cdbe0145-cd3c-4551-af87-0c7f8b53f26c tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Waiting for the task: (returnval){ [ 859.107776] env[61923]: value = "task-1377621" [ 859.107776] env[61923]: _type = "Task" [ 859.107776] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 859.117720] env[61923]: DEBUG oslo_vmware.api [None req-cdbe0145-cd3c-4551-af87-0c7f8b53f26c tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Task: {'id': task-1377621, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.202337] env[61923]: DEBUG nova.scheduler.client.report [None req-13baf863-98e6-41c7-ae75-48fd1a9cef32 tempest-ServerAddressesTestJSON-1846886542 tempest-ServerAddressesTestJSON-1846886542-project-member] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 859.222022] env[61923]: DEBUG oslo_vmware.api [None req-52eadf86-cc45-4cae-af2a-645ac3cb8658 tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Task: {'id': task-1377616, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.325239] env[61923]: DEBUG nova.compute.manager [None req-cdbe0145-cd3c-4551-af87-0c7f8b53f26c tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] [instance: e413c45d-cd89-44d4-9102-3d188907e7bb] Start spawning the instance on the hypervisor. {{(pid=61923) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 859.354088] env[61923]: DEBUG nova.virt.hardware [None req-cdbe0145-cd3c-4551-af87-0c7f8b53f26c tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-29T20:07:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-29T20:07:35Z,direct_url=,disk_format='vmdk',id=0f153f63-ae0a-45b1-b7f5-7f9b673c947f,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='4e7b5e3b3c3443c8bd5c70f8a15739f5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-29T20:07:36Z,virtual_size=,visibility=), allow threads: False {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 859.354458] env[61923]: DEBUG nova.virt.hardware [None req-cdbe0145-cd3c-4551-af87-0c7f8b53f26c tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Flavor limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 859.354629] env[61923]: DEBUG nova.virt.hardware [None req-cdbe0145-cd3c-4551-af87-0c7f8b53f26c tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Image limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 859.354822] env[61923]: DEBUG nova.virt.hardware [None req-cdbe0145-cd3c-4551-af87-0c7f8b53f26c tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Flavor pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 859.354965] env[61923]: DEBUG nova.virt.hardware [None req-cdbe0145-cd3c-4551-af87-0c7f8b53f26c tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Image pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 859.355131] env[61923]: DEBUG nova.virt.hardware [None req-cdbe0145-cd3c-4551-af87-0c7f8b53f26c tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 859.355373] env[61923]: DEBUG nova.virt.hardware [None req-cdbe0145-cd3c-4551-af87-0c7f8b53f26c tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 859.355545] env[61923]: DEBUG nova.virt.hardware [None req-cdbe0145-cd3c-4551-af87-0c7f8b53f26c tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 859.355710] env[61923]: DEBUG nova.virt.hardware [None req-cdbe0145-cd3c-4551-af87-0c7f8b53f26c tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Got 1 possible topologies {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 859.355895] env[61923]: DEBUG nova.virt.hardware [None req-cdbe0145-cd3c-4551-af87-0c7f8b53f26c tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 859.356520] env[61923]: DEBUG nova.virt.hardware [None req-cdbe0145-cd3c-4551-af87-0c7f8b53f26c tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 859.357455] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-877e7bb7-8143-48a6-8c53-a95c733c452c {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.366578] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57771481-cc73-493f-bfac-41dbbe6fdcd1 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.548104] env[61923]: DEBUG oslo_vmware.api [None req-5fbf0568-2a85-4e3c-9b63-7d91ecc77593 tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Task: {'id': task-1377619, 'name': ReconfigVM_Task, 'duration_secs': 0.296749} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 859.548104] env[61923]: DEBUG nova.virt.vmwareapi.volumeops [None req-5fbf0568-2a85-4e3c-9b63-7d91ecc77593 tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] [instance: 2719569b-8572-4199-8158-7bb367d17dc5] Reconfigured VM instance instance-00000049 to attach disk [datastore1] 2719569b-8572-4199-8158-7bb367d17dc5/2719569b-8572-4199-8158-7bb367d17dc5.vmdk or device None with type sparse {{(pid=61923) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 859.548104] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-0ea9f8b1-0eb3-4941-80a4-aad3db478e12 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.554463] env[61923]: DEBUG oslo_vmware.api [None req-5fbf0568-2a85-4e3c-9b63-7d91ecc77593 tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Waiting for the task: (returnval){ [ 859.554463] env[61923]: value = "task-1377622" [ 859.554463] env[61923]: _type = "Task" [ 859.554463] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 859.569205] env[61923]: DEBUG oslo_vmware.api [None req-5fbf0568-2a85-4e3c-9b63-7d91ecc77593 tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Task: {'id': task-1377622, 'name': Rename_Task} progress is 5%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.573223] env[61923]: DEBUG oslo_vmware.api [None req-80740a2c-9779-4afc-be30-c4469cd3724d tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] Task: {'id': task-1377620, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.076959} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 859.573223] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-80740a2c-9779-4afc-be30-c4469cd3724d tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] [instance: fc4fdf89-4615-4586-9dcd-ea90dc258361] Extended root virtual disk {{(pid=61923) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 859.573739] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4746d38-4f81-4cdd-9624-e8d768e21bae {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.579651] env[61923]: DEBUG oslo_vmware.api [None req-aa1aa995-4e31-45e6-b824-29d545be859e tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Task: {'id': task-1377618, 'name': CloneVM_Task} progress is 95%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.599057] env[61923]: DEBUG oslo_concurrency.lockutils [req-9e09373d-bccb-4db7-9ce3-bf87e67d23f6 req-22ece412-5d3d-4114-b010-122828f5b4bb service nova] Releasing lock "refresh_cache-af3c317a-4007-4cea-a060-1e7dde5ce49e" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 859.608770] env[61923]: DEBUG nova.virt.vmwareapi.volumeops [None req-80740a2c-9779-4afc-be30-c4469cd3724d tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] [instance: fc4fdf89-4615-4586-9dcd-ea90dc258361] Reconfiguring VM instance instance-00000046 to attach disk [datastore1] fc4fdf89-4615-4586-9dcd-ea90dc258361/fc4fdf89-4615-4586-9dcd-ea90dc258361.vmdk or device None with type sparse {{(pid=61923) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 859.609259] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-146b0566-3a6c-4649-93ee-f0d6cf092f24 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.632500] env[61923]: DEBUG oslo_vmware.api [None req-cdbe0145-cd3c-4551-af87-0c7f8b53f26c tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Task: {'id': task-1377621, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.634106] env[61923]: DEBUG oslo_vmware.api [None req-80740a2c-9779-4afc-be30-c4469cd3724d tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] Waiting for the task: (returnval){ [ 859.634106] env[61923]: value = "task-1377623" [ 859.634106] env[61923]: _type = "Task" [ 859.634106] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 859.642453] env[61923]: DEBUG oslo_vmware.api [None req-80740a2c-9779-4afc-be30-c4469cd3724d tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] Task: {'id': task-1377623, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.708737] env[61923]: DEBUG oslo_concurrency.lockutils [None req-13baf863-98e6-41c7-ae75-48fd1a9cef32 tempest-ServerAddressesTestJSON-1846886542 tempest-ServerAddressesTestJSON-1846886542-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.411s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 859.709418] env[61923]: DEBUG nova.compute.manager [None req-13baf863-98e6-41c7-ae75-48fd1a9cef32 tempest-ServerAddressesTestJSON-1846886542 tempest-ServerAddressesTestJSON-1846886542-project-member] [instance: 5b73e07a-d6ef-4dcf-bdae-eea91d2aeb6f] Start building networks asynchronously for instance. {{(pid=61923) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 859.712660] env[61923]: DEBUG oslo_concurrency.lockutils [None req-00703b42-f309-4672-8dc7-10f4a69426aa tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 11.849s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 859.714430] env[61923]: INFO nova.compute.claims [None req-00703b42-f309-4672-8dc7-10f4a69426aa tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] [instance: b779d183-89ae-4e4d-ae99-e514e145ed43] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 859.727523] env[61923]: DEBUG oslo_vmware.api [None req-52eadf86-cc45-4cae-af2a-645ac3cb8658 tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Task: {'id': task-1377616, 'name': PowerOnVM_Task, 'duration_secs': 1.195442} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 859.727817] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-52eadf86-cc45-4cae-af2a-645ac3cb8658 tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] [instance: 87ae37dc-d0d0-4b76-afdd-0ba3e8f6ce0b] Powered on the VM {{(pid=61923) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 859.729212] env[61923]: INFO nova.compute.manager [None req-52eadf86-cc45-4cae-af2a-645ac3cb8658 tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] [instance: 87ae37dc-d0d0-4b76-afdd-0ba3e8f6ce0b] Took 10.34 seconds to spawn the instance on the hypervisor. [ 859.729212] env[61923]: DEBUG nova.compute.manager [None req-52eadf86-cc45-4cae-af2a-645ac3cb8658 tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] [instance: 87ae37dc-d0d0-4b76-afdd-0ba3e8f6ce0b] Checking state {{(pid=61923) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 859.729338] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d96eb87-dbd8-4f62-a87c-d26e67d67965 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.065327] env[61923]: DEBUG oslo_vmware.api [None req-5fbf0568-2a85-4e3c-9b63-7d91ecc77593 tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Task: {'id': task-1377622, 'name': Rename_Task, 'duration_secs': 0.167852} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 860.068641] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-5fbf0568-2a85-4e3c-9b63-7d91ecc77593 tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] [instance: 2719569b-8572-4199-8158-7bb367d17dc5] Powering on the VM {{(pid=61923) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 860.068929] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1b432cda-f08e-4ef1-890e-f683e1e9a8e6 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.075460] env[61923]: DEBUG oslo_vmware.api [None req-aa1aa995-4e31-45e6-b824-29d545be859e tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Task: {'id': task-1377618, 'name': CloneVM_Task, 'duration_secs': 1.241518} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 860.076661] env[61923]: INFO nova.virt.vmwareapi.vmops [None req-aa1aa995-4e31-45e6-b824-29d545be859e tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] [instance: 59198f4d-4dde-4eaf-9f6c-a962cbe53c6e] Created linked-clone VM from snapshot [ 860.076984] env[61923]: DEBUG oslo_vmware.api [None req-5fbf0568-2a85-4e3c-9b63-7d91ecc77593 tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Waiting for the task: (returnval){ [ 860.076984] env[61923]: value = "task-1377624" [ 860.076984] env[61923]: _type = "Task" [ 860.076984] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 860.077664] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21456308-ffb3-4ff2-83f7-6000d0c80001 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.087764] env[61923]: DEBUG oslo_vmware.api [None req-5fbf0568-2a85-4e3c-9b63-7d91ecc77593 tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Task: {'id': task-1377624, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.091013] env[61923]: DEBUG nova.virt.vmwareapi.images [None req-aa1aa995-4e31-45e6-b824-29d545be859e tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] [instance: 59198f4d-4dde-4eaf-9f6c-a962cbe53c6e] Uploading image 9ddad7e6-f1e4-49aa-b956-1b8d3273318d {{(pid=61923) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 860.118654] env[61923]: DEBUG oslo_vmware.rw_handles [None req-aa1aa995-4e31-45e6-b824-29d545be859e tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 860.118654] env[61923]: value = "vm-292686" [ 860.118654] env[61923]: _type = "VirtualMachine" [ 860.118654] env[61923]: }. {{(pid=61923) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 860.118960] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-f0d992ca-c78a-4b0e-9089-93d52b9800d6 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.123573] env[61923]: DEBUG oslo_vmware.api [None req-cdbe0145-cd3c-4551-af87-0c7f8b53f26c tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Task: {'id': task-1377621, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.731295} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 860.124191] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-cdbe0145-cd3c-4551-af87-0c7f8b53f26c tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk to [datastore1] 444affa5-a7ed-4a17-9015-9fd5724aab64/444affa5-a7ed-4a17-9015-9fd5724aab64.vmdk {{(pid=61923) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 860.124409] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-cdbe0145-cd3c-4551-af87-0c7f8b53f26c tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] [instance: 444affa5-a7ed-4a17-9015-9fd5724aab64] Extending root virtual disk to 1048576 {{(pid=61923) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 860.124642] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-bfa8a358-ee27-436d-9562-5c3d8fce2acb {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.127803] env[61923]: DEBUG oslo_vmware.rw_handles [None req-aa1aa995-4e31-45e6-b824-29d545be859e tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Lease: (returnval){ [ 860.127803] env[61923]: value = "session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]5207ba3f-0658-d50a-70f2-56799cd72d40" [ 860.127803] env[61923]: _type = "HttpNfcLease" [ 860.127803] env[61923]: } obtained for exporting VM: (result){ [ 860.127803] env[61923]: value = "vm-292686" [ 860.127803] env[61923]: _type = "VirtualMachine" [ 860.127803] env[61923]: }. {{(pid=61923) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 860.128075] env[61923]: DEBUG oslo_vmware.api [None req-aa1aa995-4e31-45e6-b824-29d545be859e tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Waiting for the lease: (returnval){ [ 860.128075] env[61923]: value = "session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]5207ba3f-0658-d50a-70f2-56799cd72d40" [ 860.128075] env[61923]: _type = "HttpNfcLease" [ 860.128075] env[61923]: } to be ready. {{(pid=61923) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 860.133891] env[61923]: DEBUG oslo_vmware.api [None req-cdbe0145-cd3c-4551-af87-0c7f8b53f26c tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Waiting for the task: (returnval){ [ 860.133891] env[61923]: value = "task-1377626" [ 860.133891] env[61923]: _type = "Task" [ 860.133891] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 860.147624] env[61923]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 860.147624] env[61923]: value = "session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]5207ba3f-0658-d50a-70f2-56799cd72d40" [ 860.147624] env[61923]: _type = "HttpNfcLease" [ 860.147624] env[61923]: } is initializing. {{(pid=61923) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 860.154164] env[61923]: DEBUG oslo_vmware.api [None req-cdbe0145-cd3c-4551-af87-0c7f8b53f26c tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Task: {'id': task-1377626, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.159158] env[61923]: DEBUG oslo_vmware.api [None req-80740a2c-9779-4afc-be30-c4469cd3724d tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] Task: {'id': task-1377623, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.220187] env[61923]: DEBUG nova.compute.utils [None req-13baf863-98e6-41c7-ae75-48fd1a9cef32 tempest-ServerAddressesTestJSON-1846886542 tempest-ServerAddressesTestJSON-1846886542-project-member] Using /dev/sd instead of None {{(pid=61923) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 860.224015] env[61923]: DEBUG nova.compute.manager [None req-13baf863-98e6-41c7-ae75-48fd1a9cef32 tempest-ServerAddressesTestJSON-1846886542 tempest-ServerAddressesTestJSON-1846886542-project-member] [instance: 5b73e07a-d6ef-4dcf-bdae-eea91d2aeb6f] Allocating IP information in the background. {{(pid=61923) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 860.224015] env[61923]: DEBUG nova.network.neutron [None req-13baf863-98e6-41c7-ae75-48fd1a9cef32 tempest-ServerAddressesTestJSON-1846886542 tempest-ServerAddressesTestJSON-1846886542-project-member] [instance: 5b73e07a-d6ef-4dcf-bdae-eea91d2aeb6f] allocate_for_instance() {{(pid=61923) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 860.252141] env[61923]: INFO nova.compute.manager [None req-52eadf86-cc45-4cae-af2a-645ac3cb8658 tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] [instance: 87ae37dc-d0d0-4b76-afdd-0ba3e8f6ce0b] Took 31.65 seconds to build instance. [ 860.282019] env[61923]: DEBUG nova.policy [None req-13baf863-98e6-41c7-ae75-48fd1a9cef32 tempest-ServerAddressesTestJSON-1846886542 tempest-ServerAddressesTestJSON-1846886542-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '1ee4d88a54834f409b3e46c5e594bc0a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a69fe23e6b4c4526a488b2be81e6882e', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61923) authorize /opt/stack/nova/nova/policy.py:201}} [ 860.287131] env[61923]: DEBUG nova.compute.manager [req-3bbb1936-7ea4-4332-8151-631564ecab66 req-04528e91-b4c3-445b-aafc-bca7a5cec916 service nova] [instance: e413c45d-cd89-44d4-9102-3d188907e7bb] Received event network-vif-plugged-4ed56dd5-beb3-4254-932b-7290c7b85700 {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 860.287521] env[61923]: DEBUG oslo_concurrency.lockutils [req-3bbb1936-7ea4-4332-8151-631564ecab66 req-04528e91-b4c3-445b-aafc-bca7a5cec916 service nova] Acquiring lock "e413c45d-cd89-44d4-9102-3d188907e7bb-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 860.287627] env[61923]: DEBUG oslo_concurrency.lockutils [req-3bbb1936-7ea4-4332-8151-631564ecab66 req-04528e91-b4c3-445b-aafc-bca7a5cec916 service nova] Lock "e413c45d-cd89-44d4-9102-3d188907e7bb-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 860.287842] env[61923]: DEBUG oslo_concurrency.lockutils [req-3bbb1936-7ea4-4332-8151-631564ecab66 req-04528e91-b4c3-445b-aafc-bca7a5cec916 service nova] Lock "e413c45d-cd89-44d4-9102-3d188907e7bb-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 860.288026] env[61923]: DEBUG nova.compute.manager [req-3bbb1936-7ea4-4332-8151-631564ecab66 req-04528e91-b4c3-445b-aafc-bca7a5cec916 service nova] [instance: e413c45d-cd89-44d4-9102-3d188907e7bb] No waiting events found dispatching network-vif-plugged-4ed56dd5-beb3-4254-932b-7290c7b85700 {{(pid=61923) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 860.288082] env[61923]: WARNING nova.compute.manager [req-3bbb1936-7ea4-4332-8151-631564ecab66 req-04528e91-b4c3-445b-aafc-bca7a5cec916 service nova] [instance: e413c45d-cd89-44d4-9102-3d188907e7bb] Received unexpected event network-vif-plugged-4ed56dd5-beb3-4254-932b-7290c7b85700 for instance with vm_state building and task_state spawning. [ 860.396127] env[61923]: DEBUG nova.network.neutron [None req-cdbe0145-cd3c-4551-af87-0c7f8b53f26c tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] [instance: e413c45d-cd89-44d4-9102-3d188907e7bb] Successfully updated port: 4ed56dd5-beb3-4254-932b-7290c7b85700 {{(pid=61923) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 860.568073] env[61923]: DEBUG nova.network.neutron [None req-13baf863-98e6-41c7-ae75-48fd1a9cef32 tempest-ServerAddressesTestJSON-1846886542 tempest-ServerAddressesTestJSON-1846886542-project-member] [instance: 5b73e07a-d6ef-4dcf-bdae-eea91d2aeb6f] Successfully created port: 262b2cfc-fe55-4a4e-bb7c-87f81d8c7e91 {{(pid=61923) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 860.590071] env[61923]: DEBUG oslo_vmware.api [None req-5fbf0568-2a85-4e3c-9b63-7d91ecc77593 tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Task: {'id': task-1377624, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.636936] env[61923]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 860.636936] env[61923]: value = "session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]5207ba3f-0658-d50a-70f2-56799cd72d40" [ 860.636936] env[61923]: _type = "HttpNfcLease" [ 860.636936] env[61923]: } is ready. {{(pid=61923) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 860.642321] env[61923]: DEBUG oslo_vmware.rw_handles [None req-aa1aa995-4e31-45e6-b824-29d545be859e tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 860.642321] env[61923]: value = "session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]5207ba3f-0658-d50a-70f2-56799cd72d40" [ 860.642321] env[61923]: _type = "HttpNfcLease" [ 860.642321] env[61923]: }. {{(pid=61923) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 860.643301] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ebdff14-46f5-48ee-8a43-8edcd54ee90a {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.650697] env[61923]: DEBUG oslo_vmware.api [None req-cdbe0145-cd3c-4551-af87-0c7f8b53f26c tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Task: {'id': task-1377626, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.067179} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 860.657131] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-cdbe0145-cd3c-4551-af87-0c7f8b53f26c tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] [instance: 444affa5-a7ed-4a17-9015-9fd5724aab64] Extended root virtual disk {{(pid=61923) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 860.657435] env[61923]: DEBUG oslo_vmware.api [None req-80740a2c-9779-4afc-be30-c4469cd3724d tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] Task: {'id': task-1377623, 'name': ReconfigVM_Task, 'duration_secs': 0.781868} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 860.657667] env[61923]: DEBUG oslo_vmware.rw_handles [None req-aa1aa995-4e31-45e6-b824-29d545be859e tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/526fd05d-3941-e7f4-4056-812724b9c6fb/disk-0.vmdk from lease info. {{(pid=61923) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 860.657826] env[61923]: DEBUG oslo_vmware.rw_handles [None req-aa1aa995-4e31-45e6-b824-29d545be859e tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Opening URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/526fd05d-3941-e7f4-4056-812724b9c6fb/disk-0.vmdk for reading. {{(pid=61923) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 860.659826] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5675e95-3ad3-480c-bad5-69b3dbef4e64 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.662363] env[61923]: DEBUG nova.virt.vmwareapi.volumeops [None req-80740a2c-9779-4afc-be30-c4469cd3724d tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] [instance: fc4fdf89-4615-4586-9dcd-ea90dc258361] Reconfigured VM instance instance-00000046 to attach disk [datastore1] fc4fdf89-4615-4586-9dcd-ea90dc258361/fc4fdf89-4615-4586-9dcd-ea90dc258361.vmdk or device None with type sparse {{(pid=61923) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 860.663849] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-0e7289a1-54f7-49ad-a217-2684006349c5 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.734417] env[61923]: DEBUG nova.compute.manager [None req-13baf863-98e6-41c7-ae75-48fd1a9cef32 tempest-ServerAddressesTestJSON-1846886542 tempest-ServerAddressesTestJSON-1846886542-project-member] [instance: 5b73e07a-d6ef-4dcf-bdae-eea91d2aeb6f] Start building block device mappings for instance. {{(pid=61923) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 860.748634] env[61923]: DEBUG nova.virt.vmwareapi.volumeops [None req-cdbe0145-cd3c-4551-af87-0c7f8b53f26c tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] [instance: 444affa5-a7ed-4a17-9015-9fd5724aab64] Reconfiguring VM instance instance-0000004a to attach disk [datastore1] 444affa5-a7ed-4a17-9015-9fd5724aab64/444affa5-a7ed-4a17-9015-9fd5724aab64.vmdk or device None with type sparse {{(pid=61923) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 860.752347] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2d985cc0-2636-49a3-b07c-961c39dab122 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.770862] env[61923]: DEBUG oslo_vmware.api [None req-80740a2c-9779-4afc-be30-c4469cd3724d tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] Waiting for the task: (returnval){ [ 860.770862] env[61923]: value = "task-1377627" [ 860.770862] env[61923]: _type = "Task" [ 860.770862] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 860.770862] env[61923]: DEBUG oslo_concurrency.lockutils [None req-52eadf86-cc45-4cae-af2a-645ac3cb8658 tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Lock "87ae37dc-d0d0-4b76-afdd-0ba3e8f6ce0b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 45.352s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 860.775872] env[61923]: DEBUG oslo_vmware.api [None req-cdbe0145-cd3c-4551-af87-0c7f8b53f26c tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Waiting for the task: (returnval){ [ 860.775872] env[61923]: value = "task-1377628" [ 860.775872] env[61923]: _type = "Task" [ 860.775872] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 860.779794] env[61923]: DEBUG oslo_vmware.api [None req-80740a2c-9779-4afc-be30-c4469cd3724d tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] Task: {'id': task-1377627, 'name': Rename_Task} progress is 14%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.791206] env[61923]: DEBUG oslo_vmware.api [None req-cdbe0145-cd3c-4551-af87-0c7f8b53f26c tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Task: {'id': task-1377628, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.809072] env[61923]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-89daaf61-865e-43b9-82b9-b562195ba2a4 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.898210] env[61923]: DEBUG oslo_concurrency.lockutils [None req-cdbe0145-cd3c-4551-af87-0c7f8b53f26c tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Acquiring lock "refresh_cache-e413c45d-cd89-44d4-9102-3d188907e7bb" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 860.898447] env[61923]: DEBUG oslo_concurrency.lockutils [None req-cdbe0145-cd3c-4551-af87-0c7f8b53f26c tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Acquired lock "refresh_cache-e413c45d-cd89-44d4-9102-3d188907e7bb" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 860.898596] env[61923]: DEBUG nova.network.neutron [None req-cdbe0145-cd3c-4551-af87-0c7f8b53f26c tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] [instance: e413c45d-cd89-44d4-9102-3d188907e7bb] Building network info cache for instance {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 861.052311] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25e13f5d-9874-432f-b245-c2bb24449e4e {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.062618] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a5e9d8d-0d72-47a7-b139-b5e481572063 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.102030] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3748cfc6-855a-4791-96b9-a49374873cf3 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.110419] env[61923]: DEBUG oslo_vmware.api [None req-5fbf0568-2a85-4e3c-9b63-7d91ecc77593 tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Task: {'id': task-1377624, 'name': PowerOnVM_Task, 'duration_secs': 0.894774} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 861.112661] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-5fbf0568-2a85-4e3c-9b63-7d91ecc77593 tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] [instance: 2719569b-8572-4199-8158-7bb367d17dc5] Powered on the VM {{(pid=61923) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 861.113068] env[61923]: INFO nova.compute.manager [None req-5fbf0568-2a85-4e3c-9b63-7d91ecc77593 tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] [instance: 2719569b-8572-4199-8158-7bb367d17dc5] Took 9.30 seconds to spawn the instance on the hypervisor. [ 861.113141] env[61923]: DEBUG nova.compute.manager [None req-5fbf0568-2a85-4e3c-9b63-7d91ecc77593 tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] [instance: 2719569b-8572-4199-8158-7bb367d17dc5] Checking state {{(pid=61923) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 861.113940] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3911e448-868c-40e1-b331-d55b4aeb48bc {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.117515] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea87cef0-42ec-4511-8508-6c573fbe8b46 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.135409] env[61923]: DEBUG nova.compute.provider_tree [None req-00703b42-f309-4672-8dc7-10f4a69426aa tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 861.280755] env[61923]: DEBUG oslo_vmware.api [None req-80740a2c-9779-4afc-be30-c4469cd3724d tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] Task: {'id': task-1377627, 'name': Rename_Task, 'duration_secs': 0.185448} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 861.284768] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-80740a2c-9779-4afc-be30-c4469cd3724d tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] [instance: fc4fdf89-4615-4586-9dcd-ea90dc258361] Powering on the VM {{(pid=61923) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 861.285174] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-363cd50c-6068-45d8-9139-8796e7220827 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.293712] env[61923]: DEBUG oslo_vmware.api [None req-cdbe0145-cd3c-4551-af87-0c7f8b53f26c tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Task: {'id': task-1377628, 'name': ReconfigVM_Task, 'duration_secs': 0.468146} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 861.295465] env[61923]: DEBUG nova.virt.vmwareapi.volumeops [None req-cdbe0145-cd3c-4551-af87-0c7f8b53f26c tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] [instance: 444affa5-a7ed-4a17-9015-9fd5724aab64] Reconfigured VM instance instance-0000004a to attach disk [datastore1] 444affa5-a7ed-4a17-9015-9fd5724aab64/444affa5-a7ed-4a17-9015-9fd5724aab64.vmdk or device None with type sparse {{(pid=61923) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 861.296014] env[61923]: DEBUG oslo_vmware.api [None req-80740a2c-9779-4afc-be30-c4469cd3724d tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] Waiting for the task: (returnval){ [ 861.296014] env[61923]: value = "task-1377629" [ 861.296014] env[61923]: _type = "Task" [ 861.296014] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 861.296285] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-2b64942e-4846-478e-a572-bbe0c7e8ce4f {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.306376] env[61923]: DEBUG oslo_vmware.api [None req-80740a2c-9779-4afc-be30-c4469cd3724d tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] Task: {'id': task-1377629, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.307792] env[61923]: DEBUG oslo_vmware.api [None req-cdbe0145-cd3c-4551-af87-0c7f8b53f26c tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Waiting for the task: (returnval){ [ 861.307792] env[61923]: value = "task-1377630" [ 861.307792] env[61923]: _type = "Task" [ 861.307792] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 861.316283] env[61923]: DEBUG oslo_vmware.api [None req-cdbe0145-cd3c-4551-af87-0c7f8b53f26c tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Task: {'id': task-1377630, 'name': Rename_Task} progress is 6%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.432710] env[61923]: DEBUG nova.network.neutron [None req-cdbe0145-cd3c-4551-af87-0c7f8b53f26c tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] [instance: e413c45d-cd89-44d4-9102-3d188907e7bb] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 861.643515] env[61923]: DEBUG nova.scheduler.client.report [None req-00703b42-f309-4672-8dc7-10f4a69426aa tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 861.649876] env[61923]: INFO nova.compute.manager [None req-5fbf0568-2a85-4e3c-9b63-7d91ecc77593 tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] [instance: 2719569b-8572-4199-8158-7bb367d17dc5] Took 26.10 seconds to build instance. [ 861.757874] env[61923]: DEBUG nova.compute.manager [None req-13baf863-98e6-41c7-ae75-48fd1a9cef32 tempest-ServerAddressesTestJSON-1846886542 tempest-ServerAddressesTestJSON-1846886542-project-member] [instance: 5b73e07a-d6ef-4dcf-bdae-eea91d2aeb6f] Start spawning the instance on the hypervisor. {{(pid=61923) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 861.783602] env[61923]: DEBUG nova.virt.hardware [None req-13baf863-98e6-41c7-ae75-48fd1a9cef32 tempest-ServerAddressesTestJSON-1846886542 tempest-ServerAddressesTestJSON-1846886542-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-29T20:07:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-29T20:07:35Z,direct_url=,disk_format='vmdk',id=0f153f63-ae0a-45b1-b7f5-7f9b673c947f,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='4e7b5e3b3c3443c8bd5c70f8a15739f5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-29T20:07:36Z,virtual_size=,visibility=), allow threads: False {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 861.783838] env[61923]: DEBUG nova.virt.hardware [None req-13baf863-98e6-41c7-ae75-48fd1a9cef32 tempest-ServerAddressesTestJSON-1846886542 tempest-ServerAddressesTestJSON-1846886542-project-member] Flavor limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 861.784020] env[61923]: DEBUG nova.virt.hardware [None req-13baf863-98e6-41c7-ae75-48fd1a9cef32 tempest-ServerAddressesTestJSON-1846886542 tempest-ServerAddressesTestJSON-1846886542-project-member] Image limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 861.785313] env[61923]: DEBUG nova.virt.hardware [None req-13baf863-98e6-41c7-ae75-48fd1a9cef32 tempest-ServerAddressesTestJSON-1846886542 tempest-ServerAddressesTestJSON-1846886542-project-member] Flavor pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 861.785622] env[61923]: DEBUG nova.virt.hardware [None req-13baf863-98e6-41c7-ae75-48fd1a9cef32 tempest-ServerAddressesTestJSON-1846886542 tempest-ServerAddressesTestJSON-1846886542-project-member] Image pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 861.786402] env[61923]: DEBUG nova.virt.hardware [None req-13baf863-98e6-41c7-ae75-48fd1a9cef32 tempest-ServerAddressesTestJSON-1846886542 tempest-ServerAddressesTestJSON-1846886542-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 861.786818] env[61923]: DEBUG nova.virt.hardware [None req-13baf863-98e6-41c7-ae75-48fd1a9cef32 tempest-ServerAddressesTestJSON-1846886542 tempest-ServerAddressesTestJSON-1846886542-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 861.787261] env[61923]: DEBUG nova.virt.hardware [None req-13baf863-98e6-41c7-ae75-48fd1a9cef32 tempest-ServerAddressesTestJSON-1846886542 tempest-ServerAddressesTestJSON-1846886542-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 861.788689] env[61923]: DEBUG nova.virt.hardware [None req-13baf863-98e6-41c7-ae75-48fd1a9cef32 tempest-ServerAddressesTestJSON-1846886542 tempest-ServerAddressesTestJSON-1846886542-project-member] Got 1 possible topologies {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 861.789735] env[61923]: DEBUG nova.virt.hardware [None req-13baf863-98e6-41c7-ae75-48fd1a9cef32 tempest-ServerAddressesTestJSON-1846886542 tempest-ServerAddressesTestJSON-1846886542-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 861.790251] env[61923]: DEBUG nova.virt.hardware [None req-13baf863-98e6-41c7-ae75-48fd1a9cef32 tempest-ServerAddressesTestJSON-1846886542 tempest-ServerAddressesTestJSON-1846886542-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 861.791728] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24df94be-82fb-4d27-8fab-2904d9552d2b {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.804980] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be5b09db-616a-4c11-b25b-130f7a0f33ef {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.817508] env[61923]: DEBUG oslo_vmware.api [None req-80740a2c-9779-4afc-be30-c4469cd3724d tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] Task: {'id': task-1377629, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.830048] env[61923]: DEBUG nova.network.neutron [None req-cdbe0145-cd3c-4551-af87-0c7f8b53f26c tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] [instance: e413c45d-cd89-44d4-9102-3d188907e7bb] Updating instance_info_cache with network_info: [{"id": "4ed56dd5-beb3-4254-932b-7290c7b85700", "address": "fa:16:3e:c5:9b:44", "network": {"id": "545a170d-946b-4e0b-b9f5-b21c2dcda19d", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-40231173-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6e260342cd1b41ef8f0ceb15b6438daa", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3753f451-fa23-4988-9361-074fb0bd3fd4", "external-id": "nsx-vlan-transportzone-440", "segmentation_id": 440, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4ed56dd5-be", "ovs_interfaceid": "4ed56dd5-beb3-4254-932b-7290c7b85700", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 861.834656] env[61923]: DEBUG oslo_vmware.api [None req-cdbe0145-cd3c-4551-af87-0c7f8b53f26c tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Task: {'id': task-1377630, 'name': Rename_Task, 'duration_secs': 0.185689} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 861.835401] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-cdbe0145-cd3c-4551-af87-0c7f8b53f26c tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] [instance: 444affa5-a7ed-4a17-9015-9fd5724aab64] Powering on the VM {{(pid=61923) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 861.835736] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c7d60e46-2a8e-4459-a213-90a4ec256e58 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.843321] env[61923]: DEBUG oslo_vmware.api [None req-cdbe0145-cd3c-4551-af87-0c7f8b53f26c tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Waiting for the task: (returnval){ [ 861.843321] env[61923]: value = "task-1377631" [ 861.843321] env[61923]: _type = "Task" [ 861.843321] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 861.858659] env[61923]: DEBUG oslo_vmware.api [None req-cdbe0145-cd3c-4551-af87-0c7f8b53f26c tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Task: {'id': task-1377631, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.038476] env[61923]: DEBUG nova.compute.manager [req-57e1f73b-cf6b-4b9f-a94a-4c34b0ea5188 req-f7b639db-7f08-44c2-afe1-f7b4658fd912 service nova] [instance: 5b73e07a-d6ef-4dcf-bdae-eea91d2aeb6f] Received event network-vif-plugged-262b2cfc-fe55-4a4e-bb7c-87f81d8c7e91 {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 862.039146] env[61923]: DEBUG oslo_concurrency.lockutils [req-57e1f73b-cf6b-4b9f-a94a-4c34b0ea5188 req-f7b639db-7f08-44c2-afe1-f7b4658fd912 service nova] Acquiring lock "5b73e07a-d6ef-4dcf-bdae-eea91d2aeb6f-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 862.039146] env[61923]: DEBUG oslo_concurrency.lockutils [req-57e1f73b-cf6b-4b9f-a94a-4c34b0ea5188 req-f7b639db-7f08-44c2-afe1-f7b4658fd912 service nova] Lock "5b73e07a-d6ef-4dcf-bdae-eea91d2aeb6f-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 862.039146] env[61923]: DEBUG oslo_concurrency.lockutils [req-57e1f73b-cf6b-4b9f-a94a-4c34b0ea5188 req-f7b639db-7f08-44c2-afe1-f7b4658fd912 service nova] Lock "5b73e07a-d6ef-4dcf-bdae-eea91d2aeb6f-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 862.039733] env[61923]: DEBUG nova.compute.manager [req-57e1f73b-cf6b-4b9f-a94a-4c34b0ea5188 req-f7b639db-7f08-44c2-afe1-f7b4658fd912 service nova] [instance: 5b73e07a-d6ef-4dcf-bdae-eea91d2aeb6f] No waiting events found dispatching network-vif-plugged-262b2cfc-fe55-4a4e-bb7c-87f81d8c7e91 {{(pid=61923) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 862.039733] env[61923]: WARNING nova.compute.manager [req-57e1f73b-cf6b-4b9f-a94a-4c34b0ea5188 req-f7b639db-7f08-44c2-afe1-f7b4658fd912 service nova] [instance: 5b73e07a-d6ef-4dcf-bdae-eea91d2aeb6f] Received unexpected event network-vif-plugged-262b2cfc-fe55-4a4e-bb7c-87f81d8c7e91 for instance with vm_state building and task_state spawning. [ 862.143413] env[61923]: DEBUG nova.network.neutron [None req-13baf863-98e6-41c7-ae75-48fd1a9cef32 tempest-ServerAddressesTestJSON-1846886542 tempest-ServerAddressesTestJSON-1846886542-project-member] [instance: 5b73e07a-d6ef-4dcf-bdae-eea91d2aeb6f] Successfully updated port: 262b2cfc-fe55-4a4e-bb7c-87f81d8c7e91 {{(pid=61923) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 862.151220] env[61923]: DEBUG oslo_concurrency.lockutils [None req-00703b42-f309-4672-8dc7-10f4a69426aa tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.438s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 862.151688] env[61923]: DEBUG nova.compute.manager [None req-00703b42-f309-4672-8dc7-10f4a69426aa tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] [instance: b779d183-89ae-4e4d-ae99-e514e145ed43] Start building networks asynchronously for instance. {{(pid=61923) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 862.155405] env[61923]: DEBUG oslo_concurrency.lockutils [None req-62f42696-8467-4135-b6f9-a4440e44fead tempest-ServerShowV257Test-986496257 tempest-ServerShowV257Test-986496257-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 13.829s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 862.155405] env[61923]: DEBUG nova.objects.instance [None req-62f42696-8467-4135-b6f9-a4440e44fead tempest-ServerShowV257Test-986496257 tempest-ServerShowV257Test-986496257-project-member] Lazy-loading 'resources' on Instance uuid b8c60af7-3561-468c-a9a5-1c645080fc69 {{(pid=61923) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 862.156297] env[61923]: DEBUG oslo_concurrency.lockutils [None req-5fbf0568-2a85-4e3c-9b63-7d91ecc77593 tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Lock "2719569b-8572-4199-8158-7bb367d17dc5" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 46.482s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 862.309954] env[61923]: DEBUG oslo_vmware.api [None req-80740a2c-9779-4afc-be30-c4469cd3724d tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] Task: {'id': task-1377629, 'name': PowerOnVM_Task, 'duration_secs': 0.589204} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 862.310292] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-80740a2c-9779-4afc-be30-c4469cd3724d tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] [instance: fc4fdf89-4615-4586-9dcd-ea90dc258361] Powered on the VM {{(pid=61923) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 862.310514] env[61923]: INFO nova.compute.manager [None req-80740a2c-9779-4afc-be30-c4469cd3724d tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] [instance: fc4fdf89-4615-4586-9dcd-ea90dc258361] Took 17.70 seconds to spawn the instance on the hypervisor. [ 862.310699] env[61923]: DEBUG nova.compute.manager [None req-80740a2c-9779-4afc-be30-c4469cd3724d tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] [instance: fc4fdf89-4615-4586-9dcd-ea90dc258361] Checking state {{(pid=61923) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 862.311571] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de06207a-d1d1-480d-9993-5dc60a7d6187 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.337035] env[61923]: DEBUG oslo_concurrency.lockutils [None req-cdbe0145-cd3c-4551-af87-0c7f8b53f26c tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Releasing lock "refresh_cache-e413c45d-cd89-44d4-9102-3d188907e7bb" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 862.337035] env[61923]: DEBUG nova.compute.manager [None req-cdbe0145-cd3c-4551-af87-0c7f8b53f26c tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] [instance: e413c45d-cd89-44d4-9102-3d188907e7bb] Instance network_info: |[{"id": "4ed56dd5-beb3-4254-932b-7290c7b85700", "address": "fa:16:3e:c5:9b:44", "network": {"id": "545a170d-946b-4e0b-b9f5-b21c2dcda19d", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-40231173-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6e260342cd1b41ef8f0ceb15b6438daa", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3753f451-fa23-4988-9361-074fb0bd3fd4", "external-id": "nsx-vlan-transportzone-440", "segmentation_id": 440, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4ed56dd5-be", "ovs_interfaceid": "4ed56dd5-beb3-4254-932b-7290c7b85700", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61923) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 862.337035] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-cdbe0145-cd3c-4551-af87-0c7f8b53f26c tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] [instance: e413c45d-cd89-44d4-9102-3d188907e7bb] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:c5:9b:44', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '3753f451-fa23-4988-9361-074fb0bd3fd4', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '4ed56dd5-beb3-4254-932b-7290c7b85700', 'vif_model': 'vmxnet3'}] {{(pid=61923) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 862.347548] env[61923]: DEBUG oslo.service.loopingcall [None req-cdbe0145-cd3c-4551-af87-0c7f8b53f26c tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61923) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 862.347873] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e413c45d-cd89-44d4-9102-3d188907e7bb] Creating VM on the ESX host {{(pid=61923) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 862.355436] env[61923]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-1fb455af-cded-496d-a826-f982d2c52a77 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.373907] env[61923]: DEBUG nova.compute.manager [req-5a9ad83d-9fb2-42c6-8749-bef60a895e36 req-2f70dbdb-8e92-4d11-a440-10f18c00d3a0 service nova] [instance: e413c45d-cd89-44d4-9102-3d188907e7bb] Received event network-changed-4ed56dd5-beb3-4254-932b-7290c7b85700 {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 862.374587] env[61923]: DEBUG nova.compute.manager [req-5a9ad83d-9fb2-42c6-8749-bef60a895e36 req-2f70dbdb-8e92-4d11-a440-10f18c00d3a0 service nova] [instance: e413c45d-cd89-44d4-9102-3d188907e7bb] Refreshing instance network info cache due to event network-changed-4ed56dd5-beb3-4254-932b-7290c7b85700. {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 862.374682] env[61923]: DEBUG oslo_concurrency.lockutils [req-5a9ad83d-9fb2-42c6-8749-bef60a895e36 req-2f70dbdb-8e92-4d11-a440-10f18c00d3a0 service nova] Acquiring lock "refresh_cache-e413c45d-cd89-44d4-9102-3d188907e7bb" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 862.374781] env[61923]: DEBUG oslo_concurrency.lockutils [req-5a9ad83d-9fb2-42c6-8749-bef60a895e36 req-2f70dbdb-8e92-4d11-a440-10f18c00d3a0 service nova] Acquired lock "refresh_cache-e413c45d-cd89-44d4-9102-3d188907e7bb" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 862.375102] env[61923]: DEBUG nova.network.neutron [req-5a9ad83d-9fb2-42c6-8749-bef60a895e36 req-2f70dbdb-8e92-4d11-a440-10f18c00d3a0 service nova] [instance: e413c45d-cd89-44d4-9102-3d188907e7bb] Refreshing network info cache for port 4ed56dd5-beb3-4254-932b-7290c7b85700 {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 862.383022] env[61923]: DEBUG oslo_vmware.api [None req-cdbe0145-cd3c-4551-af87-0c7f8b53f26c tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Task: {'id': task-1377631, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.385229] env[61923]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 862.385229] env[61923]: value = "task-1377632" [ 862.385229] env[61923]: _type = "Task" [ 862.385229] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 862.396191] env[61923]: DEBUG oslo_vmware.api [-] Task: {'id': task-1377632, 'name': CreateVM_Task} progress is 5%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.646133] env[61923]: DEBUG oslo_concurrency.lockutils [None req-13baf863-98e6-41c7-ae75-48fd1a9cef32 tempest-ServerAddressesTestJSON-1846886542 tempest-ServerAddressesTestJSON-1846886542-project-member] Acquiring lock "refresh_cache-5b73e07a-d6ef-4dcf-bdae-eea91d2aeb6f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 862.646383] env[61923]: DEBUG oslo_concurrency.lockutils [None req-13baf863-98e6-41c7-ae75-48fd1a9cef32 tempest-ServerAddressesTestJSON-1846886542 tempest-ServerAddressesTestJSON-1846886542-project-member] Acquired lock "refresh_cache-5b73e07a-d6ef-4dcf-bdae-eea91d2aeb6f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 862.646553] env[61923]: DEBUG nova.network.neutron [None req-13baf863-98e6-41c7-ae75-48fd1a9cef32 tempest-ServerAddressesTestJSON-1846886542 tempest-ServerAddressesTestJSON-1846886542-project-member] [instance: 5b73e07a-d6ef-4dcf-bdae-eea91d2aeb6f] Building network info cache for instance {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 862.661355] env[61923]: DEBUG nova.compute.utils [None req-00703b42-f309-4672-8dc7-10f4a69426aa tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Using /dev/sd instead of None {{(pid=61923) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 862.667358] env[61923]: DEBUG nova.compute.manager [None req-00703b42-f309-4672-8dc7-10f4a69426aa tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] [instance: b779d183-89ae-4e4d-ae99-e514e145ed43] Allocating IP information in the background. {{(pid=61923) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 862.667358] env[61923]: DEBUG nova.network.neutron [None req-00703b42-f309-4672-8dc7-10f4a69426aa tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] [instance: b779d183-89ae-4e4d-ae99-e514e145ed43] allocate_for_instance() {{(pid=61923) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 862.717273] env[61923]: DEBUG nova.policy [None req-00703b42-f309-4672-8dc7-10f4a69426aa tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ced04976ec034e09a0ad122827292901', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a13d827933fa4597984afdb91dbbdd39', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61923) authorize /opt/stack/nova/nova/policy.py:201}} [ 862.831308] env[61923]: INFO nova.compute.manager [None req-80740a2c-9779-4afc-be30-c4469cd3724d tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] [instance: fc4fdf89-4615-4586-9dcd-ea90dc258361] Took 37.93 seconds to build instance. [ 862.860529] env[61923]: DEBUG oslo_vmware.api [None req-cdbe0145-cd3c-4551-af87-0c7f8b53f26c tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Task: {'id': task-1377631, 'name': PowerOnVM_Task, 'duration_secs': 0.690364} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 862.860813] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-cdbe0145-cd3c-4551-af87-0c7f8b53f26c tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] [instance: 444affa5-a7ed-4a17-9015-9fd5724aab64] Powered on the VM {{(pid=61923) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 862.861046] env[61923]: INFO nova.compute.manager [None req-cdbe0145-cd3c-4551-af87-0c7f8b53f26c tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] [instance: 444affa5-a7ed-4a17-9015-9fd5724aab64] Took 8.71 seconds to spawn the instance on the hypervisor. [ 862.861238] env[61923]: DEBUG nova.compute.manager [None req-cdbe0145-cd3c-4551-af87-0c7f8b53f26c tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] [instance: 444affa5-a7ed-4a17-9015-9fd5724aab64] Checking state {{(pid=61923) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 862.862134] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0256caf-e463-4e21-9643-81684c6f02c7 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.897321] env[61923]: DEBUG oslo_vmware.api [-] Task: {'id': task-1377632, 'name': CreateVM_Task} progress is 99%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.913633] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee05724b-b53a-4833-920d-787426acad36 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.921646] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0a22fcb-8f9a-4b03-8c8e-cefd11d50c53 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.954478] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-056fe3d0-ced0-47e1-bc76-f9aa79b99211 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.963536] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b2fc87a-3ccb-42a5-b3a9-8787eb84f304 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.980683] env[61923]: DEBUG nova.compute.provider_tree [None req-62f42696-8467-4135-b6f9-a4440e44fead tempest-ServerShowV257Test-986496257 tempest-ServerShowV257Test-986496257-project-member] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 863.043170] env[61923]: DEBUG nova.network.neutron [None req-00703b42-f309-4672-8dc7-10f4a69426aa tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] [instance: b779d183-89ae-4e4d-ae99-e514e145ed43] Successfully created port: d49a3010-c84a-4bef-beed-9ca4aa7d5dc2 {{(pid=61923) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 863.166509] env[61923]: DEBUG nova.compute.manager [None req-00703b42-f309-4672-8dc7-10f4a69426aa tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] [instance: b779d183-89ae-4e4d-ae99-e514e145ed43] Start building block device mappings for instance. {{(pid=61923) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 863.169828] env[61923]: DEBUG oslo_concurrency.lockutils [None req-144d62d8-7418-4330-9258-261ec048af9e tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] Acquiring lock "fc4fdf89-4615-4586-9dcd-ea90dc258361" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 863.187783] env[61923]: DEBUG nova.network.neutron [None req-13baf863-98e6-41c7-ae75-48fd1a9cef32 tempest-ServerAddressesTestJSON-1846886542 tempest-ServerAddressesTestJSON-1846886542-project-member] [instance: 5b73e07a-d6ef-4dcf-bdae-eea91d2aeb6f] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 863.268366] env[61923]: DEBUG nova.network.neutron [req-5a9ad83d-9fb2-42c6-8749-bef60a895e36 req-2f70dbdb-8e92-4d11-a440-10f18c00d3a0 service nova] [instance: e413c45d-cd89-44d4-9102-3d188907e7bb] Updated VIF entry in instance network info cache for port 4ed56dd5-beb3-4254-932b-7290c7b85700. {{(pid=61923) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 863.268745] env[61923]: DEBUG nova.network.neutron [req-5a9ad83d-9fb2-42c6-8749-bef60a895e36 req-2f70dbdb-8e92-4d11-a440-10f18c00d3a0 service nova] [instance: e413c45d-cd89-44d4-9102-3d188907e7bb] Updating instance_info_cache with network_info: [{"id": "4ed56dd5-beb3-4254-932b-7290c7b85700", "address": "fa:16:3e:c5:9b:44", "network": {"id": "545a170d-946b-4e0b-b9f5-b21c2dcda19d", "bridge": "br-int", "label": "tempest-MultipleCreateTestJSON-40231173-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "6e260342cd1b41ef8f0ceb15b6438daa", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3753f451-fa23-4988-9361-074fb0bd3fd4", "external-id": "nsx-vlan-transportzone-440", "segmentation_id": 440, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4ed56dd5-be", "ovs_interfaceid": "4ed56dd5-beb3-4254-932b-7290c7b85700", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 863.333814] env[61923]: DEBUG oslo_concurrency.lockutils [None req-80740a2c-9779-4afc-be30-c4469cd3724d tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] Lock "fc4fdf89-4615-4586-9dcd-ea90dc258361" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 91.209s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 863.333814] env[61923]: DEBUG oslo_concurrency.lockutils [None req-144d62d8-7418-4330-9258-261ec048af9e tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] Lock "fc4fdf89-4615-4586-9dcd-ea90dc258361" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.164s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 863.333814] env[61923]: DEBUG oslo_concurrency.lockutils [None req-144d62d8-7418-4330-9258-261ec048af9e tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] Acquiring lock "fc4fdf89-4615-4586-9dcd-ea90dc258361-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 863.334044] env[61923]: DEBUG oslo_concurrency.lockutils [None req-144d62d8-7418-4330-9258-261ec048af9e tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] Lock "fc4fdf89-4615-4586-9dcd-ea90dc258361-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 863.334081] env[61923]: DEBUG oslo_concurrency.lockutils [None req-144d62d8-7418-4330-9258-261ec048af9e tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] Lock "fc4fdf89-4615-4586-9dcd-ea90dc258361-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 863.336107] env[61923]: INFO nova.compute.manager [None req-144d62d8-7418-4330-9258-261ec048af9e tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] [instance: fc4fdf89-4615-4586-9dcd-ea90dc258361] Terminating instance [ 863.338056] env[61923]: DEBUG nova.compute.manager [None req-144d62d8-7418-4330-9258-261ec048af9e tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] [instance: fc4fdf89-4615-4586-9dcd-ea90dc258361] Start destroying the instance on the hypervisor. {{(pid=61923) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 863.338149] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-144d62d8-7418-4330-9258-261ec048af9e tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] [instance: fc4fdf89-4615-4586-9dcd-ea90dc258361] Destroying instance {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 863.339240] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-650044f3-d006-4104-9824-eff53e253293 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.347388] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-144d62d8-7418-4330-9258-261ec048af9e tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] [instance: fc4fdf89-4615-4586-9dcd-ea90dc258361] Powering off the VM {{(pid=61923) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 863.348574] env[61923]: DEBUG nova.network.neutron [None req-13baf863-98e6-41c7-ae75-48fd1a9cef32 tempest-ServerAddressesTestJSON-1846886542 tempest-ServerAddressesTestJSON-1846886542-project-member] [instance: 5b73e07a-d6ef-4dcf-bdae-eea91d2aeb6f] Updating instance_info_cache with network_info: [{"id": "262b2cfc-fe55-4a4e-bb7c-87f81d8c7e91", "address": "fa:16:3e:b6:c3:a4", "network": {"id": "fc27730b-1770-4d57-88be-96476321caa0", "bridge": "br-int", "label": "tempest-ServerAddressesTestJSON-225581996-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a69fe23e6b4c4526a488b2be81e6882e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c5652322-9f10-4996-baed-4c0aa13a1b4e", "external-id": "nsx-vlan-transportzone-941", "segmentation_id": 941, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap262b2cfc-fe", "ovs_interfaceid": "262b2cfc-fe55-4a4e-bb7c-87f81d8c7e91", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 863.350506] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-be42dbca-17b4-478f-9023-d1837adf28d6 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.356630] env[61923]: DEBUG oslo_vmware.api [None req-144d62d8-7418-4330-9258-261ec048af9e tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] Waiting for the task: (returnval){ [ 863.356630] env[61923]: value = "task-1377633" [ 863.356630] env[61923]: _type = "Task" [ 863.356630] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 863.365034] env[61923]: DEBUG oslo_vmware.api [None req-144d62d8-7418-4330-9258-261ec048af9e tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] Task: {'id': task-1377633, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.382240] env[61923]: INFO nova.compute.manager [None req-cdbe0145-cd3c-4551-af87-0c7f8b53f26c tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] [instance: 444affa5-a7ed-4a17-9015-9fd5724aab64] Took 23.30 seconds to build instance. [ 863.398260] env[61923]: DEBUG oslo_vmware.api [-] Task: {'id': task-1377632, 'name': CreateVM_Task, 'duration_secs': 0.638768} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 863.398486] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e413c45d-cd89-44d4-9102-3d188907e7bb] Created VM on the ESX host {{(pid=61923) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 863.399232] env[61923]: DEBUG oslo_concurrency.lockutils [None req-cdbe0145-cd3c-4551-af87-0c7f8b53f26c tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 863.399421] env[61923]: DEBUG oslo_concurrency.lockutils [None req-cdbe0145-cd3c-4551-af87-0c7f8b53f26c tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 863.399769] env[61923]: DEBUG oslo_concurrency.lockutils [None req-cdbe0145-cd3c-4551-af87-0c7f8b53f26c tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 863.400049] env[61923]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b5a8c359-c954-499b-8d5f-815c415f816b {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.405259] env[61923]: DEBUG oslo_vmware.api [None req-cdbe0145-cd3c-4551-af87-0c7f8b53f26c tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Waiting for the task: (returnval){ [ 863.405259] env[61923]: value = "session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52738df2-b2b5-6fcd-1891-a60e86957ddb" [ 863.405259] env[61923]: _type = "Task" [ 863.405259] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 863.415742] env[61923]: DEBUG oslo_vmware.api [None req-cdbe0145-cd3c-4551-af87-0c7f8b53f26c tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52738df2-b2b5-6fcd-1891-a60e86957ddb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.483516] env[61923]: DEBUG nova.scheduler.client.report [None req-62f42696-8467-4135-b6f9-a4440e44fead tempest-ServerShowV257Test-986496257 tempest-ServerShowV257Test-986496257-project-member] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 863.772276] env[61923]: DEBUG oslo_concurrency.lockutils [req-5a9ad83d-9fb2-42c6-8749-bef60a895e36 req-2f70dbdb-8e92-4d11-a440-10f18c00d3a0 service nova] Releasing lock "refresh_cache-e413c45d-cd89-44d4-9102-3d188907e7bb" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 863.825420] env[61923]: DEBUG nova.compute.manager [None req-12656253-9def-48f0-9500-888c645ef4dc tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] [instance: 87ae37dc-d0d0-4b76-afdd-0ba3e8f6ce0b] Checking state {{(pid=61923) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 863.825420] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bea00661-2db6-4fe1-b43c-0095361b07d7 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.852995] env[61923]: DEBUG oslo_concurrency.lockutils [None req-13baf863-98e6-41c7-ae75-48fd1a9cef32 tempest-ServerAddressesTestJSON-1846886542 tempest-ServerAddressesTestJSON-1846886542-project-member] Releasing lock "refresh_cache-5b73e07a-d6ef-4dcf-bdae-eea91d2aeb6f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 863.853353] env[61923]: DEBUG nova.compute.manager [None req-13baf863-98e6-41c7-ae75-48fd1a9cef32 tempest-ServerAddressesTestJSON-1846886542 tempest-ServerAddressesTestJSON-1846886542-project-member] [instance: 5b73e07a-d6ef-4dcf-bdae-eea91d2aeb6f] Instance network_info: |[{"id": "262b2cfc-fe55-4a4e-bb7c-87f81d8c7e91", "address": "fa:16:3e:b6:c3:a4", "network": {"id": "fc27730b-1770-4d57-88be-96476321caa0", "bridge": "br-int", "label": "tempest-ServerAddressesTestJSON-225581996-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a69fe23e6b4c4526a488b2be81e6882e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c5652322-9f10-4996-baed-4c0aa13a1b4e", "external-id": "nsx-vlan-transportzone-941", "segmentation_id": 941, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap262b2cfc-fe", "ovs_interfaceid": "262b2cfc-fe55-4a4e-bb7c-87f81d8c7e91", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61923) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 863.855020] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-13baf863-98e6-41c7-ae75-48fd1a9cef32 tempest-ServerAddressesTestJSON-1846886542 tempest-ServerAddressesTestJSON-1846886542-project-member] [instance: 5b73e07a-d6ef-4dcf-bdae-eea91d2aeb6f] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:b6:c3:a4', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'c5652322-9f10-4996-baed-4c0aa13a1b4e', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '262b2cfc-fe55-4a4e-bb7c-87f81d8c7e91', 'vif_model': 'vmxnet3'}] {{(pid=61923) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 863.862438] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-13baf863-98e6-41c7-ae75-48fd1a9cef32 tempest-ServerAddressesTestJSON-1846886542 tempest-ServerAddressesTestJSON-1846886542-project-member] Creating folder: Project (a69fe23e6b4c4526a488b2be81e6882e). Parent ref: group-v292629. {{(pid=61923) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 863.862764] env[61923]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-da94f86c-049a-4c10-bd30-c97f9b8b1563 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.879685] env[61923]: DEBUG oslo_vmware.api [None req-144d62d8-7418-4330-9258-261ec048af9e tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] Task: {'id': task-1377633, 'name': PowerOffVM_Task, 'duration_secs': 0.236456} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 863.880685] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-144d62d8-7418-4330-9258-261ec048af9e tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] [instance: fc4fdf89-4615-4586-9dcd-ea90dc258361] Powered off the VM {{(pid=61923) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 863.880873] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-144d62d8-7418-4330-9258-261ec048af9e tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] [instance: fc4fdf89-4615-4586-9dcd-ea90dc258361] Unregistering the VM {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 863.881295] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-835c3aa1-190e-46d9-8fea-3cd64bf12184 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.886016] env[61923]: DEBUG oslo_concurrency.lockutils [None req-cdbe0145-cd3c-4551-af87-0c7f8b53f26c tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Lock "444affa5-a7ed-4a17-9015-9fd5724aab64" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 42.329s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 863.886016] env[61923]: INFO nova.virt.vmwareapi.vm_util [None req-13baf863-98e6-41c7-ae75-48fd1a9cef32 tempest-ServerAddressesTestJSON-1846886542 tempest-ServerAddressesTestJSON-1846886542-project-member] Created folder: Project (a69fe23e6b4c4526a488b2be81e6882e) in parent group-v292629. [ 863.886016] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-13baf863-98e6-41c7-ae75-48fd1a9cef32 tempest-ServerAddressesTestJSON-1846886542 tempest-ServerAddressesTestJSON-1846886542-project-member] Creating folder: Instances. Parent ref: group-v292688. {{(pid=61923) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 863.886438] env[61923]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-1227ceae-7174-489b-ba42-999620e3d04f {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.895923] env[61923]: INFO nova.virt.vmwareapi.vm_util [None req-13baf863-98e6-41c7-ae75-48fd1a9cef32 tempest-ServerAddressesTestJSON-1846886542 tempest-ServerAddressesTestJSON-1846886542-project-member] Created folder: Instances in parent group-v292688. [ 863.896199] env[61923]: DEBUG oslo.service.loopingcall [None req-13baf863-98e6-41c7-ae75-48fd1a9cef32 tempest-ServerAddressesTestJSON-1846886542 tempest-ServerAddressesTestJSON-1846886542-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61923) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 863.897299] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5b73e07a-d6ef-4dcf-bdae-eea91d2aeb6f] Creating VM on the ESX host {{(pid=61923) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 863.897299] env[61923]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-5afe3074-707c-4a64-8fdd-0417ed9dfc00 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.933850] env[61923]: DEBUG oslo_vmware.api [None req-cdbe0145-cd3c-4551-af87-0c7f8b53f26c tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52738df2-b2b5-6fcd-1891-a60e86957ddb, 'name': SearchDatastore_Task, 'duration_secs': 0.014202} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 863.935699] env[61923]: DEBUG oslo_concurrency.lockutils [None req-cdbe0145-cd3c-4551-af87-0c7f8b53f26c tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 863.936070] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-cdbe0145-cd3c-4551-af87-0c7f8b53f26c tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] [instance: e413c45d-cd89-44d4-9102-3d188907e7bb] Processing image 0f153f63-ae0a-45b1-b7f5-7f9b673c947f {{(pid=61923) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 863.936453] env[61923]: DEBUG oslo_concurrency.lockutils [None req-cdbe0145-cd3c-4551-af87-0c7f8b53f26c tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 863.936679] env[61923]: DEBUG oslo_concurrency.lockutils [None req-cdbe0145-cd3c-4551-af87-0c7f8b53f26c tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 863.936927] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-cdbe0145-cd3c-4551-af87-0c7f8b53f26c tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61923) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 863.937288] env[61923]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 863.937288] env[61923]: value = "task-1377637" [ 863.937288] env[61923]: _type = "Task" [ 863.937288] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 863.937478] env[61923]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-507326a4-b070-4173-b282-c0625f1d24af {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.947669] env[61923]: DEBUG oslo_vmware.api [-] Task: {'id': task-1377637, 'name': CreateVM_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.949463] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-cdbe0145-cd3c-4551-af87-0c7f8b53f26c tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61923) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 863.949665] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-cdbe0145-cd3c-4551-af87-0c7f8b53f26c tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61923) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 863.950474] env[61923]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-359e728f-b696-497d-adfa-4baeae77db84 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.955540] env[61923]: DEBUG oslo_vmware.api [None req-cdbe0145-cd3c-4551-af87-0c7f8b53f26c tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Waiting for the task: (returnval){ [ 863.955540] env[61923]: value = "session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]5274cf1a-51f9-a1c8-3a13-8fd73598e046" [ 863.955540] env[61923]: _type = "Task" [ 863.955540] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 863.963820] env[61923]: DEBUG oslo_vmware.api [None req-cdbe0145-cd3c-4551-af87-0c7f8b53f26c tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]5274cf1a-51f9-a1c8-3a13-8fd73598e046, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.991642] env[61923]: DEBUG oslo_concurrency.lockutils [None req-62f42696-8467-4135-b6f9-a4440e44fead tempest-ServerShowV257Test-986496257 tempest-ServerShowV257Test-986496257-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.837s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 863.998033] env[61923]: DEBUG oslo_concurrency.lockutils [None req-f39c3939-2621-40b9-927d-ae5b7aafe54d tempest-ServerTagsTestJSON-19215696 tempest-ServerTagsTestJSON-19215696-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 14.640s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 863.998033] env[61923]: DEBUG nova.objects.instance [None req-f39c3939-2621-40b9-927d-ae5b7aafe54d tempest-ServerTagsTestJSON-19215696 tempest-ServerTagsTestJSON-19215696-project-member] Lazy-loading 'resources' on Instance uuid 48825592-dcdc-41d5-9fbf-500d1f31017b {{(pid=61923) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 864.020861] env[61923]: INFO nova.scheduler.client.report [None req-62f42696-8467-4135-b6f9-a4440e44fead tempest-ServerShowV257Test-986496257 tempest-ServerShowV257Test-986496257-project-member] Deleted allocations for instance b8c60af7-3561-468c-a9a5-1c645080fc69 [ 864.030523] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-144d62d8-7418-4330-9258-261ec048af9e tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] [instance: fc4fdf89-4615-4586-9dcd-ea90dc258361] Unregistered the VM {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 864.030523] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-144d62d8-7418-4330-9258-261ec048af9e tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] [instance: fc4fdf89-4615-4586-9dcd-ea90dc258361] Deleting contents of the VM from datastore datastore1 {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 864.030523] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-144d62d8-7418-4330-9258-261ec048af9e tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] Deleting the datastore file [datastore1] fc4fdf89-4615-4586-9dcd-ea90dc258361 {{(pid=61923) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 864.033092] env[61923]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-1a569fa4-0f76-4ec2-bb9a-6ed8853b7408 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.038484] env[61923]: DEBUG oslo_vmware.api [None req-144d62d8-7418-4330-9258-261ec048af9e tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] Waiting for the task: (returnval){ [ 864.038484] env[61923]: value = "task-1377638" [ 864.038484] env[61923]: _type = "Task" [ 864.038484] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 864.048249] env[61923]: DEBUG oslo_vmware.api [None req-144d62d8-7418-4330-9258-261ec048af9e tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] Task: {'id': task-1377638, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.132624] env[61923]: DEBUG nova.compute.manager [req-c7850cb1-3015-44b4-af49-9afe9a31f755 req-ee46d7f5-981f-4f05-8286-4f2f5bc9c985 service nova] [instance: 5b73e07a-d6ef-4dcf-bdae-eea91d2aeb6f] Received event network-changed-262b2cfc-fe55-4a4e-bb7c-87f81d8c7e91 {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 864.132737] env[61923]: DEBUG nova.compute.manager [req-c7850cb1-3015-44b4-af49-9afe9a31f755 req-ee46d7f5-981f-4f05-8286-4f2f5bc9c985 service nova] [instance: 5b73e07a-d6ef-4dcf-bdae-eea91d2aeb6f] Refreshing instance network info cache due to event network-changed-262b2cfc-fe55-4a4e-bb7c-87f81d8c7e91. {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 864.133864] env[61923]: DEBUG oslo_concurrency.lockutils [req-c7850cb1-3015-44b4-af49-9afe9a31f755 req-ee46d7f5-981f-4f05-8286-4f2f5bc9c985 service nova] Acquiring lock "refresh_cache-5b73e07a-d6ef-4dcf-bdae-eea91d2aeb6f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 864.133864] env[61923]: DEBUG oslo_concurrency.lockutils [req-c7850cb1-3015-44b4-af49-9afe9a31f755 req-ee46d7f5-981f-4f05-8286-4f2f5bc9c985 service nova] Acquired lock "refresh_cache-5b73e07a-d6ef-4dcf-bdae-eea91d2aeb6f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 864.133864] env[61923]: DEBUG nova.network.neutron [req-c7850cb1-3015-44b4-af49-9afe9a31f755 req-ee46d7f5-981f-4f05-8286-4f2f5bc9c985 service nova] [instance: 5b73e07a-d6ef-4dcf-bdae-eea91d2aeb6f] Refreshing network info cache for port 262b2cfc-fe55-4a4e-bb7c-87f81d8c7e91 {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 864.177407] env[61923]: DEBUG nova.compute.manager [None req-00703b42-f309-4672-8dc7-10f4a69426aa tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] [instance: b779d183-89ae-4e4d-ae99-e514e145ed43] Start spawning the instance on the hypervisor. {{(pid=61923) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 864.207292] env[61923]: DEBUG nova.virt.hardware [None req-00703b42-f309-4672-8dc7-10f4a69426aa tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-29T20:07:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-29T20:07:35Z,direct_url=,disk_format='vmdk',id=0f153f63-ae0a-45b1-b7f5-7f9b673c947f,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='4e7b5e3b3c3443c8bd5c70f8a15739f5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-29T20:07:36Z,virtual_size=,visibility=), allow threads: False {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 864.207625] env[61923]: DEBUG nova.virt.hardware [None req-00703b42-f309-4672-8dc7-10f4a69426aa tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Flavor limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 864.207849] env[61923]: DEBUG nova.virt.hardware [None req-00703b42-f309-4672-8dc7-10f4a69426aa tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Image limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 864.208160] env[61923]: DEBUG nova.virt.hardware [None req-00703b42-f309-4672-8dc7-10f4a69426aa tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Flavor pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 864.208395] env[61923]: DEBUG nova.virt.hardware [None req-00703b42-f309-4672-8dc7-10f4a69426aa tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Image pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 864.208612] env[61923]: DEBUG nova.virt.hardware [None req-00703b42-f309-4672-8dc7-10f4a69426aa tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 864.208975] env[61923]: DEBUG nova.virt.hardware [None req-00703b42-f309-4672-8dc7-10f4a69426aa tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 864.209236] env[61923]: DEBUG nova.virt.hardware [None req-00703b42-f309-4672-8dc7-10f4a69426aa tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 864.209521] env[61923]: DEBUG nova.virt.hardware [None req-00703b42-f309-4672-8dc7-10f4a69426aa tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Got 1 possible topologies {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 864.209783] env[61923]: DEBUG nova.virt.hardware [None req-00703b42-f309-4672-8dc7-10f4a69426aa tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 864.210021] env[61923]: DEBUG nova.virt.hardware [None req-00703b42-f309-4672-8dc7-10f4a69426aa tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 864.211410] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64f32bd3-cdfd-4cd7-bd7c-3ebd52f39f4f {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.223560] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52ddffee-9ad3-4c8f-a326-9ad7d20d9ef4 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.336562] env[61923]: INFO nova.compute.manager [None req-12656253-9def-48f0-9500-888c645ef4dc tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] [instance: 87ae37dc-d0d0-4b76-afdd-0ba3e8f6ce0b] instance snapshotting [ 864.339796] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5270ade6-397a-4d43-a79f-be16b61fe5fe {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.361239] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e050aac7-a48e-480a-afbc-b0815e2b87e4 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.449930] env[61923]: DEBUG oslo_vmware.api [-] Task: {'id': task-1377637, 'name': CreateVM_Task, 'duration_secs': 0.412469} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 864.450189] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5b73e07a-d6ef-4dcf-bdae-eea91d2aeb6f] Created VM on the ESX host {{(pid=61923) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 864.450928] env[61923]: DEBUG oslo_concurrency.lockutils [None req-13baf863-98e6-41c7-ae75-48fd1a9cef32 tempest-ServerAddressesTestJSON-1846886542 tempest-ServerAddressesTestJSON-1846886542-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 864.451162] env[61923]: DEBUG oslo_concurrency.lockutils [None req-13baf863-98e6-41c7-ae75-48fd1a9cef32 tempest-ServerAddressesTestJSON-1846886542 tempest-ServerAddressesTestJSON-1846886542-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 864.451593] env[61923]: DEBUG oslo_concurrency.lockutils [None req-13baf863-98e6-41c7-ae75-48fd1a9cef32 tempest-ServerAddressesTestJSON-1846886542 tempest-ServerAddressesTestJSON-1846886542-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 864.451883] env[61923]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f2122b7d-3724-413a-a850-fd85378e002c {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.456706] env[61923]: DEBUG oslo_vmware.api [None req-13baf863-98e6-41c7-ae75-48fd1a9cef32 tempest-ServerAddressesTestJSON-1846886542 tempest-ServerAddressesTestJSON-1846886542-project-member] Waiting for the task: (returnval){ [ 864.456706] env[61923]: value = "session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52fec434-592d-35fc-e291-e6eac865e156" [ 864.456706] env[61923]: _type = "Task" [ 864.456706] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 864.467882] env[61923]: DEBUG oslo_vmware.api [None req-cdbe0145-cd3c-4551-af87-0c7f8b53f26c tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]5274cf1a-51f9-a1c8-3a13-8fd73598e046, 'name': SearchDatastore_Task, 'duration_secs': 0.009955} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 864.471449] env[61923]: DEBUG oslo_vmware.api [None req-13baf863-98e6-41c7-ae75-48fd1a9cef32 tempest-ServerAddressesTestJSON-1846886542 tempest-ServerAddressesTestJSON-1846886542-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52fec434-592d-35fc-e291-e6eac865e156, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.471675] env[61923]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ed3a4e62-95e4-4665-99e9-f0cf55678fae {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.477104] env[61923]: DEBUG oslo_vmware.api [None req-cdbe0145-cd3c-4551-af87-0c7f8b53f26c tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Waiting for the task: (returnval){ [ 864.477104] env[61923]: value = "session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52acd244-6456-71c4-38be-21ebd1ffb310" [ 864.477104] env[61923]: _type = "Task" [ 864.477104] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 864.484863] env[61923]: DEBUG oslo_vmware.api [None req-cdbe0145-cd3c-4551-af87-0c7f8b53f26c tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52acd244-6456-71c4-38be-21ebd1ffb310, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.529036] env[61923]: DEBUG oslo_concurrency.lockutils [None req-62f42696-8467-4135-b6f9-a4440e44fead tempest-ServerShowV257Test-986496257 tempest-ServerShowV257Test-986496257-project-member] Lock "b8c60af7-3561-468c-a9a5-1c645080fc69" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 20.018s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 864.550934] env[61923]: DEBUG oslo_vmware.api [None req-144d62d8-7418-4330-9258-261ec048af9e tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] Task: {'id': task-1377638, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.251933} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 864.551239] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-144d62d8-7418-4330-9258-261ec048af9e tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] Deleted the datastore file {{(pid=61923) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 864.551511] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-144d62d8-7418-4330-9258-261ec048af9e tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] [instance: fc4fdf89-4615-4586-9dcd-ea90dc258361] Deleted contents of the VM from datastore datastore1 {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 864.551627] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-144d62d8-7418-4330-9258-261ec048af9e tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] [instance: fc4fdf89-4615-4586-9dcd-ea90dc258361] Instance destroyed {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 864.551845] env[61923]: INFO nova.compute.manager [None req-144d62d8-7418-4330-9258-261ec048af9e tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] [instance: fc4fdf89-4615-4586-9dcd-ea90dc258361] Took 1.21 seconds to destroy the instance on the hypervisor. [ 864.552054] env[61923]: DEBUG oslo.service.loopingcall [None req-144d62d8-7418-4330-9258-261ec048af9e tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61923) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 864.552246] env[61923]: DEBUG nova.compute.manager [-] [instance: fc4fdf89-4615-4586-9dcd-ea90dc258361] Deallocating network for instance {{(pid=61923) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 864.552339] env[61923]: DEBUG nova.network.neutron [-] [instance: fc4fdf89-4615-4586-9dcd-ea90dc258361] deallocate_for_instance() {{(pid=61923) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 864.679459] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74fa9475-c202-42e3-a412-7408960a9c5f {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.691102] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0cad330c-e256-4015-ab53-5b1e075058e3 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.728108] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-85702654-18ac-4ef8-bfbb-e2dac3a31e83 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.739325] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec240f41-cba6-4655-a6ca-08ae7bb88175 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.753718] env[61923]: DEBUG nova.compute.provider_tree [None req-f39c3939-2621-40b9-927d-ae5b7aafe54d tempest-ServerTagsTestJSON-19215696 tempest-ServerTagsTestJSON-19215696-project-member] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 864.875025] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-12656253-9def-48f0-9500-888c645ef4dc tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] [instance: 87ae37dc-d0d0-4b76-afdd-0ba3e8f6ce0b] Creating Snapshot of the VM instance {{(pid=61923) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 864.875025] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-64bc3696-bac4-403b-b64f-8e8540ae8987 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.882401] env[61923]: DEBUG oslo_vmware.api [None req-12656253-9def-48f0-9500-888c645ef4dc tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Waiting for the task: (returnval){ [ 864.882401] env[61923]: value = "task-1377639" [ 864.882401] env[61923]: _type = "Task" [ 864.882401] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 864.894895] env[61923]: DEBUG oslo_vmware.api [None req-12656253-9def-48f0-9500-888c645ef4dc tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Task: {'id': task-1377639, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.967565] env[61923]: DEBUG nova.compute.manager [req-5b7405db-16ff-4a4a-b1ea-f515b61d4868 req-f1cc69e0-0395-44dd-8554-b3add8ecc82b service nova] [instance: fc4fdf89-4615-4586-9dcd-ea90dc258361] Received event network-vif-deleted-8781f580-cc8d-4677-b6b3-2e71a0b907bc {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 864.967772] env[61923]: INFO nova.compute.manager [req-5b7405db-16ff-4a4a-b1ea-f515b61d4868 req-f1cc69e0-0395-44dd-8554-b3add8ecc82b service nova] [instance: fc4fdf89-4615-4586-9dcd-ea90dc258361] Neutron deleted interface 8781f580-cc8d-4677-b6b3-2e71a0b907bc; detaching it from the instance and deleting it from the info cache [ 864.968070] env[61923]: DEBUG nova.network.neutron [req-5b7405db-16ff-4a4a-b1ea-f515b61d4868 req-f1cc69e0-0395-44dd-8554-b3add8ecc82b service nova] [instance: fc4fdf89-4615-4586-9dcd-ea90dc258361] Updating instance_info_cache with network_info: [{"id": "a36809ea-2e4a-48ff-ac61-305da772b282", "address": "fa:16:3e:78:7e:ef", "network": {"id": "3adb8073-ba83-4101-a1fa-aaef57320dbf", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-563610619", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.15", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "c304b5b5d4df485fbda55a570494a62f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a485857d-7086-4dcf-9d65-d0dcd177fcb0", "external-id": "nsx-vlan-transportzone-232", "segmentation_id": 232, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa36809ea-2e", "ovs_interfaceid": "a36809ea-2e4a-48ff-ac61-305da772b282", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "0738a978-9439-4919-8894-4b264480593e", "address": "fa:16:3e:37:4a:a2", "network": {"id": "982ad408-258d-4c40-a80e-fb1c8b012c16", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1379323635", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.167", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c304b5b5d4df485fbda55a570494a62f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "32faf59b-014c-4f1f-8331-40df95bf741f", "external-id": "nsx-vlan-transportzone-996", "segmentation_id": 996, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0738a978-94", "ovs_interfaceid": "0738a978-9439-4919-8894-4b264480593e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 864.975975] env[61923]: DEBUG oslo_vmware.api [None req-13baf863-98e6-41c7-ae75-48fd1a9cef32 tempest-ServerAddressesTestJSON-1846886542 tempest-ServerAddressesTestJSON-1846886542-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52fec434-592d-35fc-e291-e6eac865e156, 'name': SearchDatastore_Task, 'duration_secs': 0.017203} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 864.976653] env[61923]: DEBUG oslo_concurrency.lockutils [None req-13baf863-98e6-41c7-ae75-48fd1a9cef32 tempest-ServerAddressesTestJSON-1846886542 tempest-ServerAddressesTestJSON-1846886542-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 864.976817] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-13baf863-98e6-41c7-ae75-48fd1a9cef32 tempest-ServerAddressesTestJSON-1846886542 tempest-ServerAddressesTestJSON-1846886542-project-member] [instance: 5b73e07a-d6ef-4dcf-bdae-eea91d2aeb6f] Processing image 0f153f63-ae0a-45b1-b7f5-7f9b673c947f {{(pid=61923) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 864.977586] env[61923]: DEBUG oslo_concurrency.lockutils [None req-13baf863-98e6-41c7-ae75-48fd1a9cef32 tempest-ServerAddressesTestJSON-1846886542 tempest-ServerAddressesTestJSON-1846886542-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 864.987629] env[61923]: DEBUG oslo_vmware.api [None req-cdbe0145-cd3c-4551-af87-0c7f8b53f26c tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52acd244-6456-71c4-38be-21ebd1ffb310, 'name': SearchDatastore_Task, 'duration_secs': 0.017107} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 864.988668] env[61923]: DEBUG oslo_concurrency.lockutils [None req-cdbe0145-cd3c-4551-af87-0c7f8b53f26c tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 864.988870] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-cdbe0145-cd3c-4551-af87-0c7f8b53f26c tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk to [datastore2] e413c45d-cd89-44d4-9102-3d188907e7bb/e413c45d-cd89-44d4-9102-3d188907e7bb.vmdk {{(pid=61923) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 864.989258] env[61923]: DEBUG oslo_concurrency.lockutils [None req-13baf863-98e6-41c7-ae75-48fd1a9cef32 tempest-ServerAddressesTestJSON-1846886542 tempest-ServerAddressesTestJSON-1846886542-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 864.989514] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-13baf863-98e6-41c7-ae75-48fd1a9cef32 tempest-ServerAddressesTestJSON-1846886542 tempest-ServerAddressesTestJSON-1846886542-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61923) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 864.989944] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-68aabd43-168e-4f1b-90eb-3d0bf30a7a92 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.993387] env[61923]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-7e9e1328-2de7-46ff-b473-5f61bc95a590 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.003602] env[61923]: DEBUG oslo_vmware.api [None req-cdbe0145-cd3c-4551-af87-0c7f8b53f26c tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Waiting for the task: (returnval){ [ 865.003602] env[61923]: value = "task-1377640" [ 865.003602] env[61923]: _type = "Task" [ 865.003602] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 865.008134] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-13baf863-98e6-41c7-ae75-48fd1a9cef32 tempest-ServerAddressesTestJSON-1846886542 tempest-ServerAddressesTestJSON-1846886542-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61923) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 865.008250] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-13baf863-98e6-41c7-ae75-48fd1a9cef32 tempest-ServerAddressesTestJSON-1846886542 tempest-ServerAddressesTestJSON-1846886542-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61923) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 865.010131] env[61923]: DEBUG nova.network.neutron [req-c7850cb1-3015-44b4-af49-9afe9a31f755 req-ee46d7f5-981f-4f05-8286-4f2f5bc9c985 service nova] [instance: 5b73e07a-d6ef-4dcf-bdae-eea91d2aeb6f] Updated VIF entry in instance network info cache for port 262b2cfc-fe55-4a4e-bb7c-87f81d8c7e91. {{(pid=61923) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 865.010713] env[61923]: DEBUG nova.network.neutron [req-c7850cb1-3015-44b4-af49-9afe9a31f755 req-ee46d7f5-981f-4f05-8286-4f2f5bc9c985 service nova] [instance: 5b73e07a-d6ef-4dcf-bdae-eea91d2aeb6f] Updating instance_info_cache with network_info: [{"id": "262b2cfc-fe55-4a4e-bb7c-87f81d8c7e91", "address": "fa:16:3e:b6:c3:a4", "network": {"id": "fc27730b-1770-4d57-88be-96476321caa0", "bridge": "br-int", "label": "tempest-ServerAddressesTestJSON-225581996-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a69fe23e6b4c4526a488b2be81e6882e", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c5652322-9f10-4996-baed-4c0aa13a1b4e", "external-id": "nsx-vlan-transportzone-941", "segmentation_id": 941, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap262b2cfc-fe", "ovs_interfaceid": "262b2cfc-fe55-4a4e-bb7c-87f81d8c7e91", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 865.012505] env[61923]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-758a5aca-d009-4e38-b86d-01d19a3623dd {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.022765] env[61923]: DEBUG oslo_vmware.api [None req-cdbe0145-cd3c-4551-af87-0c7f8b53f26c tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Task: {'id': task-1377640, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.024383] env[61923]: DEBUG oslo_vmware.api [None req-13baf863-98e6-41c7-ae75-48fd1a9cef32 tempest-ServerAddressesTestJSON-1846886542 tempest-ServerAddressesTestJSON-1846886542-project-member] Waiting for the task: (returnval){ [ 865.024383] env[61923]: value = "session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52ff7c14-303d-9bd7-7248-d6e9406386c7" [ 865.024383] env[61923]: _type = "Task" [ 865.024383] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 865.034962] env[61923]: DEBUG oslo_vmware.api [None req-13baf863-98e6-41c7-ae75-48fd1a9cef32 tempest-ServerAddressesTestJSON-1846886542 tempest-ServerAddressesTestJSON-1846886542-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52ff7c14-303d-9bd7-7248-d6e9406386c7, 'name': SearchDatastore_Task, 'duration_secs': 0.011591} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 865.035830] env[61923]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-84c1da09-1e56-443e-93f1-a285cb26e16c {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.042371] env[61923]: DEBUG oslo_vmware.api [None req-13baf863-98e6-41c7-ae75-48fd1a9cef32 tempest-ServerAddressesTestJSON-1846886542 tempest-ServerAddressesTestJSON-1846886542-project-member] Waiting for the task: (returnval){ [ 865.042371] env[61923]: value = "session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52a2b139-21f2-0675-4294-96d58bd78de6" [ 865.042371] env[61923]: _type = "Task" [ 865.042371] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 865.053754] env[61923]: DEBUG oslo_vmware.api [None req-13baf863-98e6-41c7-ae75-48fd1a9cef32 tempest-ServerAddressesTestJSON-1846886542 tempest-ServerAddressesTestJSON-1846886542-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52a2b139-21f2-0675-4294-96d58bd78de6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.087334] env[61923]: DEBUG nova.network.neutron [None req-00703b42-f309-4672-8dc7-10f4a69426aa tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] [instance: b779d183-89ae-4e4d-ae99-e514e145ed43] Successfully updated port: d49a3010-c84a-4bef-beed-9ca4aa7d5dc2 {{(pid=61923) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 865.257411] env[61923]: DEBUG nova.scheduler.client.report [None req-f39c3939-2621-40b9-927d-ae5b7aafe54d tempest-ServerTagsTestJSON-19215696 tempest-ServerTagsTestJSON-19215696-project-member] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 865.394835] env[61923]: DEBUG oslo_vmware.api [None req-12656253-9def-48f0-9500-888c645ef4dc tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Task: {'id': task-1377639, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.472168] env[61923]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-7bb6a2fd-41a5-4e6f-96fc-3fc8fe34c89d {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.490638] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-238c1f6c-3084-4a4c-b1c6-cc54cb016213 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.519425] env[61923]: DEBUG oslo_concurrency.lockutils [req-c7850cb1-3015-44b4-af49-9afe9a31f755 req-ee46d7f5-981f-4f05-8286-4f2f5bc9c985 service nova] Releasing lock "refresh_cache-5b73e07a-d6ef-4dcf-bdae-eea91d2aeb6f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 865.520101] env[61923]: DEBUG nova.compute.manager [req-5b7405db-16ff-4a4a-b1ea-f515b61d4868 req-f1cc69e0-0395-44dd-8554-b3add8ecc82b service nova] [instance: fc4fdf89-4615-4586-9dcd-ea90dc258361] Detach interface failed, port_id=8781f580-cc8d-4677-b6b3-2e71a0b907bc, reason: Instance fc4fdf89-4615-4586-9dcd-ea90dc258361 could not be found. {{(pid=61923) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 865.523901] env[61923]: DEBUG oslo_vmware.api [None req-cdbe0145-cd3c-4551-af87-0c7f8b53f26c tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Task: {'id': task-1377640, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.553762] env[61923]: DEBUG oslo_vmware.api [None req-13baf863-98e6-41c7-ae75-48fd1a9cef32 tempest-ServerAddressesTestJSON-1846886542 tempest-ServerAddressesTestJSON-1846886542-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52a2b139-21f2-0675-4294-96d58bd78de6, 'name': SearchDatastore_Task, 'duration_secs': 0.011606} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 865.554194] env[61923]: DEBUG oslo_concurrency.lockutils [None req-13baf863-98e6-41c7-ae75-48fd1a9cef32 tempest-ServerAddressesTestJSON-1846886542 tempest-ServerAddressesTestJSON-1846886542-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 865.554346] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-13baf863-98e6-41c7-ae75-48fd1a9cef32 tempest-ServerAddressesTestJSON-1846886542 tempest-ServerAddressesTestJSON-1846886542-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk to [datastore2] 5b73e07a-d6ef-4dcf-bdae-eea91d2aeb6f/5b73e07a-d6ef-4dcf-bdae-eea91d2aeb6f.vmdk {{(pid=61923) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 865.554627] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-85ab7541-7bbf-4269-a741-6fbf917a4734 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.561714] env[61923]: DEBUG oslo_vmware.api [None req-13baf863-98e6-41c7-ae75-48fd1a9cef32 tempest-ServerAddressesTestJSON-1846886542 tempest-ServerAddressesTestJSON-1846886542-project-member] Waiting for the task: (returnval){ [ 865.561714] env[61923]: value = "task-1377641" [ 865.561714] env[61923]: _type = "Task" [ 865.561714] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 865.570531] env[61923]: DEBUG oslo_vmware.api [None req-13baf863-98e6-41c7-ae75-48fd1a9cef32 tempest-ServerAddressesTestJSON-1846886542 tempest-ServerAddressesTestJSON-1846886542-project-member] Task: {'id': task-1377641, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.591483] env[61923]: DEBUG oslo_concurrency.lockutils [None req-00703b42-f309-4672-8dc7-10f4a69426aa tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Acquiring lock "refresh_cache-b779d183-89ae-4e4d-ae99-e514e145ed43" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 865.591700] env[61923]: DEBUG oslo_concurrency.lockutils [None req-00703b42-f309-4672-8dc7-10f4a69426aa tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Acquired lock "refresh_cache-b779d183-89ae-4e4d-ae99-e514e145ed43" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 865.591772] env[61923]: DEBUG nova.network.neutron [None req-00703b42-f309-4672-8dc7-10f4a69426aa tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] [instance: b779d183-89ae-4e4d-ae99-e514e145ed43] Building network info cache for instance {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 865.763024] env[61923]: DEBUG oslo_concurrency.lockutils [None req-f39c3939-2621-40b9-927d-ae5b7aafe54d tempest-ServerTagsTestJSON-19215696 tempest-ServerTagsTestJSON-19215696-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.768s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 865.767190] env[61923]: DEBUG oslo_concurrency.lockutils [None req-9887b108-13fe-4987-ad6f-1ef139e2a15d tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.968s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 865.770300] env[61923]: INFO nova.compute.claims [None req-9887b108-13fe-4987-ad6f-1ef139e2a15d tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] [instance: 6012d3b5-739c-4762-9bb4-09c51171dcd7] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 865.798291] env[61923]: INFO nova.scheduler.client.report [None req-f39c3939-2621-40b9-927d-ae5b7aafe54d tempest-ServerTagsTestJSON-19215696 tempest-ServerTagsTestJSON-19215696-project-member] Deleted allocations for instance 48825592-dcdc-41d5-9fbf-500d1f31017b [ 865.894721] env[61923]: DEBUG oslo_vmware.api [None req-12656253-9def-48f0-9500-888c645ef4dc tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Task: {'id': task-1377639, 'name': CreateSnapshot_Task, 'duration_secs': 0.816755} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 865.895024] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-12656253-9def-48f0-9500-888c645ef4dc tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] [instance: 87ae37dc-d0d0-4b76-afdd-0ba3e8f6ce0b] Created Snapshot of the VM instance {{(pid=61923) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 865.895909] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f65fae3-afd7-4c17-923a-600d55ea1d21 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.016143] env[61923]: DEBUG oslo_vmware.api [None req-cdbe0145-cd3c-4551-af87-0c7f8b53f26c tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Task: {'id': task-1377640, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.526322} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 866.016460] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-cdbe0145-cd3c-4551-af87-0c7f8b53f26c tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk to [datastore2] e413c45d-cd89-44d4-9102-3d188907e7bb/e413c45d-cd89-44d4-9102-3d188907e7bb.vmdk {{(pid=61923) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 866.016699] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-cdbe0145-cd3c-4551-af87-0c7f8b53f26c tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] [instance: e413c45d-cd89-44d4-9102-3d188907e7bb] Extending root virtual disk to 1048576 {{(pid=61923) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 866.016941] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-9c193762-c6ae-43de-b92a-25dbb427d900 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.030206] env[61923]: DEBUG oslo_vmware.api [None req-cdbe0145-cd3c-4551-af87-0c7f8b53f26c tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Waiting for the task: (returnval){ [ 866.030206] env[61923]: value = "task-1377642" [ 866.030206] env[61923]: _type = "Task" [ 866.030206] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 866.039044] env[61923]: DEBUG oslo_vmware.api [None req-cdbe0145-cd3c-4551-af87-0c7f8b53f26c tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Task: {'id': task-1377642, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.073193] env[61923]: DEBUG oslo_vmware.api [None req-13baf863-98e6-41c7-ae75-48fd1a9cef32 tempest-ServerAddressesTestJSON-1846886542 tempest-ServerAddressesTestJSON-1846886542-project-member] Task: {'id': task-1377641, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.127500] env[61923]: DEBUG nova.network.neutron [None req-00703b42-f309-4672-8dc7-10f4a69426aa tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] [instance: b779d183-89ae-4e4d-ae99-e514e145ed43] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 866.178171] env[61923]: DEBUG nova.compute.manager [req-62f6b981-57b6-4557-a87a-40ed593864c7 req-3c067aa4-2313-4d4f-addd-05e72b6aad26 service nova] [instance: b779d183-89ae-4e4d-ae99-e514e145ed43] Received event network-vif-plugged-d49a3010-c84a-4bef-beed-9ca4aa7d5dc2 {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 866.178418] env[61923]: DEBUG oslo_concurrency.lockutils [req-62f6b981-57b6-4557-a87a-40ed593864c7 req-3c067aa4-2313-4d4f-addd-05e72b6aad26 service nova] Acquiring lock "b779d183-89ae-4e4d-ae99-e514e145ed43-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 866.179070] env[61923]: DEBUG oslo_concurrency.lockutils [req-62f6b981-57b6-4557-a87a-40ed593864c7 req-3c067aa4-2313-4d4f-addd-05e72b6aad26 service nova] Lock "b779d183-89ae-4e4d-ae99-e514e145ed43-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 866.179274] env[61923]: DEBUG oslo_concurrency.lockutils [req-62f6b981-57b6-4557-a87a-40ed593864c7 req-3c067aa4-2313-4d4f-addd-05e72b6aad26 service nova] Lock "b779d183-89ae-4e4d-ae99-e514e145ed43-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 866.179452] env[61923]: DEBUG nova.compute.manager [req-62f6b981-57b6-4557-a87a-40ed593864c7 req-3c067aa4-2313-4d4f-addd-05e72b6aad26 service nova] [instance: b779d183-89ae-4e4d-ae99-e514e145ed43] No waiting events found dispatching network-vif-plugged-d49a3010-c84a-4bef-beed-9ca4aa7d5dc2 {{(pid=61923) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 866.179623] env[61923]: WARNING nova.compute.manager [req-62f6b981-57b6-4557-a87a-40ed593864c7 req-3c067aa4-2313-4d4f-addd-05e72b6aad26 service nova] [instance: b779d183-89ae-4e4d-ae99-e514e145ed43] Received unexpected event network-vif-plugged-d49a3010-c84a-4bef-beed-9ca4aa7d5dc2 for instance with vm_state building and task_state spawning. [ 866.179789] env[61923]: DEBUG nova.compute.manager [req-62f6b981-57b6-4557-a87a-40ed593864c7 req-3c067aa4-2313-4d4f-addd-05e72b6aad26 service nova] [instance: b779d183-89ae-4e4d-ae99-e514e145ed43] Received event network-changed-d49a3010-c84a-4bef-beed-9ca4aa7d5dc2 {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 866.179954] env[61923]: DEBUG nova.compute.manager [req-62f6b981-57b6-4557-a87a-40ed593864c7 req-3c067aa4-2313-4d4f-addd-05e72b6aad26 service nova] [instance: b779d183-89ae-4e4d-ae99-e514e145ed43] Refreshing instance network info cache due to event network-changed-d49a3010-c84a-4bef-beed-9ca4aa7d5dc2. {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 866.180144] env[61923]: DEBUG oslo_concurrency.lockutils [req-62f6b981-57b6-4557-a87a-40ed593864c7 req-3c067aa4-2313-4d4f-addd-05e72b6aad26 service nova] Acquiring lock "refresh_cache-b779d183-89ae-4e4d-ae99-e514e145ed43" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 866.294829] env[61923]: DEBUG nova.network.neutron [-] [instance: fc4fdf89-4615-4586-9dcd-ea90dc258361] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 866.307269] env[61923]: DEBUG oslo_concurrency.lockutils [None req-f39c3939-2621-40b9-927d-ae5b7aafe54d tempest-ServerTagsTestJSON-19215696 tempest-ServerTagsTestJSON-19215696-project-member] Lock "48825592-dcdc-41d5-9fbf-500d1f31017b" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 20.646s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 866.312335] env[61923]: DEBUG nova.network.neutron [None req-00703b42-f309-4672-8dc7-10f4a69426aa tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] [instance: b779d183-89ae-4e4d-ae99-e514e145ed43] Updating instance_info_cache with network_info: [{"id": "d49a3010-c84a-4bef-beed-9ca4aa7d5dc2", "address": "fa:16:3e:48:e6:a2", "network": {"id": "d2905082-7ca4-4f92-a142-bb85211bb7c4", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-885218622-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a13d827933fa4597984afdb91dbbdd39", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "03ac2c9c-6ad2-4a85-bfab-c7e336df859a", "external-id": "nsx-vlan-transportzone-379", "segmentation_id": 379, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd49a3010-c8", "ovs_interfaceid": "d49a3010-c84a-4bef-beed-9ca4aa7d5dc2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 866.415969] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-12656253-9def-48f0-9500-888c645ef4dc tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] [instance: 87ae37dc-d0d0-4b76-afdd-0ba3e8f6ce0b] Creating linked-clone VM from snapshot {{(pid=61923) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 866.416314] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-b30ddc1f-8427-4e5f-9958-b8708a8b6113 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.424843] env[61923]: DEBUG oslo_vmware.api [None req-12656253-9def-48f0-9500-888c645ef4dc tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Waiting for the task: (returnval){ [ 866.424843] env[61923]: value = "task-1377643" [ 866.424843] env[61923]: _type = "Task" [ 866.424843] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 866.434225] env[61923]: DEBUG oslo_vmware.api [None req-12656253-9def-48f0-9500-888c645ef4dc tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Task: {'id': task-1377643, 'name': CloneVM_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.542662] env[61923]: DEBUG oslo_vmware.api [None req-cdbe0145-cd3c-4551-af87-0c7f8b53f26c tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Task: {'id': task-1377642, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.096583} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 866.543998] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-cdbe0145-cd3c-4551-af87-0c7f8b53f26c tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] [instance: e413c45d-cd89-44d4-9102-3d188907e7bb] Extended root virtual disk {{(pid=61923) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 866.544995] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa284751-894e-4c05-a586-34f4aeb874be {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.572627] env[61923]: DEBUG nova.virt.vmwareapi.volumeops [None req-cdbe0145-cd3c-4551-af87-0c7f8b53f26c tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] [instance: e413c45d-cd89-44d4-9102-3d188907e7bb] Reconfiguring VM instance instance-0000004b to attach disk [datastore2] e413c45d-cd89-44d4-9102-3d188907e7bb/e413c45d-cd89-44d4-9102-3d188907e7bb.vmdk or device None with type sparse {{(pid=61923) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 866.574033] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-db555e7a-78e7-4e02-a3e0-1dcc9279c1d8 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.601515] env[61923]: DEBUG oslo_vmware.api [None req-13baf863-98e6-41c7-ae75-48fd1a9cef32 tempest-ServerAddressesTestJSON-1846886542 tempest-ServerAddressesTestJSON-1846886542-project-member] Task: {'id': task-1377641, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.701616} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 866.603013] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-13baf863-98e6-41c7-ae75-48fd1a9cef32 tempest-ServerAddressesTestJSON-1846886542 tempest-ServerAddressesTestJSON-1846886542-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk to [datastore2] 5b73e07a-d6ef-4dcf-bdae-eea91d2aeb6f/5b73e07a-d6ef-4dcf-bdae-eea91d2aeb6f.vmdk {{(pid=61923) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 866.603300] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-13baf863-98e6-41c7-ae75-48fd1a9cef32 tempest-ServerAddressesTestJSON-1846886542 tempest-ServerAddressesTestJSON-1846886542-project-member] [instance: 5b73e07a-d6ef-4dcf-bdae-eea91d2aeb6f] Extending root virtual disk to 1048576 {{(pid=61923) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 866.603659] env[61923]: DEBUG oslo_vmware.api [None req-cdbe0145-cd3c-4551-af87-0c7f8b53f26c tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Waiting for the task: (returnval){ [ 866.603659] env[61923]: value = "task-1377644" [ 866.603659] env[61923]: _type = "Task" [ 866.603659] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 866.603894] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-68d25cc6-f714-4b43-b337-b107c605c763 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.615151] env[61923]: DEBUG oslo_vmware.api [None req-cdbe0145-cd3c-4551-af87-0c7f8b53f26c tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Task: {'id': task-1377644, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.617019] env[61923]: DEBUG oslo_vmware.api [None req-13baf863-98e6-41c7-ae75-48fd1a9cef32 tempest-ServerAddressesTestJSON-1846886542 tempest-ServerAddressesTestJSON-1846886542-project-member] Waiting for the task: (returnval){ [ 866.617019] env[61923]: value = "task-1377645" [ 866.617019] env[61923]: _type = "Task" [ 866.617019] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 866.627774] env[61923]: DEBUG oslo_vmware.api [None req-13baf863-98e6-41c7-ae75-48fd1a9cef32 tempest-ServerAddressesTestJSON-1846886542 tempest-ServerAddressesTestJSON-1846886542-project-member] Task: {'id': task-1377645, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.798638] env[61923]: INFO nova.compute.manager [-] [instance: fc4fdf89-4615-4586-9dcd-ea90dc258361] Took 2.25 seconds to deallocate network for instance. [ 866.815120] env[61923]: DEBUG oslo_concurrency.lockutils [None req-00703b42-f309-4672-8dc7-10f4a69426aa tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Releasing lock "refresh_cache-b779d183-89ae-4e4d-ae99-e514e145ed43" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 866.815489] env[61923]: DEBUG nova.compute.manager [None req-00703b42-f309-4672-8dc7-10f4a69426aa tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] [instance: b779d183-89ae-4e4d-ae99-e514e145ed43] Instance network_info: |[{"id": "d49a3010-c84a-4bef-beed-9ca4aa7d5dc2", "address": "fa:16:3e:48:e6:a2", "network": {"id": "d2905082-7ca4-4f92-a142-bb85211bb7c4", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-885218622-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a13d827933fa4597984afdb91dbbdd39", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "03ac2c9c-6ad2-4a85-bfab-c7e336df859a", "external-id": "nsx-vlan-transportzone-379", "segmentation_id": 379, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd49a3010-c8", "ovs_interfaceid": "d49a3010-c84a-4bef-beed-9ca4aa7d5dc2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61923) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 866.815839] env[61923]: DEBUG oslo_concurrency.lockutils [req-62f6b981-57b6-4557-a87a-40ed593864c7 req-3c067aa4-2313-4d4f-addd-05e72b6aad26 service nova] Acquired lock "refresh_cache-b779d183-89ae-4e4d-ae99-e514e145ed43" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 866.816031] env[61923]: DEBUG nova.network.neutron [req-62f6b981-57b6-4557-a87a-40ed593864c7 req-3c067aa4-2313-4d4f-addd-05e72b6aad26 service nova] [instance: b779d183-89ae-4e4d-ae99-e514e145ed43] Refreshing network info cache for port d49a3010-c84a-4bef-beed-9ca4aa7d5dc2 {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 866.817312] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-00703b42-f309-4672-8dc7-10f4a69426aa tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] [instance: b779d183-89ae-4e4d-ae99-e514e145ed43] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:48:e6:a2', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '03ac2c9c-6ad2-4a85-bfab-c7e336df859a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'd49a3010-c84a-4bef-beed-9ca4aa7d5dc2', 'vif_model': 'vmxnet3'}] {{(pid=61923) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 866.827432] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-00703b42-f309-4672-8dc7-10f4a69426aa tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Creating folder: Project (a13d827933fa4597984afdb91dbbdd39). Parent ref: group-v292629. {{(pid=61923) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 866.827545] env[61923]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-15dcefdb-1f29-45b1-a405-f14ed0141455 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.841706] env[61923]: INFO nova.virt.vmwareapi.vm_util [None req-00703b42-f309-4672-8dc7-10f4a69426aa tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Created folder: Project (a13d827933fa4597984afdb91dbbdd39) in parent group-v292629. [ 866.841894] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-00703b42-f309-4672-8dc7-10f4a69426aa tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Creating folder: Instances. Parent ref: group-v292693. {{(pid=61923) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 866.842205] env[61923]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-8b4d5e4a-2a9a-487d-af56-7825fe875c39 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.851498] env[61923]: INFO nova.virt.vmwareapi.vm_util [None req-00703b42-f309-4672-8dc7-10f4a69426aa tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Created folder: Instances in parent group-v292693. [ 866.851818] env[61923]: DEBUG oslo.service.loopingcall [None req-00703b42-f309-4672-8dc7-10f4a69426aa tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61923) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 866.852032] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b779d183-89ae-4e4d-ae99-e514e145ed43] Creating VM on the ESX host {{(pid=61923) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 866.852281] env[61923]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-bd9c5028-364f-4e79-85e1-76a9d9ea4254 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.881962] env[61923]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 866.881962] env[61923]: value = "task-1377648" [ 866.881962] env[61923]: _type = "Task" [ 866.881962] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 866.890850] env[61923]: DEBUG oslo_vmware.api [-] Task: {'id': task-1377648, 'name': CreateVM_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.934796] env[61923]: DEBUG oslo_vmware.api [None req-12656253-9def-48f0-9500-888c645ef4dc tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Task: {'id': task-1377643, 'name': CloneVM_Task} progress is 94%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.011176] env[61923]: DEBUG nova.compute.manager [req-2037cbb0-2abe-4044-a1ed-02ac69ec3e19 req-87f20f56-965b-4df1-aefe-8dcb60546511 service nova] [instance: fc4fdf89-4615-4586-9dcd-ea90dc258361] Received event network-vif-deleted-a36809ea-2e4a-48ff-ac61-305da772b282 {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 867.011176] env[61923]: DEBUG nova.compute.manager [req-2037cbb0-2abe-4044-a1ed-02ac69ec3e19 req-87f20f56-965b-4df1-aefe-8dcb60546511 service nova] [instance: fc4fdf89-4615-4586-9dcd-ea90dc258361] Received event network-vif-deleted-0738a978-9439-4919-8894-4b264480593e {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 867.051540] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac0ff9da-6f82-4692-be9d-b2d519b5b0bc {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.060753] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7a85876-ecf0-4beb-b5e2-f211a25e7cfb {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.095309] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63c32357-7c84-43e5-b1ea-d40df6a50002 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.103906] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c716139-726f-489d-a623-35186352847f {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.124024] env[61923]: DEBUG nova.compute.provider_tree [None req-9887b108-13fe-4987-ad6f-1ef139e2a15d tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 867.131852] env[61923]: DEBUG oslo_vmware.api [None req-cdbe0145-cd3c-4551-af87-0c7f8b53f26c tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Task: {'id': task-1377644, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.137286] env[61923]: DEBUG oslo_vmware.api [None req-13baf863-98e6-41c7-ae75-48fd1a9cef32 tempest-ServerAddressesTestJSON-1846886542 tempest-ServerAddressesTestJSON-1846886542-project-member] Task: {'id': task-1377645, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.073488} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 867.138363] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-13baf863-98e6-41c7-ae75-48fd1a9cef32 tempest-ServerAddressesTestJSON-1846886542 tempest-ServerAddressesTestJSON-1846886542-project-member] [instance: 5b73e07a-d6ef-4dcf-bdae-eea91d2aeb6f] Extended root virtual disk {{(pid=61923) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 867.139200] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0f2f368-18c6-43b7-9bc8-3ebf5c5cf5e7 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.164382] env[61923]: DEBUG nova.virt.vmwareapi.volumeops [None req-13baf863-98e6-41c7-ae75-48fd1a9cef32 tempest-ServerAddressesTestJSON-1846886542 tempest-ServerAddressesTestJSON-1846886542-project-member] [instance: 5b73e07a-d6ef-4dcf-bdae-eea91d2aeb6f] Reconfiguring VM instance instance-0000004c to attach disk [datastore2] 5b73e07a-d6ef-4dcf-bdae-eea91d2aeb6f/5b73e07a-d6ef-4dcf-bdae-eea91d2aeb6f.vmdk or device None with type sparse {{(pid=61923) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 867.165351] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-16c6da8f-6090-408c-9761-dd64adc76bb7 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.188072] env[61923]: DEBUG oslo_vmware.api [None req-13baf863-98e6-41c7-ae75-48fd1a9cef32 tempest-ServerAddressesTestJSON-1846886542 tempest-ServerAddressesTestJSON-1846886542-project-member] Waiting for the task: (returnval){ [ 867.188072] env[61923]: value = "task-1377649" [ 867.188072] env[61923]: _type = "Task" [ 867.188072] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 867.200221] env[61923]: DEBUG oslo_vmware.api [None req-13baf863-98e6-41c7-ae75-48fd1a9cef32 tempest-ServerAddressesTestJSON-1846886542 tempest-ServerAddressesTestJSON-1846886542-project-member] Task: {'id': task-1377649, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.310216] env[61923]: DEBUG oslo_concurrency.lockutils [None req-144d62d8-7418-4330-9258-261ec048af9e tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 867.396146] env[61923]: DEBUG oslo_vmware.api [-] Task: {'id': task-1377648, 'name': CreateVM_Task, 'duration_secs': 0.413202} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 867.396413] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b779d183-89ae-4e4d-ae99-e514e145ed43] Created VM on the ESX host {{(pid=61923) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 867.397237] env[61923]: DEBUG oslo_concurrency.lockutils [None req-00703b42-f309-4672-8dc7-10f4a69426aa tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 867.397413] env[61923]: DEBUG oslo_concurrency.lockutils [None req-00703b42-f309-4672-8dc7-10f4a69426aa tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 867.397794] env[61923]: DEBUG oslo_concurrency.lockutils [None req-00703b42-f309-4672-8dc7-10f4a69426aa tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 867.398165] env[61923]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c88093ce-dcd5-492a-925b-8acb6f1f84b7 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.403405] env[61923]: DEBUG oslo_vmware.api [None req-00703b42-f309-4672-8dc7-10f4a69426aa tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Waiting for the task: (returnval){ [ 867.403405] env[61923]: value = "session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]5249a3a9-bde4-82e5-3167-783789c1d40b" [ 867.403405] env[61923]: _type = "Task" [ 867.403405] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 867.412097] env[61923]: DEBUG oslo_vmware.api [None req-00703b42-f309-4672-8dc7-10f4a69426aa tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]5249a3a9-bde4-82e5-3167-783789c1d40b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.436339] env[61923]: DEBUG oslo_vmware.api [None req-12656253-9def-48f0-9500-888c645ef4dc tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Task: {'id': task-1377643, 'name': CloneVM_Task} progress is 94%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.619922] env[61923]: DEBUG oslo_vmware.api [None req-cdbe0145-cd3c-4551-af87-0c7f8b53f26c tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Task: {'id': task-1377644, 'name': ReconfigVM_Task, 'duration_secs': 0.614931} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 867.620341] env[61923]: DEBUG nova.virt.vmwareapi.volumeops [None req-cdbe0145-cd3c-4551-af87-0c7f8b53f26c tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] [instance: e413c45d-cd89-44d4-9102-3d188907e7bb] Reconfigured VM instance instance-0000004b to attach disk [datastore2] e413c45d-cd89-44d4-9102-3d188907e7bb/e413c45d-cd89-44d4-9102-3d188907e7bb.vmdk or device None with type sparse {{(pid=61923) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 867.621196] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-7f9b184f-5c0b-451c-9756-6e5c21bd0044 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.631714] env[61923]: DEBUG oslo_vmware.api [None req-cdbe0145-cd3c-4551-af87-0c7f8b53f26c tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Waiting for the task: (returnval){ [ 867.631714] env[61923]: value = "task-1377650" [ 867.631714] env[61923]: _type = "Task" [ 867.631714] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 867.637842] env[61923]: DEBUG nova.scheduler.client.report [None req-9887b108-13fe-4987-ad6f-1ef139e2a15d tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 867.648797] env[61923]: DEBUG oslo_vmware.api [None req-cdbe0145-cd3c-4551-af87-0c7f8b53f26c tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Task: {'id': task-1377650, 'name': Rename_Task} progress is 10%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.698734] env[61923]: DEBUG oslo_vmware.api [None req-13baf863-98e6-41c7-ae75-48fd1a9cef32 tempest-ServerAddressesTestJSON-1846886542 tempest-ServerAddressesTestJSON-1846886542-project-member] Task: {'id': task-1377649, 'name': ReconfigVM_Task, 'duration_secs': 0.455888} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 867.699178] env[61923]: DEBUG nova.virt.vmwareapi.volumeops [None req-13baf863-98e6-41c7-ae75-48fd1a9cef32 tempest-ServerAddressesTestJSON-1846886542 tempest-ServerAddressesTestJSON-1846886542-project-member] [instance: 5b73e07a-d6ef-4dcf-bdae-eea91d2aeb6f] Reconfigured VM instance instance-0000004c to attach disk [datastore2] 5b73e07a-d6ef-4dcf-bdae-eea91d2aeb6f/5b73e07a-d6ef-4dcf-bdae-eea91d2aeb6f.vmdk or device None with type sparse {{(pid=61923) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 867.699888] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-3a6c83d2-bf72-42e5-83d4-0de6fc9b1026 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.710308] env[61923]: DEBUG oslo_vmware.api [None req-13baf863-98e6-41c7-ae75-48fd1a9cef32 tempest-ServerAddressesTestJSON-1846886542 tempest-ServerAddressesTestJSON-1846886542-project-member] Waiting for the task: (returnval){ [ 867.710308] env[61923]: value = "task-1377651" [ 867.710308] env[61923]: _type = "Task" [ 867.710308] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 867.724664] env[61923]: DEBUG oslo_vmware.api [None req-13baf863-98e6-41c7-ae75-48fd1a9cef32 tempest-ServerAddressesTestJSON-1846886542 tempest-ServerAddressesTestJSON-1846886542-project-member] Task: {'id': task-1377651, 'name': Rename_Task} progress is 6%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.749746] env[61923]: DEBUG nova.network.neutron [req-62f6b981-57b6-4557-a87a-40ed593864c7 req-3c067aa4-2313-4d4f-addd-05e72b6aad26 service nova] [instance: b779d183-89ae-4e4d-ae99-e514e145ed43] Updated VIF entry in instance network info cache for port d49a3010-c84a-4bef-beed-9ca4aa7d5dc2. {{(pid=61923) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 867.750216] env[61923]: DEBUG nova.network.neutron [req-62f6b981-57b6-4557-a87a-40ed593864c7 req-3c067aa4-2313-4d4f-addd-05e72b6aad26 service nova] [instance: b779d183-89ae-4e4d-ae99-e514e145ed43] Updating instance_info_cache with network_info: [{"id": "d49a3010-c84a-4bef-beed-9ca4aa7d5dc2", "address": "fa:16:3e:48:e6:a2", "network": {"id": "d2905082-7ca4-4f92-a142-bb85211bb7c4", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-885218622-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a13d827933fa4597984afdb91dbbdd39", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "03ac2c9c-6ad2-4a85-bfab-c7e336df859a", "external-id": "nsx-vlan-transportzone-379", "segmentation_id": 379, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd49a3010-c8", "ovs_interfaceid": "d49a3010-c84a-4bef-beed-9ca4aa7d5dc2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 867.920076] env[61923]: DEBUG oslo_vmware.api [None req-00703b42-f309-4672-8dc7-10f4a69426aa tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]5249a3a9-bde4-82e5-3167-783789c1d40b, 'name': SearchDatastore_Task, 'duration_secs': 0.013288} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 867.920421] env[61923]: DEBUG oslo_concurrency.lockutils [None req-00703b42-f309-4672-8dc7-10f4a69426aa tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 867.921705] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-00703b42-f309-4672-8dc7-10f4a69426aa tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] [instance: b779d183-89ae-4e4d-ae99-e514e145ed43] Processing image 0f153f63-ae0a-45b1-b7f5-7f9b673c947f {{(pid=61923) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 867.922356] env[61923]: DEBUG oslo_concurrency.lockutils [None req-00703b42-f309-4672-8dc7-10f4a69426aa tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 867.922463] env[61923]: DEBUG oslo_concurrency.lockutils [None req-00703b42-f309-4672-8dc7-10f4a69426aa tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 867.922695] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-00703b42-f309-4672-8dc7-10f4a69426aa tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61923) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 867.923277] env[61923]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-657c95f9-717c-4438-9576-b3cc7bd765f1 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.936336] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-00703b42-f309-4672-8dc7-10f4a69426aa tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61923) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 867.936336] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-00703b42-f309-4672-8dc7-10f4a69426aa tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61923) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 867.937806] env[61923]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-939de0b4-bac4-4f61-b581-a6e43cfd77d4 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.944451] env[61923]: DEBUG oslo_vmware.api [None req-12656253-9def-48f0-9500-888c645ef4dc tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Task: {'id': task-1377643, 'name': CloneVM_Task} progress is 100%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.948391] env[61923]: DEBUG oslo_vmware.api [None req-00703b42-f309-4672-8dc7-10f4a69426aa tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Waiting for the task: (returnval){ [ 867.948391] env[61923]: value = "session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52ed4991-f5fa-414e-7e1a-d86d4054f5b3" [ 867.948391] env[61923]: _type = "Task" [ 867.948391] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 867.957712] env[61923]: DEBUG oslo_vmware.api [None req-00703b42-f309-4672-8dc7-10f4a69426aa tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52ed4991-f5fa-414e-7e1a-d86d4054f5b3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.142026] env[61923]: DEBUG oslo_vmware.api [None req-cdbe0145-cd3c-4551-af87-0c7f8b53f26c tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Task: {'id': task-1377650, 'name': Rename_Task, 'duration_secs': 0.265731} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 868.142384] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-cdbe0145-cd3c-4551-af87-0c7f8b53f26c tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] [instance: e413c45d-cd89-44d4-9102-3d188907e7bb] Powering on the VM {{(pid=61923) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 868.142641] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-885a1ef3-ab0b-4d04-88c7-98989a51d56d {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.144774] env[61923]: DEBUG oslo_concurrency.lockutils [None req-9887b108-13fe-4987-ad6f-1ef139e2a15d tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.378s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 868.145297] env[61923]: DEBUG nova.compute.manager [None req-9887b108-13fe-4987-ad6f-1ef139e2a15d tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] [instance: 6012d3b5-739c-4762-9bb4-09c51171dcd7] Start building networks asynchronously for instance. {{(pid=61923) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 868.149454] env[61923]: DEBUG oslo_concurrency.lockutils [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 15.945s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 868.157644] env[61923]: DEBUG oslo_vmware.api [None req-cdbe0145-cd3c-4551-af87-0c7f8b53f26c tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Waiting for the task: (returnval){ [ 868.157644] env[61923]: value = "task-1377652" [ 868.157644] env[61923]: _type = "Task" [ 868.157644] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 868.176480] env[61923]: DEBUG oslo_vmware.api [None req-cdbe0145-cd3c-4551-af87-0c7f8b53f26c tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Task: {'id': task-1377652, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.222466] env[61923]: DEBUG oslo_vmware.api [None req-13baf863-98e6-41c7-ae75-48fd1a9cef32 tempest-ServerAddressesTestJSON-1846886542 tempest-ServerAddressesTestJSON-1846886542-project-member] Task: {'id': task-1377651, 'name': Rename_Task} progress is 14%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.256013] env[61923]: DEBUG oslo_concurrency.lockutils [req-62f6b981-57b6-4557-a87a-40ed593864c7 req-3c067aa4-2313-4d4f-addd-05e72b6aad26 service nova] Releasing lock "refresh_cache-b779d183-89ae-4e4d-ae99-e514e145ed43" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 868.438897] env[61923]: DEBUG oslo_vmware.api [None req-12656253-9def-48f0-9500-888c645ef4dc tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Task: {'id': task-1377643, 'name': CloneVM_Task, 'duration_secs': 1.597798} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 868.439337] env[61923]: INFO nova.virt.vmwareapi.vmops [None req-12656253-9def-48f0-9500-888c645ef4dc tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] [instance: 87ae37dc-d0d0-4b76-afdd-0ba3e8f6ce0b] Created linked-clone VM from snapshot [ 868.440635] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-efc96045-5e1b-4801-8261-89bf97507aef {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.448837] env[61923]: DEBUG nova.virt.vmwareapi.images [None req-12656253-9def-48f0-9500-888c645ef4dc tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] [instance: 87ae37dc-d0d0-4b76-afdd-0ba3e8f6ce0b] Uploading image cb483b67-0b09-46ef-8d9d-6fb9b7e0f932 {{(pid=61923) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 868.467942] env[61923]: DEBUG oslo_vmware.api [None req-00703b42-f309-4672-8dc7-10f4a69426aa tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52ed4991-f5fa-414e-7e1a-d86d4054f5b3, 'name': SearchDatastore_Task, 'duration_secs': 0.011382} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 868.467942] env[61923]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-995f7136-5fec-4740-8272-e35a3106c115 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.472789] env[61923]: DEBUG oslo_vmware.api [None req-00703b42-f309-4672-8dc7-10f4a69426aa tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Waiting for the task: (returnval){ [ 868.472789] env[61923]: value = "session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]5298a9fd-b76e-48a3-ffd6-c6101d1538e1" [ 868.472789] env[61923]: _type = "Task" [ 868.472789] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 868.482180] env[61923]: DEBUG oslo_vmware.rw_handles [None req-12656253-9def-48f0-9500-888c645ef4dc tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 868.482180] env[61923]: value = "vm-292692" [ 868.482180] env[61923]: _type = "VirtualMachine" [ 868.482180] env[61923]: }. {{(pid=61923) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 868.482739] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-6cba4cec-942d-4b82-8642-f4f29e18c997 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.487605] env[61923]: DEBUG oslo_vmware.api [None req-00703b42-f309-4672-8dc7-10f4a69426aa tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]5298a9fd-b76e-48a3-ffd6-c6101d1538e1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.495352] env[61923]: DEBUG oslo_vmware.rw_handles [None req-12656253-9def-48f0-9500-888c645ef4dc tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Lease: (returnval){ [ 868.495352] env[61923]: value = "session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]520c0e17-40f4-8615-6a88-ed4415d33305" [ 868.495352] env[61923]: _type = "HttpNfcLease" [ 868.495352] env[61923]: } obtained for exporting VM: (result){ [ 868.495352] env[61923]: value = "vm-292692" [ 868.495352] env[61923]: _type = "VirtualMachine" [ 868.495352] env[61923]: }. {{(pid=61923) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 868.495352] env[61923]: DEBUG oslo_vmware.api [None req-12656253-9def-48f0-9500-888c645ef4dc tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Waiting for the lease: (returnval){ [ 868.495352] env[61923]: value = "session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]520c0e17-40f4-8615-6a88-ed4415d33305" [ 868.495352] env[61923]: _type = "HttpNfcLease" [ 868.495352] env[61923]: } to be ready. {{(pid=61923) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 868.501269] env[61923]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 868.501269] env[61923]: value = "session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]520c0e17-40f4-8615-6a88-ed4415d33305" [ 868.501269] env[61923]: _type = "HttpNfcLease" [ 868.501269] env[61923]: } is initializing. {{(pid=61923) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 868.666123] env[61923]: DEBUG nova.compute.utils [None req-9887b108-13fe-4987-ad6f-1ef139e2a15d tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Using /dev/sd instead of None {{(pid=61923) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 868.669750] env[61923]: DEBUG nova.compute.manager [None req-9887b108-13fe-4987-ad6f-1ef139e2a15d tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] [instance: 6012d3b5-739c-4762-9bb4-09c51171dcd7] Allocating IP information in the background. {{(pid=61923) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 868.671283] env[61923]: DEBUG nova.network.neutron [None req-9887b108-13fe-4987-ad6f-1ef139e2a15d tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] [instance: 6012d3b5-739c-4762-9bb4-09c51171dcd7] allocate_for_instance() {{(pid=61923) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 868.690955] env[61923]: DEBUG oslo_vmware.api [None req-cdbe0145-cd3c-4551-af87-0c7f8b53f26c tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Task: {'id': task-1377652, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.726485] env[61923]: DEBUG oslo_vmware.api [None req-13baf863-98e6-41c7-ae75-48fd1a9cef32 tempest-ServerAddressesTestJSON-1846886542 tempest-ServerAddressesTestJSON-1846886542-project-member] Task: {'id': task-1377651, 'name': Rename_Task} progress is 14%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.753197] env[61923]: DEBUG nova.policy [None req-9887b108-13fe-4987-ad6f-1ef139e2a15d tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'cd5039bf8f4344e490df828e643d2779', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '912300a09b80452e85e3cd13c4a644cc', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61923) authorize /opt/stack/nova/nova/policy.py:201}} [ 868.987102] env[61923]: DEBUG oslo_vmware.api [None req-00703b42-f309-4672-8dc7-10f4a69426aa tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]5298a9fd-b76e-48a3-ffd6-c6101d1538e1, 'name': SearchDatastore_Task, 'duration_secs': 0.011051} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 868.987607] env[61923]: DEBUG oslo_concurrency.lockutils [None req-00703b42-f309-4672-8dc7-10f4a69426aa tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 868.987735] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-00703b42-f309-4672-8dc7-10f4a69426aa tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk to [datastore2] b779d183-89ae-4e4d-ae99-e514e145ed43/b779d183-89ae-4e4d-ae99-e514e145ed43.vmdk {{(pid=61923) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 868.988205] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ac4c852a-4920-482d-88e3-6ce738cdbe1b {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.001065] env[61923]: DEBUG oslo_vmware.api [None req-00703b42-f309-4672-8dc7-10f4a69426aa tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Waiting for the task: (returnval){ [ 869.001065] env[61923]: value = "task-1377654" [ 869.001065] env[61923]: _type = "Task" [ 869.001065] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 869.007034] env[61923]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 869.007034] env[61923]: value = "session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]520c0e17-40f4-8615-6a88-ed4415d33305" [ 869.007034] env[61923]: _type = "HttpNfcLease" [ 869.007034] env[61923]: } is ready. {{(pid=61923) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 869.007813] env[61923]: DEBUG oslo_vmware.rw_handles [None req-12656253-9def-48f0-9500-888c645ef4dc tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 869.007813] env[61923]: value = "session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]520c0e17-40f4-8615-6a88-ed4415d33305" [ 869.007813] env[61923]: _type = "HttpNfcLease" [ 869.007813] env[61923]: }. {{(pid=61923) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 869.008923] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cba5386d-c258-4714-a003-873d3661a108 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.017255] env[61923]: DEBUG oslo_vmware.api [None req-00703b42-f309-4672-8dc7-10f4a69426aa tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Task: {'id': task-1377654, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 869.022666] env[61923]: DEBUG oslo_vmware.rw_handles [None req-12656253-9def-48f0-9500-888c645ef4dc tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5235c29c-aba2-46b7-1408-4e88ce392c19/disk-0.vmdk from lease info. {{(pid=61923) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 869.022666] env[61923]: DEBUG oslo_vmware.rw_handles [None req-12656253-9def-48f0-9500-888c645ef4dc tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Opening URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5235c29c-aba2-46b7-1408-4e88ce392c19/disk-0.vmdk for reading. {{(pid=61923) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 869.131126] env[61923]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-ee7d9fda-2e73-43db-b00f-f31e763327fb {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.181613] env[61923]: DEBUG nova.compute.manager [None req-9887b108-13fe-4987-ad6f-1ef139e2a15d tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] [instance: 6012d3b5-739c-4762-9bb4-09c51171dcd7] Start building block device mappings for instance. {{(pid=61923) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 869.195856] env[61923]: DEBUG oslo_vmware.api [None req-cdbe0145-cd3c-4551-af87-0c7f8b53f26c tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Task: {'id': task-1377652, 'name': PowerOnVM_Task, 'duration_secs': 0.887947} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 869.196181] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-cdbe0145-cd3c-4551-af87-0c7f8b53f26c tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] [instance: e413c45d-cd89-44d4-9102-3d188907e7bb] Powered on the VM {{(pid=61923) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 869.196437] env[61923]: INFO nova.compute.manager [None req-cdbe0145-cd3c-4551-af87-0c7f8b53f26c tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] [instance: e413c45d-cd89-44d4-9102-3d188907e7bb] Took 9.87 seconds to spawn the instance on the hypervisor. [ 869.196579] env[61923]: DEBUG nova.compute.manager [None req-cdbe0145-cd3c-4551-af87-0c7f8b53f26c tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] [instance: e413c45d-cd89-44d4-9102-3d188907e7bb] Checking state {{(pid=61923) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 869.198073] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cba80974-8a0f-4f54-979d-184c02c32b13 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.223572] env[61923]: DEBUG nova.compute.resource_tracker [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Instance f1839f61-4314-48fe-8ab6-14b5e49d569d actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61923) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 869.223572] env[61923]: DEBUG nova.compute.resource_tracker [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Instance 422d06d6-6932-46c3-bb25-841e6f65c028 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61923) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 869.223572] env[61923]: WARNING nova.compute.resource_tracker [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Instance 73692517-1816-4e9b-ab2d-8265c683e83d is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 869.223572] env[61923]: DEBUG nova.compute.resource_tracker [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Instance 59198f4d-4dde-4eaf-9f6c-a962cbe53c6e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61923) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 869.225802] env[61923]: WARNING nova.compute.resource_tracker [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Instance fc4fdf89-4615-4586-9dcd-ea90dc258361 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 869.225802] env[61923]: DEBUG nova.compute.resource_tracker [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Instance af3c317a-4007-4cea-a060-1e7dde5ce49e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61923) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 869.225802] env[61923]: DEBUG nova.compute.resource_tracker [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Instance 87ae37dc-d0d0-4b76-afdd-0ba3e8f6ce0b actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61923) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 869.225802] env[61923]: DEBUG nova.compute.resource_tracker [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Instance 2719569b-8572-4199-8158-7bb367d17dc5 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61923) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 869.225802] env[61923]: DEBUG nova.compute.resource_tracker [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Instance 444affa5-a7ed-4a17-9015-9fd5724aab64 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61923) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 869.227077] env[61923]: DEBUG nova.compute.resource_tracker [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Instance e413c45d-cd89-44d4-9102-3d188907e7bb actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61923) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 869.227077] env[61923]: DEBUG nova.compute.resource_tracker [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Instance 5b73e07a-d6ef-4dcf-bdae-eea91d2aeb6f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61923) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 869.227077] env[61923]: DEBUG nova.compute.resource_tracker [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Instance b779d183-89ae-4e4d-ae99-e514e145ed43 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61923) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 869.227077] env[61923]: DEBUG nova.compute.resource_tracker [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Instance 6012d3b5-739c-4762-9bb4-09c51171dcd7 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61923) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 869.227077] env[61923]: DEBUG nova.compute.resource_tracker [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Total usable vcpus: 48, total allocated vcpus: 11 {{(pid=61923) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 869.227077] env[61923]: DEBUG nova.compute.resource_tracker [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=2624MB phys_disk=200GB used_disk=11GB total_vcpus=48 used_vcpus=11 pci_stats=[] {{(pid=61923) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 869.235768] env[61923]: DEBUG oslo_vmware.api [None req-13baf863-98e6-41c7-ae75-48fd1a9cef32 tempest-ServerAddressesTestJSON-1846886542 tempest-ServerAddressesTestJSON-1846886542-project-member] Task: {'id': task-1377651, 'name': Rename_Task, 'duration_secs': 1.087277} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 869.235768] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-13baf863-98e6-41c7-ae75-48fd1a9cef32 tempest-ServerAddressesTestJSON-1846886542 tempest-ServerAddressesTestJSON-1846886542-project-member] [instance: 5b73e07a-d6ef-4dcf-bdae-eea91d2aeb6f] Powering on the VM {{(pid=61923) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 869.235768] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b39882d3-e82a-45af-a153-dd064e2f0aaf {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.242285] env[61923]: DEBUG oslo_vmware.api [None req-13baf863-98e6-41c7-ae75-48fd1a9cef32 tempest-ServerAddressesTestJSON-1846886542 tempest-ServerAddressesTestJSON-1846886542-project-member] Waiting for the task: (returnval){ [ 869.242285] env[61923]: value = "task-1377655" [ 869.242285] env[61923]: _type = "Task" [ 869.242285] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 869.251902] env[61923]: DEBUG oslo_vmware.api [None req-13baf863-98e6-41c7-ae75-48fd1a9cef32 tempest-ServerAddressesTestJSON-1846886542 tempest-ServerAddressesTestJSON-1846886542-project-member] Task: {'id': task-1377655, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 869.261391] env[61923]: DEBUG nova.network.neutron [None req-9887b108-13fe-4987-ad6f-1ef139e2a15d tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] [instance: 6012d3b5-739c-4762-9bb4-09c51171dcd7] Successfully created port: 269a41b0-75e7-46af-9497-8eae3102df11 {{(pid=61923) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 869.489536] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c31ac079-03f1-4453-9045-8deccac36904 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.501460] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97d7d5be-8f26-462e-a7f4-6bcd5ae649e9 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.518855] env[61923]: DEBUG oslo_vmware.api [None req-00703b42-f309-4672-8dc7-10f4a69426aa tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Task: {'id': task-1377654, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 869.554128] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3edbf1c-0b29-40d3-bf41-39c763e441ec {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.565146] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6e35e5d-9049-4de3-b83a-5f50c7028069 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.583061] env[61923]: DEBUG nova.compute.provider_tree [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Updating inventory in ProviderTree for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 177, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 869.719461] env[61923]: INFO nova.compute.manager [None req-cdbe0145-cd3c-4551-af87-0c7f8b53f26c tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] [instance: e413c45d-cd89-44d4-9102-3d188907e7bb] Took 25.25 seconds to build instance. [ 869.758216] env[61923]: DEBUG oslo_vmware.api [None req-13baf863-98e6-41c7-ae75-48fd1a9cef32 tempest-ServerAddressesTestJSON-1846886542 tempest-ServerAddressesTestJSON-1846886542-project-member] Task: {'id': task-1377655, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 870.018856] env[61923]: DEBUG oslo_vmware.api [None req-00703b42-f309-4672-8dc7-10f4a69426aa tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Task: {'id': task-1377654, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.651658} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 870.019797] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-00703b42-f309-4672-8dc7-10f4a69426aa tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk to [datastore2] b779d183-89ae-4e4d-ae99-e514e145ed43/b779d183-89ae-4e4d-ae99-e514e145ed43.vmdk {{(pid=61923) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 870.020239] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-00703b42-f309-4672-8dc7-10f4a69426aa tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] [instance: b779d183-89ae-4e4d-ae99-e514e145ed43] Extending root virtual disk to 1048576 {{(pid=61923) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 870.020660] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-50dc420e-1f2b-49fa-ba21-82d94baf2af0 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.029390] env[61923]: DEBUG oslo_vmware.api [None req-00703b42-f309-4672-8dc7-10f4a69426aa tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Waiting for the task: (returnval){ [ 870.029390] env[61923]: value = "task-1377656" [ 870.029390] env[61923]: _type = "Task" [ 870.029390] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 870.040974] env[61923]: DEBUG oslo_vmware.api [None req-00703b42-f309-4672-8dc7-10f4a69426aa tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Task: {'id': task-1377656, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 870.109024] env[61923]: ERROR nova.scheduler.client.report [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] [req-284dbb6b-3460-4cd9-8ec3-57c34d4428d6] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 177, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID f81803f2-f7f8-4939-a757-a77d34a1d0a2. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-284dbb6b-3460-4cd9-8ec3-57c34d4428d6"}]} [ 870.126699] env[61923]: DEBUG nova.scheduler.client.report [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Refreshing inventories for resource provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 870.148458] env[61923]: DEBUG nova.scheduler.client.report [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Updating ProviderTree inventory for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 870.148458] env[61923]: DEBUG nova.compute.provider_tree [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Updating inventory in ProviderTree for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 870.166031] env[61923]: DEBUG nova.scheduler.client.report [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Refreshing aggregate associations for resource provider f81803f2-f7f8-4939-a757-a77d34a1d0a2, aggregates: None {{(pid=61923) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 870.186496] env[61923]: DEBUG nova.scheduler.client.report [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Refreshing trait associations for resource provider f81803f2-f7f8-4939-a757-a77d34a1d0a2, traits: COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NODE,HW_ARCH_X86_64,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_IMAGE_TYPE_ISO {{(pid=61923) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 870.197928] env[61923]: DEBUG nova.compute.manager [None req-9887b108-13fe-4987-ad6f-1ef139e2a15d tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] [instance: 6012d3b5-739c-4762-9bb4-09c51171dcd7] Start spawning the instance on the hypervisor. {{(pid=61923) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 870.223632] env[61923]: DEBUG oslo_concurrency.lockutils [None req-cdbe0145-cd3c-4551-af87-0c7f8b53f26c tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Lock "e413c45d-cd89-44d4-9102-3d188907e7bb" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 48.639s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 870.258143] env[61923]: DEBUG oslo_vmware.api [None req-13baf863-98e6-41c7-ae75-48fd1a9cef32 tempest-ServerAddressesTestJSON-1846886542 tempest-ServerAddressesTestJSON-1846886542-project-member] Task: {'id': task-1377655, 'name': PowerOnVM_Task, 'duration_secs': 0.869053} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 870.258380] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-13baf863-98e6-41c7-ae75-48fd1a9cef32 tempest-ServerAddressesTestJSON-1846886542 tempest-ServerAddressesTestJSON-1846886542-project-member] [instance: 5b73e07a-d6ef-4dcf-bdae-eea91d2aeb6f] Powered on the VM {{(pid=61923) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 870.258594] env[61923]: INFO nova.compute.manager [None req-13baf863-98e6-41c7-ae75-48fd1a9cef32 tempest-ServerAddressesTestJSON-1846886542 tempest-ServerAddressesTestJSON-1846886542-project-member] [instance: 5b73e07a-d6ef-4dcf-bdae-eea91d2aeb6f] Took 8.50 seconds to spawn the instance on the hypervisor. [ 870.258773] env[61923]: DEBUG nova.compute.manager [None req-13baf863-98e6-41c7-ae75-48fd1a9cef32 tempest-ServerAddressesTestJSON-1846886542 tempest-ServerAddressesTestJSON-1846886542-project-member] [instance: 5b73e07a-d6ef-4dcf-bdae-eea91d2aeb6f] Checking state {{(pid=61923) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 870.259661] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90a7e7d3-ba06-4f91-aeb8-bc47f914d7f7 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.428326] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fea2bc1c-2bed-45a9-b8a4-7a94664960c7 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.438731] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ececaa8d-2ea9-4550-aca9-b5f508622257 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.480032] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d2ced0f-fa2a-4600-a755-18682c0dba6c {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.491278] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f66265ce-50a7-4539-b921-a167bf99c665 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.512784] env[61923]: DEBUG nova.compute.provider_tree [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Updating inventory in ProviderTree for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 177, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 870.546481] env[61923]: DEBUG oslo_vmware.api [None req-00703b42-f309-4672-8dc7-10f4a69426aa tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Task: {'id': task-1377656, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.108593} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 870.547334] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-00703b42-f309-4672-8dc7-10f4a69426aa tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] [instance: b779d183-89ae-4e4d-ae99-e514e145ed43] Extended root virtual disk {{(pid=61923) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 870.548364] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0cfe40e-0cd3-4b53-95af-34c606f33ca9 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.587870] env[61923]: DEBUG nova.virt.vmwareapi.volumeops [None req-00703b42-f309-4672-8dc7-10f4a69426aa tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] [instance: b779d183-89ae-4e4d-ae99-e514e145ed43] Reconfiguring VM instance instance-0000004d to attach disk [datastore2] b779d183-89ae-4e4d-ae99-e514e145ed43/b779d183-89ae-4e4d-ae99-e514e145ed43.vmdk or device None with type sparse {{(pid=61923) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 870.590568] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7286e55b-cd2b-4380-af48-2f1729a32836 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.619026] env[61923]: DEBUG oslo_concurrency.lockutils [None req-6d4df172-3edf-4dbb-b96f-afa60a7b76e1 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Acquiring lock "a511392e-9ab6-42fb-b07c-f90e9786dcc1" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 870.619399] env[61923]: DEBUG oslo_concurrency.lockutils [None req-6d4df172-3edf-4dbb-b96f-afa60a7b76e1 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Lock "a511392e-9ab6-42fb-b07c-f90e9786dcc1" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 870.629051] env[61923]: DEBUG oslo_vmware.api [None req-00703b42-f309-4672-8dc7-10f4a69426aa tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Waiting for the task: (returnval){ [ 870.629051] env[61923]: value = "task-1377657" [ 870.629051] env[61923]: _type = "Task" [ 870.629051] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 870.641020] env[61923]: DEBUG oslo_vmware.api [None req-00703b42-f309-4672-8dc7-10f4a69426aa tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Task: {'id': task-1377657, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 870.783644] env[61923]: INFO nova.compute.manager [None req-13baf863-98e6-41c7-ae75-48fd1a9cef32 tempest-ServerAddressesTestJSON-1846886542 tempest-ServerAddressesTestJSON-1846886542-project-member] [instance: 5b73e07a-d6ef-4dcf-bdae-eea91d2aeb6f] Took 25.03 seconds to build instance. [ 870.911899] env[61923]: DEBUG nova.virt.hardware [None req-9887b108-13fe-4987-ad6f-1ef139e2a15d tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-29T20:07:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-29T20:07:35Z,direct_url=,disk_format='vmdk',id=0f153f63-ae0a-45b1-b7f5-7f9b673c947f,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='4e7b5e3b3c3443c8bd5c70f8a15739f5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-29T20:07:36Z,virtual_size=,visibility=), allow threads: False {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 870.914038] env[61923]: DEBUG nova.virt.hardware [None req-9887b108-13fe-4987-ad6f-1ef139e2a15d tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Flavor limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 870.914038] env[61923]: DEBUG nova.virt.hardware [None req-9887b108-13fe-4987-ad6f-1ef139e2a15d tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Image limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 870.914038] env[61923]: DEBUG nova.virt.hardware [None req-9887b108-13fe-4987-ad6f-1ef139e2a15d tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Flavor pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 870.914038] env[61923]: DEBUG nova.virt.hardware [None req-9887b108-13fe-4987-ad6f-1ef139e2a15d tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Image pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 870.914038] env[61923]: DEBUG nova.virt.hardware [None req-9887b108-13fe-4987-ad6f-1ef139e2a15d tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 870.914038] env[61923]: DEBUG nova.virt.hardware [None req-9887b108-13fe-4987-ad6f-1ef139e2a15d tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 870.914038] env[61923]: DEBUG nova.virt.hardware [None req-9887b108-13fe-4987-ad6f-1ef139e2a15d tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 870.914038] env[61923]: DEBUG nova.virt.hardware [None req-9887b108-13fe-4987-ad6f-1ef139e2a15d tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Got 1 possible topologies {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 870.914038] env[61923]: DEBUG nova.virt.hardware [None req-9887b108-13fe-4987-ad6f-1ef139e2a15d tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 870.914038] env[61923]: DEBUG nova.virt.hardware [None req-9887b108-13fe-4987-ad6f-1ef139e2a15d tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 870.916125] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee1323c0-2ed0-4d79-b099-dfe795d4c40c {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.926912] env[61923]: DEBUG oslo_vmware.rw_handles [None req-aa1aa995-4e31-45e6-b824-29d545be859e tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/526fd05d-3941-e7f4-4056-812724b9c6fb/disk-0.vmdk. {{(pid=61923) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 870.928210] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53829851-c013-4084-9cec-8cc890fb722c {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.932986] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2edc9689-41a2-449b-a293-6c42af742df6 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.940212] env[61923]: DEBUG oslo_vmware.rw_handles [None req-aa1aa995-4e31-45e6-b824-29d545be859e tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/526fd05d-3941-e7f4-4056-812724b9c6fb/disk-0.vmdk is in state: ready. {{(pid=61923) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 870.940428] env[61923]: ERROR oslo_vmware.rw_handles [None req-aa1aa995-4e31-45e6-b824-29d545be859e tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Aborting lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/526fd05d-3941-e7f4-4056-812724b9c6fb/disk-0.vmdk due to incomplete transfer. [ 870.948198] env[61923]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-005ef356-fda2-47a6-b845-3617a22e375f {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.961562] env[61923]: DEBUG oslo_vmware.rw_handles [None req-aa1aa995-4e31-45e6-b824-29d545be859e tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Closed VMDK read handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/526fd05d-3941-e7f4-4056-812724b9c6fb/disk-0.vmdk. {{(pid=61923) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 870.961860] env[61923]: DEBUG nova.virt.vmwareapi.images [None req-aa1aa995-4e31-45e6-b824-29d545be859e tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] [instance: 59198f4d-4dde-4eaf-9f6c-a962cbe53c6e] Uploaded image 9ddad7e6-f1e4-49aa-b956-1b8d3273318d to the Glance image server {{(pid=61923) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 870.964413] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-aa1aa995-4e31-45e6-b824-29d545be859e tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] [instance: 59198f4d-4dde-4eaf-9f6c-a962cbe53c6e] Destroying the VM {{(pid=61923) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 870.964729] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-81563c1c-4840-422e-b279-02bda7213c9a {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.977079] env[61923]: DEBUG oslo_vmware.api [None req-aa1aa995-4e31-45e6-b824-29d545be859e tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Waiting for the task: (returnval){ [ 870.977079] env[61923]: value = "task-1377658" [ 870.977079] env[61923]: _type = "Task" [ 870.977079] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 870.987548] env[61923]: DEBUG oslo_vmware.api [None req-aa1aa995-4e31-45e6-b824-29d545be859e tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Task: {'id': task-1377658, 'name': Destroy_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.057148] env[61923]: DEBUG nova.scheduler.client.report [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Updated inventory for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 with generation 87 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 177, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 871.057458] env[61923]: DEBUG nova.compute.provider_tree [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Updating resource provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 generation from 87 to 88 during operation: update_inventory {{(pid=61923) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 871.057554] env[61923]: DEBUG nova.compute.provider_tree [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Updating inventory in ProviderTree for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 177, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 871.128100] env[61923]: DEBUG nova.compute.manager [None req-6d4df172-3edf-4dbb-b96f-afa60a7b76e1 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] [instance: a511392e-9ab6-42fb-b07c-f90e9786dcc1] Starting instance... {{(pid=61923) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 871.155661] env[61923]: DEBUG oslo_vmware.api [None req-00703b42-f309-4672-8dc7-10f4a69426aa tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Task: {'id': task-1377657, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.286191] env[61923]: DEBUG oslo_concurrency.lockutils [None req-13baf863-98e6-41c7-ae75-48fd1a9cef32 tempest-ServerAddressesTestJSON-1846886542 tempest-ServerAddressesTestJSON-1846886542-project-member] Lock "5b73e07a-d6ef-4dcf-bdae-eea91d2aeb6f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 49.298s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 871.486764] env[61923]: DEBUG oslo_vmware.api [None req-aa1aa995-4e31-45e6-b824-29d545be859e tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Task: {'id': task-1377658, 'name': Destroy_Task} progress is 33%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.515035] env[61923]: DEBUG nova.compute.manager [req-97bdc3e0-8e2a-4c19-bc95-05949818b156 req-92dd7fee-0d95-48e8-82fb-ed6c55493c9c service nova] [instance: 6012d3b5-739c-4762-9bb4-09c51171dcd7] Received event network-vif-plugged-269a41b0-75e7-46af-9497-8eae3102df11 {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 871.515269] env[61923]: DEBUG oslo_concurrency.lockutils [req-97bdc3e0-8e2a-4c19-bc95-05949818b156 req-92dd7fee-0d95-48e8-82fb-ed6c55493c9c service nova] Acquiring lock "6012d3b5-739c-4762-9bb4-09c51171dcd7-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 871.515523] env[61923]: DEBUG oslo_concurrency.lockutils [req-97bdc3e0-8e2a-4c19-bc95-05949818b156 req-92dd7fee-0d95-48e8-82fb-ed6c55493c9c service nova] Lock "6012d3b5-739c-4762-9bb4-09c51171dcd7-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 871.515709] env[61923]: DEBUG oslo_concurrency.lockutils [req-97bdc3e0-8e2a-4c19-bc95-05949818b156 req-92dd7fee-0d95-48e8-82fb-ed6c55493c9c service nova] Lock "6012d3b5-739c-4762-9bb4-09c51171dcd7-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 871.515933] env[61923]: DEBUG nova.compute.manager [req-97bdc3e0-8e2a-4c19-bc95-05949818b156 req-92dd7fee-0d95-48e8-82fb-ed6c55493c9c service nova] [instance: 6012d3b5-739c-4762-9bb4-09c51171dcd7] No waiting events found dispatching network-vif-plugged-269a41b0-75e7-46af-9497-8eae3102df11 {{(pid=61923) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 871.517190] env[61923]: WARNING nova.compute.manager [req-97bdc3e0-8e2a-4c19-bc95-05949818b156 req-92dd7fee-0d95-48e8-82fb-ed6c55493c9c service nova] [instance: 6012d3b5-739c-4762-9bb4-09c51171dcd7] Received unexpected event network-vif-plugged-269a41b0-75e7-46af-9497-8eae3102df11 for instance with vm_state building and task_state spawning. [ 871.563307] env[61923]: DEBUG nova.compute.resource_tracker [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61923) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 871.563614] env[61923]: DEBUG oslo_concurrency.lockutils [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 3.414s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 871.563965] env[61923]: DEBUG oslo_concurrency.lockutils [None req-b8ed9576-aa00-419c-8225-19ac3c654fd7 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 16.139s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 871.564238] env[61923]: DEBUG oslo_concurrency.lockutils [None req-b8ed9576-aa00-419c-8225-19ac3c654fd7 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 871.567469] env[61923]: DEBUG oslo_concurrency.lockutils [None req-144d62d8-7418-4330-9258-261ec048af9e tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 4.257s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 871.567469] env[61923]: DEBUG oslo_concurrency.lockutils [None req-144d62d8-7418-4330-9258-261ec048af9e tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 871.591530] env[61923]: INFO nova.scheduler.client.report [None req-b8ed9576-aa00-419c-8225-19ac3c654fd7 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Deleted allocations for instance 73692517-1816-4e9b-ab2d-8265c683e83d [ 871.597876] env[61923]: INFO nova.scheduler.client.report [None req-144d62d8-7418-4330-9258-261ec048af9e tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] Deleted allocations for instance fc4fdf89-4615-4586-9dcd-ea90dc258361 [ 871.615413] env[61923]: DEBUG nova.network.neutron [None req-9887b108-13fe-4987-ad6f-1ef139e2a15d tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] [instance: 6012d3b5-739c-4762-9bb4-09c51171dcd7] Successfully updated port: 269a41b0-75e7-46af-9497-8eae3102df11 {{(pid=61923) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 871.661879] env[61923]: DEBUG oslo_vmware.api [None req-00703b42-f309-4672-8dc7-10f4a69426aa tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Task: {'id': task-1377657, 'name': ReconfigVM_Task, 'duration_secs': 0.629062} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 871.661879] env[61923]: DEBUG nova.virt.vmwareapi.volumeops [None req-00703b42-f309-4672-8dc7-10f4a69426aa tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] [instance: b779d183-89ae-4e4d-ae99-e514e145ed43] Reconfigured VM instance instance-0000004d to attach disk [datastore2] b779d183-89ae-4e4d-ae99-e514e145ed43/b779d183-89ae-4e4d-ae99-e514e145ed43.vmdk or device None with type sparse {{(pid=61923) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 871.661879] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-1eee0d84-c6e2-4885-987d-a8bbd245e5d6 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.664667] env[61923]: DEBUG oslo_concurrency.lockutils [None req-6d4df172-3edf-4dbb-b96f-afa60a7b76e1 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 871.665461] env[61923]: DEBUG oslo_concurrency.lockutils [None req-6d4df172-3edf-4dbb-b96f-afa60a7b76e1 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.001s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 871.667400] env[61923]: INFO nova.compute.claims [None req-6d4df172-3edf-4dbb-b96f-afa60a7b76e1 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] [instance: a511392e-9ab6-42fb-b07c-f90e9786dcc1] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 871.672339] env[61923]: DEBUG oslo_vmware.api [None req-00703b42-f309-4672-8dc7-10f4a69426aa tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Waiting for the task: (returnval){ [ 871.672339] env[61923]: value = "task-1377659" [ 871.672339] env[61923]: _type = "Task" [ 871.672339] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 871.688326] env[61923]: DEBUG oslo_vmware.api [None req-00703b42-f309-4672-8dc7-10f4a69426aa tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Task: {'id': task-1377659, 'name': Rename_Task} progress is 5%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.986099] env[61923]: DEBUG oslo_vmware.api [None req-aa1aa995-4e31-45e6-b824-29d545be859e tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Task: {'id': task-1377658, 'name': Destroy_Task, 'duration_secs': 0.833695} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 871.986781] env[61923]: INFO nova.virt.vmwareapi.vm_util [None req-aa1aa995-4e31-45e6-b824-29d545be859e tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] [instance: 59198f4d-4dde-4eaf-9f6c-a962cbe53c6e] Destroyed the VM [ 871.987287] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-aa1aa995-4e31-45e6-b824-29d545be859e tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] [instance: 59198f4d-4dde-4eaf-9f6c-a962cbe53c6e] Deleting Snapshot of the VM instance {{(pid=61923) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 871.987624] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-cfc848a5-edce-4b0d-83f5-7733a777293a {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.995879] env[61923]: DEBUG oslo_vmware.api [None req-aa1aa995-4e31-45e6-b824-29d545be859e tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Waiting for the task: (returnval){ [ 871.995879] env[61923]: value = "task-1377660" [ 871.995879] env[61923]: _type = "Task" [ 871.995879] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 872.007031] env[61923]: DEBUG oslo_vmware.api [None req-aa1aa995-4e31-45e6-b824-29d545be859e tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Task: {'id': task-1377660, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.108116] env[61923]: DEBUG oslo_concurrency.lockutils [None req-b8ed9576-aa00-419c-8225-19ac3c654fd7 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Lock "73692517-1816-4e9b-ab2d-8265c683e83d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 19.909s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 872.109410] env[61923]: DEBUG oslo_concurrency.lockutils [None req-7e66bb89-e0a5-4005-a075-6cc387f412c9 tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Acquiring lock "444affa5-a7ed-4a17-9015-9fd5724aab64" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 872.109671] env[61923]: DEBUG oslo_concurrency.lockutils [None req-7e66bb89-e0a5-4005-a075-6cc387f412c9 tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Lock "444affa5-a7ed-4a17-9015-9fd5724aab64" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 872.109877] env[61923]: DEBUG oslo_concurrency.lockutils [None req-7e66bb89-e0a5-4005-a075-6cc387f412c9 tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Acquiring lock "444affa5-a7ed-4a17-9015-9fd5724aab64-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 872.110093] env[61923]: DEBUG oslo_concurrency.lockutils [None req-7e66bb89-e0a5-4005-a075-6cc387f412c9 tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Lock "444affa5-a7ed-4a17-9015-9fd5724aab64-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 872.110310] env[61923]: DEBUG oslo_concurrency.lockutils [None req-7e66bb89-e0a5-4005-a075-6cc387f412c9 tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Lock "444affa5-a7ed-4a17-9015-9fd5724aab64-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 872.112991] env[61923]: INFO nova.compute.manager [None req-7e66bb89-e0a5-4005-a075-6cc387f412c9 tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] [instance: 444affa5-a7ed-4a17-9015-9fd5724aab64] Terminating instance [ 872.115247] env[61923]: DEBUG nova.compute.manager [None req-7e66bb89-e0a5-4005-a075-6cc387f412c9 tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] [instance: 444affa5-a7ed-4a17-9015-9fd5724aab64] Start destroying the instance on the hypervisor. {{(pid=61923) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 872.115456] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-7e66bb89-e0a5-4005-a075-6cc387f412c9 tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] [instance: 444affa5-a7ed-4a17-9015-9fd5724aab64] Destroying instance {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 872.116728] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0989ca5e-27fe-4005-b6ac-2905d1e2ff64 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.120225] env[61923]: DEBUG oslo_concurrency.lockutils [None req-9887b108-13fe-4987-ad6f-1ef139e2a15d tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Acquiring lock "refresh_cache-6012d3b5-739c-4762-9bb4-09c51171dcd7" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 872.120359] env[61923]: DEBUG oslo_concurrency.lockutils [None req-9887b108-13fe-4987-ad6f-1ef139e2a15d tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Acquired lock "refresh_cache-6012d3b5-739c-4762-9bb4-09c51171dcd7" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 872.120513] env[61923]: DEBUG nova.network.neutron [None req-9887b108-13fe-4987-ad6f-1ef139e2a15d tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] [instance: 6012d3b5-739c-4762-9bb4-09c51171dcd7] Building network info cache for instance {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 872.130559] env[61923]: DEBUG oslo_concurrency.lockutils [None req-144d62d8-7418-4330-9258-261ec048af9e tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] Lock "fc4fdf89-4615-4586-9dcd-ea90dc258361" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 8.797s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 872.147604] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-7e66bb89-e0a5-4005-a075-6cc387f412c9 tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] [instance: 444affa5-a7ed-4a17-9015-9fd5724aab64] Powering off the VM {{(pid=61923) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 872.147916] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3cbc026c-65f3-473c-9afd-55afd1afc51d {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.156364] env[61923]: DEBUG oslo_vmware.api [None req-7e66bb89-e0a5-4005-a075-6cc387f412c9 tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Waiting for the task: (returnval){ [ 872.156364] env[61923]: value = "task-1377661" [ 872.156364] env[61923]: _type = "Task" [ 872.156364] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 872.166808] env[61923]: DEBUG oslo_vmware.api [None req-7e66bb89-e0a5-4005-a075-6cc387f412c9 tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Task: {'id': task-1377661, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.188069] env[61923]: DEBUG oslo_vmware.api [None req-00703b42-f309-4672-8dc7-10f4a69426aa tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Task: {'id': task-1377659, 'name': Rename_Task} progress is 99%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.220865] env[61923]: DEBUG oslo_concurrency.lockutils [None req-89f7a2ee-0580-4457-b803-a0236f2ecef8 tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Acquiring lock "e413c45d-cd89-44d4-9102-3d188907e7bb" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 872.221188] env[61923]: DEBUG oslo_concurrency.lockutils [None req-89f7a2ee-0580-4457-b803-a0236f2ecef8 tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Lock "e413c45d-cd89-44d4-9102-3d188907e7bb" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 872.221409] env[61923]: DEBUG oslo_concurrency.lockutils [None req-89f7a2ee-0580-4457-b803-a0236f2ecef8 tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Acquiring lock "e413c45d-cd89-44d4-9102-3d188907e7bb-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 872.221621] env[61923]: DEBUG oslo_concurrency.lockutils [None req-89f7a2ee-0580-4457-b803-a0236f2ecef8 tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Lock "e413c45d-cd89-44d4-9102-3d188907e7bb-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 872.222020] env[61923]: DEBUG oslo_concurrency.lockutils [None req-89f7a2ee-0580-4457-b803-a0236f2ecef8 tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Lock "e413c45d-cd89-44d4-9102-3d188907e7bb-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 872.224098] env[61923]: INFO nova.compute.manager [None req-89f7a2ee-0580-4457-b803-a0236f2ecef8 tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] [instance: e413c45d-cd89-44d4-9102-3d188907e7bb] Terminating instance [ 872.226293] env[61923]: DEBUG nova.compute.manager [None req-89f7a2ee-0580-4457-b803-a0236f2ecef8 tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] [instance: e413c45d-cd89-44d4-9102-3d188907e7bb] Start destroying the instance on the hypervisor. {{(pid=61923) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 872.226499] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-89f7a2ee-0580-4457-b803-a0236f2ecef8 tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] [instance: e413c45d-cd89-44d4-9102-3d188907e7bb] Destroying instance {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 872.227425] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba342cde-d538-43d7-8bb6-d872f4bf1f3a {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.236479] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-89f7a2ee-0580-4457-b803-a0236f2ecef8 tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] [instance: e413c45d-cd89-44d4-9102-3d188907e7bb] Powering off the VM {{(pid=61923) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 872.236945] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-880eb252-9e68-4ce9-baaa-7f9f3b482b5e {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.245577] env[61923]: DEBUG oslo_vmware.api [None req-89f7a2ee-0580-4457-b803-a0236f2ecef8 tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Waiting for the task: (returnval){ [ 872.245577] env[61923]: value = "task-1377662" [ 872.245577] env[61923]: _type = "Task" [ 872.245577] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 872.257443] env[61923]: DEBUG oslo_vmware.api [None req-89f7a2ee-0580-4457-b803-a0236f2ecef8 tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Task: {'id': task-1377662, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.456801] env[61923]: DEBUG oslo_concurrency.lockutils [None req-59b20242-d3b6-4451-b636-b9e52b91d9f2 tempest-ServerAddressesTestJSON-1846886542 tempest-ServerAddressesTestJSON-1846886542-project-member] Acquiring lock "5b73e07a-d6ef-4dcf-bdae-eea91d2aeb6f" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 872.457235] env[61923]: DEBUG oslo_concurrency.lockutils [None req-59b20242-d3b6-4451-b636-b9e52b91d9f2 tempest-ServerAddressesTestJSON-1846886542 tempest-ServerAddressesTestJSON-1846886542-project-member] Lock "5b73e07a-d6ef-4dcf-bdae-eea91d2aeb6f" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 872.457420] env[61923]: DEBUG oslo_concurrency.lockutils [None req-59b20242-d3b6-4451-b636-b9e52b91d9f2 tempest-ServerAddressesTestJSON-1846886542 tempest-ServerAddressesTestJSON-1846886542-project-member] Acquiring lock "5b73e07a-d6ef-4dcf-bdae-eea91d2aeb6f-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 872.457608] env[61923]: DEBUG oslo_concurrency.lockutils [None req-59b20242-d3b6-4451-b636-b9e52b91d9f2 tempest-ServerAddressesTestJSON-1846886542 tempest-ServerAddressesTestJSON-1846886542-project-member] Lock "5b73e07a-d6ef-4dcf-bdae-eea91d2aeb6f-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 872.457779] env[61923]: DEBUG oslo_concurrency.lockutils [None req-59b20242-d3b6-4451-b636-b9e52b91d9f2 tempest-ServerAddressesTestJSON-1846886542 tempest-ServerAddressesTestJSON-1846886542-project-member] Lock "5b73e07a-d6ef-4dcf-bdae-eea91d2aeb6f-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 872.460470] env[61923]: INFO nova.compute.manager [None req-59b20242-d3b6-4451-b636-b9e52b91d9f2 tempest-ServerAddressesTestJSON-1846886542 tempest-ServerAddressesTestJSON-1846886542-project-member] [instance: 5b73e07a-d6ef-4dcf-bdae-eea91d2aeb6f] Terminating instance [ 872.463502] env[61923]: DEBUG nova.compute.manager [None req-59b20242-d3b6-4451-b636-b9e52b91d9f2 tempest-ServerAddressesTestJSON-1846886542 tempest-ServerAddressesTestJSON-1846886542-project-member] [instance: 5b73e07a-d6ef-4dcf-bdae-eea91d2aeb6f] Start destroying the instance on the hypervisor. {{(pid=61923) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 872.463960] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-59b20242-d3b6-4451-b636-b9e52b91d9f2 tempest-ServerAddressesTestJSON-1846886542 tempest-ServerAddressesTestJSON-1846886542-project-member] [instance: 5b73e07a-d6ef-4dcf-bdae-eea91d2aeb6f] Destroying instance {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 872.464804] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-856c3c71-8043-463b-ba16-71b1f9f57c2e {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.474897] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-59b20242-d3b6-4451-b636-b9e52b91d9f2 tempest-ServerAddressesTestJSON-1846886542 tempest-ServerAddressesTestJSON-1846886542-project-member] [instance: 5b73e07a-d6ef-4dcf-bdae-eea91d2aeb6f] Powering off the VM {{(pid=61923) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 872.475109] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d7ab0bc7-f1f7-46ae-98cf-bc3e297eeff2 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.483480] env[61923]: DEBUG oslo_vmware.api [None req-59b20242-d3b6-4451-b636-b9e52b91d9f2 tempest-ServerAddressesTestJSON-1846886542 tempest-ServerAddressesTestJSON-1846886542-project-member] Waiting for the task: (returnval){ [ 872.483480] env[61923]: value = "task-1377663" [ 872.483480] env[61923]: _type = "Task" [ 872.483480] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 872.498776] env[61923]: DEBUG oslo_vmware.api [None req-59b20242-d3b6-4451-b636-b9e52b91d9f2 tempest-ServerAddressesTestJSON-1846886542 tempest-ServerAddressesTestJSON-1846886542-project-member] Task: {'id': task-1377663, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.512419] env[61923]: DEBUG oslo_vmware.api [None req-aa1aa995-4e31-45e6-b824-29d545be859e tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Task: {'id': task-1377660, 'name': RemoveSnapshot_Task} progress is 97%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.634017] env[61923]: DEBUG oslo_concurrency.lockutils [None req-55b58374-2c35-46b9-a508-d4833264ecac tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Acquiring lock "4cb3c397-ef1a-4267-a300-26da48206f3f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 872.634267] env[61923]: DEBUG oslo_concurrency.lockutils [None req-55b58374-2c35-46b9-a508-d4833264ecac tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Lock "4cb3c397-ef1a-4267-a300-26da48206f3f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 872.671108] env[61923]: DEBUG oslo_vmware.api [None req-7e66bb89-e0a5-4005-a075-6cc387f412c9 tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Task: {'id': task-1377661, 'name': PowerOffVM_Task, 'duration_secs': 0.432849} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 872.671757] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-7e66bb89-e0a5-4005-a075-6cc387f412c9 tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] [instance: 444affa5-a7ed-4a17-9015-9fd5724aab64] Powered off the VM {{(pid=61923) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 872.671995] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-7e66bb89-e0a5-4005-a075-6cc387f412c9 tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] [instance: 444affa5-a7ed-4a17-9015-9fd5724aab64] Unregistering the VM {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 872.672319] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-5d3284ea-4f1c-43a1-9be5-a8a54ecf2dc7 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.680131] env[61923]: DEBUG nova.network.neutron [None req-9887b108-13fe-4987-ad6f-1ef139e2a15d tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] [instance: 6012d3b5-739c-4762-9bb4-09c51171dcd7] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 872.691592] env[61923]: DEBUG oslo_vmware.api [None req-00703b42-f309-4672-8dc7-10f4a69426aa tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Task: {'id': task-1377659, 'name': Rename_Task} progress is 99%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.760229] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-7e66bb89-e0a5-4005-a075-6cc387f412c9 tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] [instance: 444affa5-a7ed-4a17-9015-9fd5724aab64] Unregistered the VM {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 872.760229] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-7e66bb89-e0a5-4005-a075-6cc387f412c9 tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] [instance: 444affa5-a7ed-4a17-9015-9fd5724aab64] Deleting contents of the VM from datastore datastore1 {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 872.760229] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-7e66bb89-e0a5-4005-a075-6cc387f412c9 tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Deleting the datastore file [datastore1] 444affa5-a7ed-4a17-9015-9fd5724aab64 {{(pid=61923) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 872.762217] env[61923]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-8674aee7-4846-480e-acf7-e449516398b9 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.772457] env[61923]: DEBUG oslo_vmware.api [None req-89f7a2ee-0580-4457-b803-a0236f2ecef8 tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Task: {'id': task-1377662, 'name': PowerOffVM_Task, 'duration_secs': 0.350321} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 872.777119] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-89f7a2ee-0580-4457-b803-a0236f2ecef8 tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] [instance: e413c45d-cd89-44d4-9102-3d188907e7bb] Powered off the VM {{(pid=61923) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 872.778207] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-89f7a2ee-0580-4457-b803-a0236f2ecef8 tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] [instance: e413c45d-cd89-44d4-9102-3d188907e7bb] Unregistering the VM {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 872.778374] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-0dedf614-17e7-4bfc-9c98-c4e1580724d4 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.783886] env[61923]: DEBUG oslo_vmware.api [None req-7e66bb89-e0a5-4005-a075-6cc387f412c9 tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Waiting for the task: (returnval){ [ 872.783886] env[61923]: value = "task-1377665" [ 872.783886] env[61923]: _type = "Task" [ 872.783886] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 872.802749] env[61923]: DEBUG oslo_vmware.api [None req-7e66bb89-e0a5-4005-a075-6cc387f412c9 tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Task: {'id': task-1377665, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.864598] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-89f7a2ee-0580-4457-b803-a0236f2ecef8 tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] [instance: e413c45d-cd89-44d4-9102-3d188907e7bb] Unregistered the VM {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 872.864805] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-89f7a2ee-0580-4457-b803-a0236f2ecef8 tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] [instance: e413c45d-cd89-44d4-9102-3d188907e7bb] Deleting contents of the VM from datastore datastore2 {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 872.865474] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-89f7a2ee-0580-4457-b803-a0236f2ecef8 tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Deleting the datastore file [datastore2] e413c45d-cd89-44d4-9102-3d188907e7bb {{(pid=61923) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 872.865474] env[61923]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a40be2ef-7da0-4ae4-84c7-3a7ed498c8e2 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.874081] env[61923]: DEBUG oslo_vmware.api [None req-89f7a2ee-0580-4457-b803-a0236f2ecef8 tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Waiting for the task: (returnval){ [ 872.874081] env[61923]: value = "task-1377667" [ 872.874081] env[61923]: _type = "Task" [ 872.874081] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 872.885473] env[61923]: DEBUG oslo_vmware.api [None req-89f7a2ee-0580-4457-b803-a0236f2ecef8 tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Task: {'id': task-1377667, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.900019] env[61923]: DEBUG nova.network.neutron [None req-9887b108-13fe-4987-ad6f-1ef139e2a15d tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] [instance: 6012d3b5-739c-4762-9bb4-09c51171dcd7] Updating instance_info_cache with network_info: [{"id": "269a41b0-75e7-46af-9497-8eae3102df11", "address": "fa:16:3e:e1:02:b8", "network": {"id": "a8b7e1b4-1ec6-4226-bdd6-535dbce450d2", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1928758125-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "912300a09b80452e85e3cd13c4a644cc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7806fe18-2b89-4386-87b1-f22876f82af2", "external-id": "nsx-vlan-transportzone-727", "segmentation_id": 727, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap269a41b0-75", "ovs_interfaceid": "269a41b0-75e7-46af-9497-8eae3102df11", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 872.929137] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f89d0fac-09d4-403a-91ba-61f793b7d4b1 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.938891] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a05054d1-eec3-4ece-ac5c-416fd876f792 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.976616] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f5afeeb-b49a-493c-bf46-b4baea22c944 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.989167] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2acbdf74-7d62-4555-8957-90708b72b15a {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.005775] env[61923]: DEBUG nova.compute.provider_tree [None req-6d4df172-3edf-4dbb-b96f-afa60a7b76e1 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 873.013525] env[61923]: DEBUG oslo_vmware.api [None req-59b20242-d3b6-4451-b636-b9e52b91d9f2 tempest-ServerAddressesTestJSON-1846886542 tempest-ServerAddressesTestJSON-1846886542-project-member] Task: {'id': task-1377663, 'name': PowerOffVM_Task} progress is 100%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.021200] env[61923]: DEBUG oslo_vmware.api [None req-aa1aa995-4e31-45e6-b824-29d545be859e tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Task: {'id': task-1377660, 'name': RemoveSnapshot_Task, 'duration_secs': 0.866748} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 873.023322] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-aa1aa995-4e31-45e6-b824-29d545be859e tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] [instance: 59198f4d-4dde-4eaf-9f6c-a962cbe53c6e] Deleted Snapshot of the VM instance {{(pid=61923) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 873.023322] env[61923]: INFO nova.compute.manager [None req-aa1aa995-4e31-45e6-b824-29d545be859e tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] [instance: 59198f4d-4dde-4eaf-9f6c-a962cbe53c6e] Took 17.12 seconds to snapshot the instance on the hypervisor. [ 873.138884] env[61923]: DEBUG nova.compute.manager [None req-55b58374-2c35-46b9-a508-d4833264ecac tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] [instance: 4cb3c397-ef1a-4267-a300-26da48206f3f] Starting instance... {{(pid=61923) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 873.187046] env[61923]: DEBUG oslo_vmware.api [None req-00703b42-f309-4672-8dc7-10f4a69426aa tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Task: {'id': task-1377659, 'name': Rename_Task, 'duration_secs': 1.211073} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 873.187943] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-00703b42-f309-4672-8dc7-10f4a69426aa tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] [instance: b779d183-89ae-4e4d-ae99-e514e145ed43] Powering on the VM {{(pid=61923) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 873.188229] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e91e12ec-dbc9-4aca-9a52-04378fb1a489 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.197805] env[61923]: DEBUG oslo_vmware.api [None req-00703b42-f309-4672-8dc7-10f4a69426aa tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Waiting for the task: (returnval){ [ 873.197805] env[61923]: value = "task-1377668" [ 873.197805] env[61923]: _type = "Task" [ 873.197805] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 873.210147] env[61923]: DEBUG oslo_vmware.api [None req-00703b42-f309-4672-8dc7-10f4a69426aa tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Task: {'id': task-1377668, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.274022] env[61923]: DEBUG oslo_concurrency.lockutils [None req-eb5e01f1-148c-4608-96e7-93572113f79a tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Acquiring lock "4c2c973a-5d23-4b8a-8b99-6b634e337054" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 873.274385] env[61923]: DEBUG oslo_concurrency.lockutils [None req-eb5e01f1-148c-4608-96e7-93572113f79a tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Lock "4c2c973a-5d23-4b8a-8b99-6b634e337054" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 873.298575] env[61923]: DEBUG oslo_vmware.api [None req-7e66bb89-e0a5-4005-a075-6cc387f412c9 tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Task: {'id': task-1377665, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.246946} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 873.298846] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-7e66bb89-e0a5-4005-a075-6cc387f412c9 tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Deleted the datastore file {{(pid=61923) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 873.299032] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-7e66bb89-e0a5-4005-a075-6cc387f412c9 tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] [instance: 444affa5-a7ed-4a17-9015-9fd5724aab64] Deleted contents of the VM from datastore datastore1 {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 873.299259] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-7e66bb89-e0a5-4005-a075-6cc387f412c9 tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] [instance: 444affa5-a7ed-4a17-9015-9fd5724aab64] Instance destroyed {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 873.299406] env[61923]: INFO nova.compute.manager [None req-7e66bb89-e0a5-4005-a075-6cc387f412c9 tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] [instance: 444affa5-a7ed-4a17-9015-9fd5724aab64] Took 1.18 seconds to destroy the instance on the hypervisor. [ 873.299645] env[61923]: DEBUG oslo.service.loopingcall [None req-7e66bb89-e0a5-4005-a075-6cc387f412c9 tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61923) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 873.299832] env[61923]: DEBUG nova.compute.manager [-] [instance: 444affa5-a7ed-4a17-9015-9fd5724aab64] Deallocating network for instance {{(pid=61923) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 873.299923] env[61923]: DEBUG nova.network.neutron [-] [instance: 444affa5-a7ed-4a17-9015-9fd5724aab64] deallocate_for_instance() {{(pid=61923) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 873.385355] env[61923]: DEBUG oslo_vmware.api [None req-89f7a2ee-0580-4457-b803-a0236f2ecef8 tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Task: {'id': task-1377667, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.203065} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 873.385661] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-89f7a2ee-0580-4457-b803-a0236f2ecef8 tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Deleted the datastore file {{(pid=61923) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 873.385661] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-89f7a2ee-0580-4457-b803-a0236f2ecef8 tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] [instance: e413c45d-cd89-44d4-9102-3d188907e7bb] Deleted contents of the VM from datastore datastore2 {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 873.385853] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-89f7a2ee-0580-4457-b803-a0236f2ecef8 tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] [instance: e413c45d-cd89-44d4-9102-3d188907e7bb] Instance destroyed {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 873.386199] env[61923]: INFO nova.compute.manager [None req-89f7a2ee-0580-4457-b803-a0236f2ecef8 tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] [instance: e413c45d-cd89-44d4-9102-3d188907e7bb] Took 1.16 seconds to destroy the instance on the hypervisor. [ 873.386299] env[61923]: DEBUG oslo.service.loopingcall [None req-89f7a2ee-0580-4457-b803-a0236f2ecef8 tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61923) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 873.386499] env[61923]: DEBUG nova.compute.manager [-] [instance: e413c45d-cd89-44d4-9102-3d188907e7bb] Deallocating network for instance {{(pid=61923) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 873.386607] env[61923]: DEBUG nova.network.neutron [-] [instance: e413c45d-cd89-44d4-9102-3d188907e7bb] deallocate_for_instance() {{(pid=61923) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 873.407018] env[61923]: DEBUG oslo_concurrency.lockutils [None req-9887b108-13fe-4987-ad6f-1ef139e2a15d tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Releasing lock "refresh_cache-6012d3b5-739c-4762-9bb4-09c51171dcd7" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 873.407018] env[61923]: DEBUG nova.compute.manager [None req-9887b108-13fe-4987-ad6f-1ef139e2a15d tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] [instance: 6012d3b5-739c-4762-9bb4-09c51171dcd7] Instance network_info: |[{"id": "269a41b0-75e7-46af-9497-8eae3102df11", "address": "fa:16:3e:e1:02:b8", "network": {"id": "a8b7e1b4-1ec6-4226-bdd6-535dbce450d2", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1928758125-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "912300a09b80452e85e3cd13c4a644cc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7806fe18-2b89-4386-87b1-f22876f82af2", "external-id": "nsx-vlan-transportzone-727", "segmentation_id": 727, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap269a41b0-75", "ovs_interfaceid": "269a41b0-75e7-46af-9497-8eae3102df11", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61923) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 873.407018] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-9887b108-13fe-4987-ad6f-1ef139e2a15d tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] [instance: 6012d3b5-739c-4762-9bb4-09c51171dcd7] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:e1:02:b8', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '7806fe18-2b89-4386-87b1-f22876f82af2', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '269a41b0-75e7-46af-9497-8eae3102df11', 'vif_model': 'vmxnet3'}] {{(pid=61923) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 873.420667] env[61923]: DEBUG oslo.service.loopingcall [None req-9887b108-13fe-4987-ad6f-1ef139e2a15d tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61923) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 873.421759] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6012d3b5-739c-4762-9bb4-09c51171dcd7] Creating VM on the ESX host {{(pid=61923) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 873.421998] env[61923]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-74428fa7-2424-4405-a317-205e06dbb8c0 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.444823] env[61923]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 873.444823] env[61923]: value = "task-1377669" [ 873.444823] env[61923]: _type = "Task" [ 873.444823] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 873.453577] env[61923]: DEBUG oslo_vmware.api [-] Task: {'id': task-1377669, 'name': CreateVM_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.497457] env[61923]: DEBUG oslo_vmware.api [None req-59b20242-d3b6-4451-b636-b9e52b91d9f2 tempest-ServerAddressesTestJSON-1846886542 tempest-ServerAddressesTestJSON-1846886542-project-member] Task: {'id': task-1377663, 'name': PowerOffVM_Task, 'duration_secs': 0.542979} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 873.497533] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-59b20242-d3b6-4451-b636-b9e52b91d9f2 tempest-ServerAddressesTestJSON-1846886542 tempest-ServerAddressesTestJSON-1846886542-project-member] [instance: 5b73e07a-d6ef-4dcf-bdae-eea91d2aeb6f] Powered off the VM {{(pid=61923) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 873.497795] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-59b20242-d3b6-4451-b636-b9e52b91d9f2 tempest-ServerAddressesTestJSON-1846886542 tempest-ServerAddressesTestJSON-1846886542-project-member] [instance: 5b73e07a-d6ef-4dcf-bdae-eea91d2aeb6f] Unregistering the VM {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 873.498140] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-381a26cc-3b87-49ae-8033-b67938debb05 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.516201] env[61923]: DEBUG nova.scheduler.client.report [None req-6d4df172-3edf-4dbb-b96f-afa60a7b76e1 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 177, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 873.551232] env[61923]: DEBUG nova.compute.manager [req-710a94cb-e6a7-4a10-853a-827b189b2cae req-22acdd62-407b-4fa8-8b1b-6354d7bd4c65 service nova] [instance: 6012d3b5-739c-4762-9bb4-09c51171dcd7] Received event network-changed-269a41b0-75e7-46af-9497-8eae3102df11 {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 873.551232] env[61923]: DEBUG nova.compute.manager [req-710a94cb-e6a7-4a10-853a-827b189b2cae req-22acdd62-407b-4fa8-8b1b-6354d7bd4c65 service nova] [instance: 6012d3b5-739c-4762-9bb4-09c51171dcd7] Refreshing instance network info cache due to event network-changed-269a41b0-75e7-46af-9497-8eae3102df11. {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 873.551458] env[61923]: DEBUG oslo_concurrency.lockutils [req-710a94cb-e6a7-4a10-853a-827b189b2cae req-22acdd62-407b-4fa8-8b1b-6354d7bd4c65 service nova] Acquiring lock "refresh_cache-6012d3b5-739c-4762-9bb4-09c51171dcd7" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 873.554026] env[61923]: DEBUG oslo_concurrency.lockutils [req-710a94cb-e6a7-4a10-853a-827b189b2cae req-22acdd62-407b-4fa8-8b1b-6354d7bd4c65 service nova] Acquired lock "refresh_cache-6012d3b5-739c-4762-9bb4-09c51171dcd7" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 873.554026] env[61923]: DEBUG nova.network.neutron [req-710a94cb-e6a7-4a10-853a-827b189b2cae req-22acdd62-407b-4fa8-8b1b-6354d7bd4c65 service nova] [instance: 6012d3b5-739c-4762-9bb4-09c51171dcd7] Refreshing network info cache for port 269a41b0-75e7-46af-9497-8eae3102df11 {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 873.581067] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-59b20242-d3b6-4451-b636-b9e52b91d9f2 tempest-ServerAddressesTestJSON-1846886542 tempest-ServerAddressesTestJSON-1846886542-project-member] [instance: 5b73e07a-d6ef-4dcf-bdae-eea91d2aeb6f] Unregistered the VM {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 873.581757] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-59b20242-d3b6-4451-b636-b9e52b91d9f2 tempest-ServerAddressesTestJSON-1846886542 tempest-ServerAddressesTestJSON-1846886542-project-member] [instance: 5b73e07a-d6ef-4dcf-bdae-eea91d2aeb6f] Deleting contents of the VM from datastore datastore2 {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 873.581879] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-59b20242-d3b6-4451-b636-b9e52b91d9f2 tempest-ServerAddressesTestJSON-1846886542 tempest-ServerAddressesTestJSON-1846886542-project-member] Deleting the datastore file [datastore2] 5b73e07a-d6ef-4dcf-bdae-eea91d2aeb6f {{(pid=61923) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 873.582288] env[61923]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d2e49406-4042-4496-b0d8-3edef654c44b {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.592534] env[61923]: DEBUG oslo_vmware.api [None req-59b20242-d3b6-4451-b636-b9e52b91d9f2 tempest-ServerAddressesTestJSON-1846886542 tempest-ServerAddressesTestJSON-1846886542-project-member] Waiting for the task: (returnval){ [ 873.592534] env[61923]: value = "task-1377671" [ 873.592534] env[61923]: _type = "Task" [ 873.592534] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 873.603658] env[61923]: DEBUG oslo_vmware.api [None req-59b20242-d3b6-4451-b636-b9e52b91d9f2 tempest-ServerAddressesTestJSON-1846886542 tempest-ServerAddressesTestJSON-1846886542-project-member] Task: {'id': task-1377671, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.669417] env[61923]: DEBUG oslo_concurrency.lockutils [None req-55b58374-2c35-46b9-a508-d4833264ecac tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 873.712028] env[61923]: DEBUG oslo_vmware.api [None req-00703b42-f309-4672-8dc7-10f4a69426aa tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Task: {'id': task-1377668, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.778827] env[61923]: DEBUG nova.compute.manager [None req-eb5e01f1-148c-4608-96e7-93572113f79a tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: 4c2c973a-5d23-4b8a-8b99-6b634e337054] Starting instance... {{(pid=61923) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 873.876825] env[61923]: DEBUG nova.compute.manager [req-b2065585-38b2-4160-9279-1f8ec44d1167 req-4d2101ff-eb1f-4697-babe-6486020880d3 service nova] [instance: e413c45d-cd89-44d4-9102-3d188907e7bb] Received event network-vif-deleted-4ed56dd5-beb3-4254-932b-7290c7b85700 {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 873.877098] env[61923]: INFO nova.compute.manager [req-b2065585-38b2-4160-9279-1f8ec44d1167 req-4d2101ff-eb1f-4697-babe-6486020880d3 service nova] [instance: e413c45d-cd89-44d4-9102-3d188907e7bb] Neutron deleted interface 4ed56dd5-beb3-4254-932b-7290c7b85700; detaching it from the instance and deleting it from the info cache [ 873.877318] env[61923]: DEBUG nova.network.neutron [req-b2065585-38b2-4160-9279-1f8ec44d1167 req-4d2101ff-eb1f-4697-babe-6486020880d3 service nova] [instance: e413c45d-cd89-44d4-9102-3d188907e7bb] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 873.957059] env[61923]: DEBUG oslo_vmware.api [-] Task: {'id': task-1377669, 'name': CreateVM_Task, 'duration_secs': 0.397364} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 873.957059] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6012d3b5-739c-4762-9bb4-09c51171dcd7] Created VM on the ESX host {{(pid=61923) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 873.957614] env[61923]: DEBUG oslo_concurrency.lockutils [None req-9887b108-13fe-4987-ad6f-1ef139e2a15d tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 873.957792] env[61923]: DEBUG oslo_concurrency.lockutils [None req-9887b108-13fe-4987-ad6f-1ef139e2a15d tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 873.958157] env[61923]: DEBUG oslo_concurrency.lockutils [None req-9887b108-13fe-4987-ad6f-1ef139e2a15d tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 873.958423] env[61923]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-87564db2-4d90-43a6-8cef-620f52231b29 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.964108] env[61923]: DEBUG oslo_vmware.api [None req-9887b108-13fe-4987-ad6f-1ef139e2a15d tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Waiting for the task: (returnval){ [ 873.964108] env[61923]: value = "session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]5281da27-c755-8fe0-09dd-0a1496b32011" [ 873.964108] env[61923]: _type = "Task" [ 873.964108] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 873.972723] env[61923]: DEBUG oslo_vmware.api [None req-9887b108-13fe-4987-ad6f-1ef139e2a15d tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]5281da27-c755-8fe0-09dd-0a1496b32011, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.024008] env[61923]: DEBUG oslo_concurrency.lockutils [None req-6d4df172-3edf-4dbb-b96f-afa60a7b76e1 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.358s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 874.024645] env[61923]: DEBUG nova.compute.manager [None req-6d4df172-3edf-4dbb-b96f-afa60a7b76e1 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] [instance: a511392e-9ab6-42fb-b07c-f90e9786dcc1] Start building networks asynchronously for instance. {{(pid=61923) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 874.027679] env[61923]: DEBUG oslo_concurrency.lockutils [None req-55b58374-2c35-46b9-a508-d4833264ecac tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.358s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 874.029296] env[61923]: INFO nova.compute.claims [None req-55b58374-2c35-46b9-a508-d4833264ecac tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] [instance: 4cb3c397-ef1a-4267-a300-26da48206f3f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 874.102474] env[61923]: DEBUG oslo_vmware.api [None req-59b20242-d3b6-4451-b636-b9e52b91d9f2 tempest-ServerAddressesTestJSON-1846886542 tempest-ServerAddressesTestJSON-1846886542-project-member] Task: {'id': task-1377671, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.278778} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 874.102756] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-59b20242-d3b6-4451-b636-b9e52b91d9f2 tempest-ServerAddressesTestJSON-1846886542 tempest-ServerAddressesTestJSON-1846886542-project-member] Deleted the datastore file {{(pid=61923) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 874.102946] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-59b20242-d3b6-4451-b636-b9e52b91d9f2 tempest-ServerAddressesTestJSON-1846886542 tempest-ServerAddressesTestJSON-1846886542-project-member] [instance: 5b73e07a-d6ef-4dcf-bdae-eea91d2aeb6f] Deleted contents of the VM from datastore datastore2 {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 874.103181] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-59b20242-d3b6-4451-b636-b9e52b91d9f2 tempest-ServerAddressesTestJSON-1846886542 tempest-ServerAddressesTestJSON-1846886542-project-member] [instance: 5b73e07a-d6ef-4dcf-bdae-eea91d2aeb6f] Instance destroyed {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 874.103364] env[61923]: INFO nova.compute.manager [None req-59b20242-d3b6-4451-b636-b9e52b91d9f2 tempest-ServerAddressesTestJSON-1846886542 tempest-ServerAddressesTestJSON-1846886542-project-member] [instance: 5b73e07a-d6ef-4dcf-bdae-eea91d2aeb6f] Took 1.64 seconds to destroy the instance on the hypervisor. [ 874.103608] env[61923]: DEBUG oslo.service.loopingcall [None req-59b20242-d3b6-4451-b636-b9e52b91d9f2 tempest-ServerAddressesTestJSON-1846886542 tempest-ServerAddressesTestJSON-1846886542-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61923) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 874.104124] env[61923]: DEBUG nova.compute.manager [-] [instance: 5b73e07a-d6ef-4dcf-bdae-eea91d2aeb6f] Deallocating network for instance {{(pid=61923) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 874.104232] env[61923]: DEBUG nova.network.neutron [-] [instance: 5b73e07a-d6ef-4dcf-bdae-eea91d2aeb6f] deallocate_for_instance() {{(pid=61923) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 874.186349] env[61923]: DEBUG nova.network.neutron [-] [instance: 444affa5-a7ed-4a17-9015-9fd5724aab64] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 874.209732] env[61923]: DEBUG oslo_vmware.api [None req-00703b42-f309-4672-8dc7-10f4a69426aa tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Task: {'id': task-1377668, 'name': PowerOnVM_Task, 'duration_secs': 0.632507} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 874.210055] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-00703b42-f309-4672-8dc7-10f4a69426aa tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] [instance: b779d183-89ae-4e4d-ae99-e514e145ed43] Powered on the VM {{(pid=61923) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 874.210207] env[61923]: INFO nova.compute.manager [None req-00703b42-f309-4672-8dc7-10f4a69426aa tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] [instance: b779d183-89ae-4e4d-ae99-e514e145ed43] Took 10.03 seconds to spawn the instance on the hypervisor. [ 874.210413] env[61923]: DEBUG nova.compute.manager [None req-00703b42-f309-4672-8dc7-10f4a69426aa tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] [instance: b779d183-89ae-4e4d-ae99-e514e145ed43] Checking state {{(pid=61923) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 874.211234] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c3e4805-2163-4aff-abbe-cf51b6cce83d {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.278772] env[61923]: DEBUG nova.network.neutron [req-710a94cb-e6a7-4a10-853a-827b189b2cae req-22acdd62-407b-4fa8-8b1b-6354d7bd4c65 service nova] [instance: 6012d3b5-739c-4762-9bb4-09c51171dcd7] Updated VIF entry in instance network info cache for port 269a41b0-75e7-46af-9497-8eae3102df11. {{(pid=61923) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 874.279238] env[61923]: DEBUG nova.network.neutron [req-710a94cb-e6a7-4a10-853a-827b189b2cae req-22acdd62-407b-4fa8-8b1b-6354d7bd4c65 service nova] [instance: 6012d3b5-739c-4762-9bb4-09c51171dcd7] Updating instance_info_cache with network_info: [{"id": "269a41b0-75e7-46af-9497-8eae3102df11", "address": "fa:16:3e:e1:02:b8", "network": {"id": "a8b7e1b4-1ec6-4226-bdd6-535dbce450d2", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1928758125-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "912300a09b80452e85e3cd13c4a644cc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7806fe18-2b89-4386-87b1-f22876f82af2", "external-id": "nsx-vlan-transportzone-727", "segmentation_id": 727, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap269a41b0-75", "ovs_interfaceid": "269a41b0-75e7-46af-9497-8eae3102df11", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 874.309941] env[61923]: DEBUG oslo_concurrency.lockutils [None req-eb5e01f1-148c-4608-96e7-93572113f79a tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 874.349868] env[61923]: DEBUG nova.network.neutron [-] [instance: e413c45d-cd89-44d4-9102-3d188907e7bb] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 874.381195] env[61923]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-b671dede-a58a-4814-804b-6e2974503766 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.393197] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4e0a034-9e10-4478-9333-3b63aebda7c0 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.426138] env[61923]: DEBUG nova.compute.manager [req-b2065585-38b2-4160-9279-1f8ec44d1167 req-4d2101ff-eb1f-4697-babe-6486020880d3 service nova] [instance: e413c45d-cd89-44d4-9102-3d188907e7bb] Detach interface failed, port_id=4ed56dd5-beb3-4254-932b-7290c7b85700, reason: Instance e413c45d-cd89-44d4-9102-3d188907e7bb could not be found. {{(pid=61923) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 874.475968] env[61923]: DEBUG oslo_vmware.api [None req-9887b108-13fe-4987-ad6f-1ef139e2a15d tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]5281da27-c755-8fe0-09dd-0a1496b32011, 'name': SearchDatastore_Task, 'duration_secs': 0.015466} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 874.476321] env[61923]: DEBUG oslo_concurrency.lockutils [None req-9887b108-13fe-4987-ad6f-1ef139e2a15d tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 874.476558] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-9887b108-13fe-4987-ad6f-1ef139e2a15d tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] [instance: 6012d3b5-739c-4762-9bb4-09c51171dcd7] Processing image 0f153f63-ae0a-45b1-b7f5-7f9b673c947f {{(pid=61923) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 874.476793] env[61923]: DEBUG oslo_concurrency.lockutils [None req-9887b108-13fe-4987-ad6f-1ef139e2a15d tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 874.476941] env[61923]: DEBUG oslo_concurrency.lockutils [None req-9887b108-13fe-4987-ad6f-1ef139e2a15d tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 874.477218] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-9887b108-13fe-4987-ad6f-1ef139e2a15d tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61923) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 874.477490] env[61923]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-8dff9a5d-f95a-41c3-bfba-1936b59c4202 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.488435] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-9887b108-13fe-4987-ad6f-1ef139e2a15d tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61923) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 874.488661] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-9887b108-13fe-4987-ad6f-1ef139e2a15d tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61923) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 874.490440] env[61923]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-be06aa14-9ddc-4882-a853-7aeb75b11f3b {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.497037] env[61923]: DEBUG oslo_vmware.api [None req-9887b108-13fe-4987-ad6f-1ef139e2a15d tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Waiting for the task: (returnval){ [ 874.497037] env[61923]: value = "session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52b79837-995d-ffbe-f2db-b9a155cfca7a" [ 874.497037] env[61923]: _type = "Task" [ 874.497037] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 874.506714] env[61923]: DEBUG oslo_vmware.api [None req-9887b108-13fe-4987-ad6f-1ef139e2a15d tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52b79837-995d-ffbe-f2db-b9a155cfca7a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.534310] env[61923]: DEBUG nova.compute.utils [None req-6d4df172-3edf-4dbb-b96f-afa60a7b76e1 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Using /dev/sd instead of None {{(pid=61923) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 874.538121] env[61923]: DEBUG nova.compute.manager [None req-6d4df172-3edf-4dbb-b96f-afa60a7b76e1 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] [instance: a511392e-9ab6-42fb-b07c-f90e9786dcc1] Allocating IP information in the background. {{(pid=61923) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 874.538373] env[61923]: DEBUG nova.network.neutron [None req-6d4df172-3edf-4dbb-b96f-afa60a7b76e1 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] [instance: a511392e-9ab6-42fb-b07c-f90e9786dcc1] allocate_for_instance() {{(pid=61923) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 874.586121] env[61923]: DEBUG nova.policy [None req-6d4df172-3edf-4dbb-b96f-afa60a7b76e1 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '03056b1a9c2143efa96ebb46ee1fda1a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0d24f4a47d1f4b79a7d999a4dacb99ba', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61923) authorize /opt/stack/nova/nova/policy.py:201}} [ 874.693125] env[61923]: INFO nova.compute.manager [-] [instance: 444affa5-a7ed-4a17-9015-9fd5724aab64] Took 1.39 seconds to deallocate network for instance. [ 874.729953] env[61923]: INFO nova.compute.manager [None req-00703b42-f309-4672-8dc7-10f4a69426aa tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] [instance: b779d183-89ae-4e4d-ae99-e514e145ed43] Took 26.89 seconds to build instance. [ 874.785993] env[61923]: DEBUG oslo_concurrency.lockutils [req-710a94cb-e6a7-4a10-853a-827b189b2cae req-22acdd62-407b-4fa8-8b1b-6354d7bd4c65 service nova] Releasing lock "refresh_cache-6012d3b5-739c-4762-9bb4-09c51171dcd7" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 874.854343] env[61923]: INFO nova.compute.manager [-] [instance: e413c45d-cd89-44d4-9102-3d188907e7bb] Took 1.47 seconds to deallocate network for instance. [ 875.009689] env[61923]: DEBUG oslo_vmware.api [None req-9887b108-13fe-4987-ad6f-1ef139e2a15d tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52b79837-995d-ffbe-f2db-b9a155cfca7a, 'name': SearchDatastore_Task, 'duration_secs': 0.01779} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 875.010799] env[61923]: DEBUG nova.network.neutron [None req-6d4df172-3edf-4dbb-b96f-afa60a7b76e1 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] [instance: a511392e-9ab6-42fb-b07c-f90e9786dcc1] Successfully created port: 0325cd38-bd49-4132-85e3-c66fc9efe49f {{(pid=61923) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 875.013448] env[61923]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-dbe2333e-1d27-4e21-bef2-66a7519b457b {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.020286] env[61923]: DEBUG oslo_vmware.api [None req-9887b108-13fe-4987-ad6f-1ef139e2a15d tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Waiting for the task: (returnval){ [ 875.020286] env[61923]: value = "session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]524b78a4-2d40-82d3-571c-ff178c412adb" [ 875.020286] env[61923]: _type = "Task" [ 875.020286] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 875.031400] env[61923]: DEBUG oslo_vmware.api [None req-9887b108-13fe-4987-ad6f-1ef139e2a15d tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]524b78a4-2d40-82d3-571c-ff178c412adb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.031400] env[61923]: DEBUG nova.network.neutron [-] [instance: 5b73e07a-d6ef-4dcf-bdae-eea91d2aeb6f] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 875.048868] env[61923]: DEBUG nova.compute.manager [None req-6d4df172-3edf-4dbb-b96f-afa60a7b76e1 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] [instance: a511392e-9ab6-42fb-b07c-f90e9786dcc1] Start building block device mappings for instance. {{(pid=61923) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 875.198726] env[61923]: DEBUG oslo_concurrency.lockutils [None req-7e66bb89-e0a5-4005-a075-6cc387f412c9 tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 875.232502] env[61923]: DEBUG oslo_concurrency.lockutils [None req-00703b42-f309-4672-8dc7-10f4a69426aa tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Lock "b779d183-89ae-4e4d-ae99-e514e145ed43" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 42.085s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 875.272550] env[61923]: DEBUG oslo_concurrency.lockutils [None req-3192dd2d-7513-4a93-8a8c-c00e6e67889c tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] Acquiring lock "263f0866-49af-4ab6-8132-d993642ebedb" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 875.272844] env[61923]: DEBUG oslo_concurrency.lockutils [None req-3192dd2d-7513-4a93-8a8c-c00e6e67889c tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] Lock "263f0866-49af-4ab6-8132-d993642ebedb" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 875.315104] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87216a0f-f721-49b6-adfa-80abeb9043ac {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.324364] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5a273df-cabc-440c-88e3-20700cd39423 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.361672] env[61923]: DEBUG oslo_concurrency.lockutils [None req-89f7a2ee-0580-4457-b803-a0236f2ecef8 tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 875.362924] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9213c2f3-464e-4d1a-a7bf-265e90b49a06 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.372264] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07ff71c2-8d34-45f9-9b53-cf65a84d9d17 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.387275] env[61923]: DEBUG nova.compute.provider_tree [None req-55b58374-2c35-46b9-a508-d4833264ecac tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Updating inventory in ProviderTree for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 875.532146] env[61923]: INFO nova.compute.manager [-] [instance: 5b73e07a-d6ef-4dcf-bdae-eea91d2aeb6f] Took 1.43 seconds to deallocate network for instance. [ 875.532545] env[61923]: DEBUG oslo_vmware.api [None req-9887b108-13fe-4987-ad6f-1ef139e2a15d tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]524b78a4-2d40-82d3-571c-ff178c412adb, 'name': SearchDatastore_Task, 'duration_secs': 0.018365} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 875.534336] env[61923]: DEBUG oslo_concurrency.lockutils [None req-9887b108-13fe-4987-ad6f-1ef139e2a15d tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 875.534608] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-9887b108-13fe-4987-ad6f-1ef139e2a15d tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk to [datastore2] 6012d3b5-739c-4762-9bb4-09c51171dcd7/6012d3b5-739c-4762-9bb4-09c51171dcd7.vmdk {{(pid=61923) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 875.534873] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ac24f642-c912-44bf-85ba-dcc88adcf9ab {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.547583] env[61923]: DEBUG oslo_vmware.api [None req-9887b108-13fe-4987-ad6f-1ef139e2a15d tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Waiting for the task: (returnval){ [ 875.547583] env[61923]: value = "task-1377672" [ 875.547583] env[61923]: _type = "Task" [ 875.547583] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 875.562743] env[61923]: DEBUG oslo_vmware.api [None req-9887b108-13fe-4987-ad6f-1ef139e2a15d tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Task: {'id': task-1377672, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.588386] env[61923]: DEBUG nova.compute.manager [req-cc1bd282-7a49-4b21-9d90-6941b779597a req-2288f9fe-b158-45ab-96c4-976a6ff95bd7 service nova] [instance: 444affa5-a7ed-4a17-9015-9fd5724aab64] Received event network-vif-deleted-7b131b11-927e-4c0c-98f2-714b63683d77 {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 875.588559] env[61923]: DEBUG nova.compute.manager [req-cc1bd282-7a49-4b21-9d90-6941b779597a req-2288f9fe-b158-45ab-96c4-976a6ff95bd7 service nova] [instance: 5b73e07a-d6ef-4dcf-bdae-eea91d2aeb6f] Received event network-vif-deleted-262b2cfc-fe55-4a4e-bb7c-87f81d8c7e91 {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 875.777652] env[61923]: DEBUG nova.compute.manager [None req-3192dd2d-7513-4a93-8a8c-c00e6e67889c tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] [instance: 263f0866-49af-4ab6-8132-d993642ebedb] Starting instance... {{(pid=61923) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 875.910609] env[61923]: DEBUG nova.compute.manager [req-4277850f-faad-47be-a133-01d84ea4c920 req-fba41451-c310-4eb0-842b-3cbb262c265b service nova] [instance: b779d183-89ae-4e4d-ae99-e514e145ed43] Received event network-changed-d49a3010-c84a-4bef-beed-9ca4aa7d5dc2 {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 875.910609] env[61923]: DEBUG nova.compute.manager [req-4277850f-faad-47be-a133-01d84ea4c920 req-fba41451-c310-4eb0-842b-3cbb262c265b service nova] [instance: b779d183-89ae-4e4d-ae99-e514e145ed43] Refreshing instance network info cache due to event network-changed-d49a3010-c84a-4bef-beed-9ca4aa7d5dc2. {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 875.910609] env[61923]: DEBUG oslo_concurrency.lockutils [req-4277850f-faad-47be-a133-01d84ea4c920 req-fba41451-c310-4eb0-842b-3cbb262c265b service nova] Acquiring lock "refresh_cache-b779d183-89ae-4e4d-ae99-e514e145ed43" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 875.910609] env[61923]: DEBUG oslo_concurrency.lockutils [req-4277850f-faad-47be-a133-01d84ea4c920 req-fba41451-c310-4eb0-842b-3cbb262c265b service nova] Acquired lock "refresh_cache-b779d183-89ae-4e4d-ae99-e514e145ed43" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 875.910609] env[61923]: DEBUG nova.network.neutron [req-4277850f-faad-47be-a133-01d84ea4c920 req-fba41451-c310-4eb0-842b-3cbb262c265b service nova] [instance: b779d183-89ae-4e4d-ae99-e514e145ed43] Refreshing network info cache for port d49a3010-c84a-4bef-beed-9ca4aa7d5dc2 {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 875.911974] env[61923]: ERROR nova.scheduler.client.report [None req-55b58374-2c35-46b9-a508-d4833264ecac tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] [req-ba3bdf43-b353-4946-881f-32273dda68b2] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID f81803f2-f7f8-4939-a757-a77d34a1d0a2. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-ba3bdf43-b353-4946-881f-32273dda68b2"}]} [ 875.929810] env[61923]: DEBUG nova.scheduler.client.report [None req-55b58374-2c35-46b9-a508-d4833264ecac tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Refreshing inventories for resource provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 875.953551] env[61923]: DEBUG nova.scheduler.client.report [None req-55b58374-2c35-46b9-a508-d4833264ecac tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Updating ProviderTree inventory for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 177, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 875.953842] env[61923]: DEBUG nova.compute.provider_tree [None req-55b58374-2c35-46b9-a508-d4833264ecac tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Updating inventory in ProviderTree for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 177, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 875.971485] env[61923]: DEBUG nova.scheduler.client.report [None req-55b58374-2c35-46b9-a508-d4833264ecac tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Refreshing aggregate associations for resource provider f81803f2-f7f8-4939-a757-a77d34a1d0a2, aggregates: None {{(pid=61923) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 875.998036] env[61923]: DEBUG nova.scheduler.client.report [None req-55b58374-2c35-46b9-a508-d4833264ecac tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Refreshing trait associations for resource provider f81803f2-f7f8-4939-a757-a77d34a1d0a2, traits: COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NODE,HW_ARCH_X86_64,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_IMAGE_TYPE_ISO {{(pid=61923) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 876.041707] env[61923]: DEBUG oslo_concurrency.lockutils [None req-59b20242-d3b6-4451-b636-b9e52b91d9f2 tempest-ServerAddressesTestJSON-1846886542 tempest-ServerAddressesTestJSON-1846886542-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 876.064454] env[61923]: DEBUG nova.compute.manager [None req-6d4df172-3edf-4dbb-b96f-afa60a7b76e1 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] [instance: a511392e-9ab6-42fb-b07c-f90e9786dcc1] Start spawning the instance on the hypervisor. {{(pid=61923) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 876.067074] env[61923]: DEBUG oslo_vmware.api [None req-9887b108-13fe-4987-ad6f-1ef139e2a15d tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Task: {'id': task-1377672, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.103180] env[61923]: DEBUG nova.virt.hardware [None req-6d4df172-3edf-4dbb-b96f-afa60a7b76e1 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-29T20:07:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-29T20:07:35Z,direct_url=,disk_format='vmdk',id=0f153f63-ae0a-45b1-b7f5-7f9b673c947f,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='4e7b5e3b3c3443c8bd5c70f8a15739f5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-29T20:07:36Z,virtual_size=,visibility=), allow threads: False {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 876.105739] env[61923]: DEBUG nova.virt.hardware [None req-6d4df172-3edf-4dbb-b96f-afa60a7b76e1 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Flavor limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 876.105739] env[61923]: DEBUG nova.virt.hardware [None req-6d4df172-3edf-4dbb-b96f-afa60a7b76e1 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Image limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 876.105739] env[61923]: DEBUG nova.virt.hardware [None req-6d4df172-3edf-4dbb-b96f-afa60a7b76e1 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Flavor pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 876.105739] env[61923]: DEBUG nova.virt.hardware [None req-6d4df172-3edf-4dbb-b96f-afa60a7b76e1 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Image pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 876.105739] env[61923]: DEBUG nova.virt.hardware [None req-6d4df172-3edf-4dbb-b96f-afa60a7b76e1 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 876.105739] env[61923]: DEBUG nova.virt.hardware [None req-6d4df172-3edf-4dbb-b96f-afa60a7b76e1 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 876.105739] env[61923]: DEBUG nova.virt.hardware [None req-6d4df172-3edf-4dbb-b96f-afa60a7b76e1 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 876.105739] env[61923]: DEBUG nova.virt.hardware [None req-6d4df172-3edf-4dbb-b96f-afa60a7b76e1 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Got 1 possible topologies {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 876.106358] env[61923]: DEBUG nova.virt.hardware [None req-6d4df172-3edf-4dbb-b96f-afa60a7b76e1 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 876.106390] env[61923]: DEBUG nova.virt.hardware [None req-6d4df172-3edf-4dbb-b96f-afa60a7b76e1 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 876.107458] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b45b27b7-54e6-41e4-a35b-af7e631a073c {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.121052] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2854afda-2338-4790-8c27-e0077da39cf3 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.237456] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2c4d8a4-0119-40f3-aff3-e9c63bf9c74a {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.245975] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-add728d9-c8a7-4c7a-8936-9fe95e6475fa {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.277866] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1b51683-a97e-46e7-b6aa-baa22970a047 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.288089] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4990f2b7-6fc3-4cd1-9084-0531ac315daa {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.302729] env[61923]: DEBUG nova.compute.provider_tree [None req-55b58374-2c35-46b9-a508-d4833264ecac tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 876.305019] env[61923]: DEBUG oslo_concurrency.lockutils [None req-3192dd2d-7513-4a93-8a8c-c00e6e67889c tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 876.562671] env[61923]: DEBUG oslo_vmware.api [None req-9887b108-13fe-4987-ad6f-1ef139e2a15d tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Task: {'id': task-1377672, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.697521} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 876.562986] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-9887b108-13fe-4987-ad6f-1ef139e2a15d tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk to [datastore2] 6012d3b5-739c-4762-9bb4-09c51171dcd7/6012d3b5-739c-4762-9bb4-09c51171dcd7.vmdk {{(pid=61923) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 876.563316] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-9887b108-13fe-4987-ad6f-1ef139e2a15d tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] [instance: 6012d3b5-739c-4762-9bb4-09c51171dcd7] Extending root virtual disk to 1048576 {{(pid=61923) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 876.563650] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ae686055-6508-4846-98ed-46251b2d341a {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.572798] env[61923]: DEBUG oslo_vmware.api [None req-9887b108-13fe-4987-ad6f-1ef139e2a15d tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Waiting for the task: (returnval){ [ 876.572798] env[61923]: value = "task-1377673" [ 876.572798] env[61923]: _type = "Task" [ 876.572798] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 876.586796] env[61923]: DEBUG oslo_vmware.api [None req-9887b108-13fe-4987-ad6f-1ef139e2a15d tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Task: {'id': task-1377673, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.683718] env[61923]: DEBUG nova.network.neutron [None req-6d4df172-3edf-4dbb-b96f-afa60a7b76e1 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] [instance: a511392e-9ab6-42fb-b07c-f90e9786dcc1] Successfully updated port: 0325cd38-bd49-4132-85e3-c66fc9efe49f {{(pid=61923) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 876.806956] env[61923]: DEBUG nova.scheduler.client.report [None req-55b58374-2c35-46b9-a508-d4833264ecac tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 177, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 876.830890] env[61923]: DEBUG nova.network.neutron [req-4277850f-faad-47be-a133-01d84ea4c920 req-fba41451-c310-4eb0-842b-3cbb262c265b service nova] [instance: b779d183-89ae-4e4d-ae99-e514e145ed43] Updated VIF entry in instance network info cache for port d49a3010-c84a-4bef-beed-9ca4aa7d5dc2. {{(pid=61923) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 876.831284] env[61923]: DEBUG nova.network.neutron [req-4277850f-faad-47be-a133-01d84ea4c920 req-fba41451-c310-4eb0-842b-3cbb262c265b service nova] [instance: b779d183-89ae-4e4d-ae99-e514e145ed43] Updating instance_info_cache with network_info: [{"id": "d49a3010-c84a-4bef-beed-9ca4aa7d5dc2", "address": "fa:16:3e:48:e6:a2", "network": {"id": "d2905082-7ca4-4f92-a142-bb85211bb7c4", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-885218622-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.251", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a13d827933fa4597984afdb91dbbdd39", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "03ac2c9c-6ad2-4a85-bfab-c7e336df859a", "external-id": "nsx-vlan-transportzone-379", "segmentation_id": 379, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd49a3010-c8", "ovs_interfaceid": "d49a3010-c84a-4bef-beed-9ca4aa7d5dc2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 877.084419] env[61923]: DEBUG oslo_vmware.api [None req-9887b108-13fe-4987-ad6f-1ef139e2a15d tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Task: {'id': task-1377673, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.082857} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 877.084419] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-9887b108-13fe-4987-ad6f-1ef139e2a15d tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] [instance: 6012d3b5-739c-4762-9bb4-09c51171dcd7] Extended root virtual disk {{(pid=61923) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 877.084949] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e27b392-5515-4d98-ac48-ded58823a0f0 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.107810] env[61923]: DEBUG nova.virt.vmwareapi.volumeops [None req-9887b108-13fe-4987-ad6f-1ef139e2a15d tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] [instance: 6012d3b5-739c-4762-9bb4-09c51171dcd7] Reconfiguring VM instance instance-0000004e to attach disk [datastore2] 6012d3b5-739c-4762-9bb4-09c51171dcd7/6012d3b5-739c-4762-9bb4-09c51171dcd7.vmdk or device None with type sparse {{(pid=61923) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 877.108468] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9fe59f88-6f09-4a5c-aae3-c6f75c5b14b4 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.145055] env[61923]: DEBUG oslo_vmware.api [None req-9887b108-13fe-4987-ad6f-1ef139e2a15d tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Waiting for the task: (returnval){ [ 877.145055] env[61923]: value = "task-1377674" [ 877.145055] env[61923]: _type = "Task" [ 877.145055] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 877.156844] env[61923]: DEBUG oslo_vmware.api [None req-9887b108-13fe-4987-ad6f-1ef139e2a15d tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Task: {'id': task-1377674, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.185759] env[61923]: DEBUG oslo_concurrency.lockutils [None req-6d4df172-3edf-4dbb-b96f-afa60a7b76e1 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Acquiring lock "refresh_cache-a511392e-9ab6-42fb-b07c-f90e9786dcc1" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 877.185899] env[61923]: DEBUG oslo_concurrency.lockutils [None req-6d4df172-3edf-4dbb-b96f-afa60a7b76e1 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Acquired lock "refresh_cache-a511392e-9ab6-42fb-b07c-f90e9786dcc1" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 877.186070] env[61923]: DEBUG nova.network.neutron [None req-6d4df172-3edf-4dbb-b96f-afa60a7b76e1 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] [instance: a511392e-9ab6-42fb-b07c-f90e9786dcc1] Building network info cache for instance {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 877.311823] env[61923]: DEBUG oslo_concurrency.lockutils [None req-55b58374-2c35-46b9-a508-d4833264ecac tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.284s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 877.312463] env[61923]: DEBUG nova.compute.manager [None req-55b58374-2c35-46b9-a508-d4833264ecac tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] [instance: 4cb3c397-ef1a-4267-a300-26da48206f3f] Start building networks asynchronously for instance. {{(pid=61923) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 877.315418] env[61923]: DEBUG oslo_concurrency.lockutils [None req-eb5e01f1-148c-4608-96e7-93572113f79a tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 3.006s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 877.317378] env[61923]: INFO nova.compute.claims [None req-eb5e01f1-148c-4608-96e7-93572113f79a tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: 4c2c973a-5d23-4b8a-8b99-6b634e337054] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 877.334379] env[61923]: DEBUG oslo_concurrency.lockutils [req-4277850f-faad-47be-a133-01d84ea4c920 req-fba41451-c310-4eb0-842b-3cbb262c265b service nova] Releasing lock "refresh_cache-b779d183-89ae-4e4d-ae99-e514e145ed43" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 877.615046] env[61923]: DEBUG nova.compute.manager [req-baf58d7d-026c-4342-8b21-338c5ca062ef req-b4ea68e1-3274-4800-87c1-ade01eaebb8c service nova] [instance: a511392e-9ab6-42fb-b07c-f90e9786dcc1] Received event network-vif-plugged-0325cd38-bd49-4132-85e3-c66fc9efe49f {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 877.615153] env[61923]: DEBUG oslo_concurrency.lockutils [req-baf58d7d-026c-4342-8b21-338c5ca062ef req-b4ea68e1-3274-4800-87c1-ade01eaebb8c service nova] Acquiring lock "a511392e-9ab6-42fb-b07c-f90e9786dcc1-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 877.615340] env[61923]: DEBUG oslo_concurrency.lockutils [req-baf58d7d-026c-4342-8b21-338c5ca062ef req-b4ea68e1-3274-4800-87c1-ade01eaebb8c service nova] Lock "a511392e-9ab6-42fb-b07c-f90e9786dcc1-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 877.615557] env[61923]: DEBUG oslo_concurrency.lockutils [req-baf58d7d-026c-4342-8b21-338c5ca062ef req-b4ea68e1-3274-4800-87c1-ade01eaebb8c service nova] Lock "a511392e-9ab6-42fb-b07c-f90e9786dcc1-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 877.615754] env[61923]: DEBUG nova.compute.manager [req-baf58d7d-026c-4342-8b21-338c5ca062ef req-b4ea68e1-3274-4800-87c1-ade01eaebb8c service nova] [instance: a511392e-9ab6-42fb-b07c-f90e9786dcc1] No waiting events found dispatching network-vif-plugged-0325cd38-bd49-4132-85e3-c66fc9efe49f {{(pid=61923) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 877.615939] env[61923]: WARNING nova.compute.manager [req-baf58d7d-026c-4342-8b21-338c5ca062ef req-b4ea68e1-3274-4800-87c1-ade01eaebb8c service nova] [instance: a511392e-9ab6-42fb-b07c-f90e9786dcc1] Received unexpected event network-vif-plugged-0325cd38-bd49-4132-85e3-c66fc9efe49f for instance with vm_state building and task_state spawning. [ 877.616158] env[61923]: DEBUG nova.compute.manager [req-baf58d7d-026c-4342-8b21-338c5ca062ef req-b4ea68e1-3274-4800-87c1-ade01eaebb8c service nova] [instance: a511392e-9ab6-42fb-b07c-f90e9786dcc1] Received event network-changed-0325cd38-bd49-4132-85e3-c66fc9efe49f {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 877.616319] env[61923]: DEBUG nova.compute.manager [req-baf58d7d-026c-4342-8b21-338c5ca062ef req-b4ea68e1-3274-4800-87c1-ade01eaebb8c service nova] [instance: a511392e-9ab6-42fb-b07c-f90e9786dcc1] Refreshing instance network info cache due to event network-changed-0325cd38-bd49-4132-85e3-c66fc9efe49f. {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 877.616500] env[61923]: DEBUG oslo_concurrency.lockutils [req-baf58d7d-026c-4342-8b21-338c5ca062ef req-b4ea68e1-3274-4800-87c1-ade01eaebb8c service nova] Acquiring lock "refresh_cache-a511392e-9ab6-42fb-b07c-f90e9786dcc1" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 877.657245] env[61923]: DEBUG oslo_vmware.api [None req-9887b108-13fe-4987-ad6f-1ef139e2a15d tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Task: {'id': task-1377674, 'name': ReconfigVM_Task, 'duration_secs': 0.325928} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 877.657626] env[61923]: DEBUG nova.virt.vmwareapi.volumeops [None req-9887b108-13fe-4987-ad6f-1ef139e2a15d tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] [instance: 6012d3b5-739c-4762-9bb4-09c51171dcd7] Reconfigured VM instance instance-0000004e to attach disk [datastore2] 6012d3b5-739c-4762-9bb4-09c51171dcd7/6012d3b5-739c-4762-9bb4-09c51171dcd7.vmdk or device None with type sparse {{(pid=61923) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 877.658278] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ff957554-3e83-4d5c-b805-eea339a90acf {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.667073] env[61923]: DEBUG oslo_vmware.api [None req-9887b108-13fe-4987-ad6f-1ef139e2a15d tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Waiting for the task: (returnval){ [ 877.667073] env[61923]: value = "task-1377675" [ 877.667073] env[61923]: _type = "Task" [ 877.667073] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 877.676423] env[61923]: DEBUG oslo_vmware.api [None req-9887b108-13fe-4987-ad6f-1ef139e2a15d tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Task: {'id': task-1377675, 'name': Rename_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.721310] env[61923]: DEBUG nova.network.neutron [None req-6d4df172-3edf-4dbb-b96f-afa60a7b76e1 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] [instance: a511392e-9ab6-42fb-b07c-f90e9786dcc1] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 877.822907] env[61923]: DEBUG nova.compute.utils [None req-55b58374-2c35-46b9-a508-d4833264ecac tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Using /dev/sd instead of None {{(pid=61923) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 877.827610] env[61923]: DEBUG nova.compute.manager [None req-55b58374-2c35-46b9-a508-d4833264ecac tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] [instance: 4cb3c397-ef1a-4267-a300-26da48206f3f] Allocating IP information in the background. {{(pid=61923) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 877.827610] env[61923]: DEBUG nova.network.neutron [None req-55b58374-2c35-46b9-a508-d4833264ecac tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] [instance: 4cb3c397-ef1a-4267-a300-26da48206f3f] allocate_for_instance() {{(pid=61923) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 877.873867] env[61923]: DEBUG nova.network.neutron [None req-6d4df172-3edf-4dbb-b96f-afa60a7b76e1 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] [instance: a511392e-9ab6-42fb-b07c-f90e9786dcc1] Updating instance_info_cache with network_info: [{"id": "0325cd38-bd49-4132-85e3-c66fc9efe49f", "address": "fa:16:3e:7f:89:31", "network": {"id": "07fc148e-c3ba-4e53-bec2-36500ec38add", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1374533437-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0d24f4a47d1f4b79a7d999a4dacb99ba", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b2907cce-d529-4809-af05-d29397bed211", "external-id": "nsx-vlan-transportzone-427", "segmentation_id": 427, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0325cd38-bd", "ovs_interfaceid": "0325cd38-bd49-4132-85e3-c66fc9efe49f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 877.887346] env[61923]: DEBUG nova.policy [None req-55b58374-2c35-46b9-a508-d4833264ecac tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'eb00f8aff92b44e9a0e739366a1d2662', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7abfe95660904cd4b8824257795d8e8a', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61923) authorize /opt/stack/nova/nova/policy.py:201}} [ 878.167298] env[61923]: DEBUG nova.network.neutron [None req-55b58374-2c35-46b9-a508-d4833264ecac tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] [instance: 4cb3c397-ef1a-4267-a300-26da48206f3f] Successfully created port: 83e705c4-d432-4fdf-8486-d31b46e1e6a5 {{(pid=61923) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 878.184213] env[61923]: DEBUG oslo_vmware.api [None req-9887b108-13fe-4987-ad6f-1ef139e2a15d tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Task: {'id': task-1377675, 'name': Rename_Task, 'duration_secs': 0.151792} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 878.184511] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-9887b108-13fe-4987-ad6f-1ef139e2a15d tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] [instance: 6012d3b5-739c-4762-9bb4-09c51171dcd7] Powering on the VM {{(pid=61923) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 878.184755] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-01ebd518-7f32-4f35-844b-c69150c710c6 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.195951] env[61923]: DEBUG oslo_vmware.api [None req-9887b108-13fe-4987-ad6f-1ef139e2a15d tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Waiting for the task: (returnval){ [ 878.195951] env[61923]: value = "task-1377676" [ 878.195951] env[61923]: _type = "Task" [ 878.195951] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 878.204537] env[61923]: DEBUG oslo_vmware.api [None req-9887b108-13fe-4987-ad6f-1ef139e2a15d tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Task: {'id': task-1377676, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.253222] env[61923]: DEBUG oslo_vmware.rw_handles [None req-12656253-9def-48f0-9500-888c645ef4dc tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5235c29c-aba2-46b7-1408-4e88ce392c19/disk-0.vmdk. {{(pid=61923) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 878.254738] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5a32c24-7d82-4685-ac9d-dc18bfccf3cc {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.261498] env[61923]: DEBUG oslo_vmware.rw_handles [None req-12656253-9def-48f0-9500-888c645ef4dc tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5235c29c-aba2-46b7-1408-4e88ce392c19/disk-0.vmdk is in state: ready. {{(pid=61923) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 878.261768] env[61923]: ERROR oslo_vmware.rw_handles [None req-12656253-9def-48f0-9500-888c645ef4dc tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Aborting lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5235c29c-aba2-46b7-1408-4e88ce392c19/disk-0.vmdk due to incomplete transfer. [ 878.261891] env[61923]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-cd0e3b44-f032-4b0c-86ba-9f4ccfb5973e {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.270063] env[61923]: DEBUG oslo_vmware.rw_handles [None req-12656253-9def-48f0-9500-888c645ef4dc tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Closed VMDK read handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/5235c29c-aba2-46b7-1408-4e88ce392c19/disk-0.vmdk. {{(pid=61923) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 878.270327] env[61923]: DEBUG nova.virt.vmwareapi.images [None req-12656253-9def-48f0-9500-888c645ef4dc tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] [instance: 87ae37dc-d0d0-4b76-afdd-0ba3e8f6ce0b] Uploaded image cb483b67-0b09-46ef-8d9d-6fb9b7e0f932 to the Glance image server {{(pid=61923) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 878.271964] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-12656253-9def-48f0-9500-888c645ef4dc tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] [instance: 87ae37dc-d0d0-4b76-afdd-0ba3e8f6ce0b] Destroying the VM {{(pid=61923) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 878.273280] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-76c9753e-3f0f-4595-bb0d-8559feae8563 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.281713] env[61923]: DEBUG oslo_vmware.api [None req-12656253-9def-48f0-9500-888c645ef4dc tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Waiting for the task: (returnval){ [ 878.281713] env[61923]: value = "task-1377677" [ 878.281713] env[61923]: _type = "Task" [ 878.281713] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 878.295814] env[61923]: DEBUG oslo_vmware.api [None req-12656253-9def-48f0-9500-888c645ef4dc tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Task: {'id': task-1377677, 'name': Destroy_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.327673] env[61923]: DEBUG nova.compute.manager [None req-55b58374-2c35-46b9-a508-d4833264ecac tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] [instance: 4cb3c397-ef1a-4267-a300-26da48206f3f] Start building block device mappings for instance. {{(pid=61923) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 878.382714] env[61923]: DEBUG oslo_concurrency.lockutils [None req-6d4df172-3edf-4dbb-b96f-afa60a7b76e1 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Releasing lock "refresh_cache-a511392e-9ab6-42fb-b07c-f90e9786dcc1" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 878.383106] env[61923]: DEBUG nova.compute.manager [None req-6d4df172-3edf-4dbb-b96f-afa60a7b76e1 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] [instance: a511392e-9ab6-42fb-b07c-f90e9786dcc1] Instance network_info: |[{"id": "0325cd38-bd49-4132-85e3-c66fc9efe49f", "address": "fa:16:3e:7f:89:31", "network": {"id": "07fc148e-c3ba-4e53-bec2-36500ec38add", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1374533437-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0d24f4a47d1f4b79a7d999a4dacb99ba", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b2907cce-d529-4809-af05-d29397bed211", "external-id": "nsx-vlan-transportzone-427", "segmentation_id": 427, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0325cd38-bd", "ovs_interfaceid": "0325cd38-bd49-4132-85e3-c66fc9efe49f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61923) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 878.383744] env[61923]: DEBUG oslo_concurrency.lockutils [req-baf58d7d-026c-4342-8b21-338c5ca062ef req-b4ea68e1-3274-4800-87c1-ade01eaebb8c service nova] Acquired lock "refresh_cache-a511392e-9ab6-42fb-b07c-f90e9786dcc1" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 878.383940] env[61923]: DEBUG nova.network.neutron [req-baf58d7d-026c-4342-8b21-338c5ca062ef req-b4ea68e1-3274-4800-87c1-ade01eaebb8c service nova] [instance: a511392e-9ab6-42fb-b07c-f90e9786dcc1] Refreshing network info cache for port 0325cd38-bd49-4132-85e3-c66fc9efe49f {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 878.385253] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-6d4df172-3edf-4dbb-b96f-afa60a7b76e1 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] [instance: a511392e-9ab6-42fb-b07c-f90e9786dcc1] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:7f:89:31', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'b2907cce-d529-4809-af05-d29397bed211', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '0325cd38-bd49-4132-85e3-c66fc9efe49f', 'vif_model': 'vmxnet3'}] {{(pid=61923) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 878.399100] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-6d4df172-3edf-4dbb-b96f-afa60a7b76e1 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Creating folder: Project (0d24f4a47d1f4b79a7d999a4dacb99ba). Parent ref: group-v292629. {{(pid=61923) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 878.405112] env[61923]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-473d34f6-4df0-43b5-a9e0-222fcbdb7e3c {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.422055] env[61923]: INFO nova.virt.vmwareapi.vm_util [None req-6d4df172-3edf-4dbb-b96f-afa60a7b76e1 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Created folder: Project (0d24f4a47d1f4b79a7d999a4dacb99ba) in parent group-v292629. [ 878.422055] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-6d4df172-3edf-4dbb-b96f-afa60a7b76e1 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Creating folder: Instances. Parent ref: group-v292697. {{(pid=61923) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 878.422055] env[61923]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-65e0453b-ccee-485f-9a45-feaec58844ff {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.434305] env[61923]: INFO nova.virt.vmwareapi.vm_util [None req-6d4df172-3edf-4dbb-b96f-afa60a7b76e1 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Created folder: Instances in parent group-v292697. [ 878.434576] env[61923]: DEBUG oslo.service.loopingcall [None req-6d4df172-3edf-4dbb-b96f-afa60a7b76e1 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61923) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 878.434781] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a511392e-9ab6-42fb-b07c-f90e9786dcc1] Creating VM on the ESX host {{(pid=61923) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 878.435009] env[61923]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-52715640-0ed5-4492-b63a-3e3d7a8aeb59 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.466995] env[61923]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 878.466995] env[61923]: value = "task-1377680" [ 878.466995] env[61923]: _type = "Task" [ 878.466995] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 878.476826] env[61923]: DEBUG oslo_vmware.api [-] Task: {'id': task-1377680, 'name': CreateVM_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.640055] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99151fed-68be-4f6d-b00d-654ac80eb64e {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.650239] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0583b1eb-36c5-4e5e-9be0-cc8c61525adc {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.692219] env[61923]: DEBUG nova.network.neutron [req-baf58d7d-026c-4342-8b21-338c5ca062ef req-b4ea68e1-3274-4800-87c1-ade01eaebb8c service nova] [instance: a511392e-9ab6-42fb-b07c-f90e9786dcc1] Updated VIF entry in instance network info cache for port 0325cd38-bd49-4132-85e3-c66fc9efe49f. {{(pid=61923) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 878.692812] env[61923]: DEBUG nova.network.neutron [req-baf58d7d-026c-4342-8b21-338c5ca062ef req-b4ea68e1-3274-4800-87c1-ade01eaebb8c service nova] [instance: a511392e-9ab6-42fb-b07c-f90e9786dcc1] Updating instance_info_cache with network_info: [{"id": "0325cd38-bd49-4132-85e3-c66fc9efe49f", "address": "fa:16:3e:7f:89:31", "network": {"id": "07fc148e-c3ba-4e53-bec2-36500ec38add", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1374533437-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0d24f4a47d1f4b79a7d999a4dacb99ba", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b2907cce-d529-4809-af05-d29397bed211", "external-id": "nsx-vlan-transportzone-427", "segmentation_id": 427, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0325cd38-bd", "ovs_interfaceid": "0325cd38-bd49-4132-85e3-c66fc9efe49f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 878.694743] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3e46b1e-304e-4b3f-b90e-1b4d0e652b98 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.712215] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-187a136b-4168-4856-812f-84afac68c43b {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.719236] env[61923]: DEBUG oslo_vmware.api [None req-9887b108-13fe-4987-ad6f-1ef139e2a15d tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Task: {'id': task-1377676, 'name': PowerOnVM_Task} progress is 92%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.731389] env[61923]: DEBUG nova.compute.provider_tree [None req-eb5e01f1-148c-4608-96e7-93572113f79a tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 878.796959] env[61923]: DEBUG oslo_vmware.api [None req-12656253-9def-48f0-9500-888c645ef4dc tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Task: {'id': task-1377677, 'name': Destroy_Task, 'duration_secs': 0.513167} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 878.797436] env[61923]: INFO nova.virt.vmwareapi.vm_util [None req-12656253-9def-48f0-9500-888c645ef4dc tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] [instance: 87ae37dc-d0d0-4b76-afdd-0ba3e8f6ce0b] Destroyed the VM [ 878.797799] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-12656253-9def-48f0-9500-888c645ef4dc tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] [instance: 87ae37dc-d0d0-4b76-afdd-0ba3e8f6ce0b] Deleting Snapshot of the VM instance {{(pid=61923) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 878.798199] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-6126068a-4f3d-4e8a-85b3-1b57e23daba3 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.806709] env[61923]: DEBUG oslo_vmware.api [None req-12656253-9def-48f0-9500-888c645ef4dc tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Waiting for the task: (returnval){ [ 878.806709] env[61923]: value = "task-1377681" [ 878.806709] env[61923]: _type = "Task" [ 878.806709] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 878.818327] env[61923]: DEBUG oslo_vmware.api [None req-12656253-9def-48f0-9500-888c645ef4dc tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Task: {'id': task-1377681, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.977389] env[61923]: DEBUG oslo_vmware.api [-] Task: {'id': task-1377680, 'name': CreateVM_Task, 'duration_secs': 0.437615} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 878.977586] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a511392e-9ab6-42fb-b07c-f90e9786dcc1] Created VM on the ESX host {{(pid=61923) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 878.978293] env[61923]: DEBUG oslo_concurrency.lockutils [None req-6d4df172-3edf-4dbb-b96f-afa60a7b76e1 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 878.978468] env[61923]: DEBUG oslo_concurrency.lockutils [None req-6d4df172-3edf-4dbb-b96f-afa60a7b76e1 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 878.978797] env[61923]: DEBUG oslo_concurrency.lockutils [None req-6d4df172-3edf-4dbb-b96f-afa60a7b76e1 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 878.979094] env[61923]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3fbf3da9-8a82-4837-948f-f94eb2275470 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.983559] env[61923]: DEBUG oslo_vmware.api [None req-6d4df172-3edf-4dbb-b96f-afa60a7b76e1 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Waiting for the task: (returnval){ [ 878.983559] env[61923]: value = "session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]5227ddfc-fdbd-71ab-867b-b2c717bee843" [ 878.983559] env[61923]: _type = "Task" [ 878.983559] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 878.991023] env[61923]: DEBUG oslo_vmware.api [None req-6d4df172-3edf-4dbb-b96f-afa60a7b76e1 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]5227ddfc-fdbd-71ab-867b-b2c717bee843, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.199309] env[61923]: DEBUG oslo_concurrency.lockutils [req-baf58d7d-026c-4342-8b21-338c5ca062ef req-b4ea68e1-3274-4800-87c1-ade01eaebb8c service nova] Releasing lock "refresh_cache-a511392e-9ab6-42fb-b07c-f90e9786dcc1" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 879.210435] env[61923]: DEBUG oslo_vmware.api [None req-9887b108-13fe-4987-ad6f-1ef139e2a15d tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Task: {'id': task-1377676, 'name': PowerOnVM_Task, 'duration_secs': 0.657023} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 879.211016] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-9887b108-13fe-4987-ad6f-1ef139e2a15d tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] [instance: 6012d3b5-739c-4762-9bb4-09c51171dcd7] Powered on the VM {{(pid=61923) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 879.211515] env[61923]: INFO nova.compute.manager [None req-9887b108-13fe-4987-ad6f-1ef139e2a15d tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] [instance: 6012d3b5-739c-4762-9bb4-09c51171dcd7] Took 9.01 seconds to spawn the instance on the hypervisor. [ 879.211799] env[61923]: DEBUG nova.compute.manager [None req-9887b108-13fe-4987-ad6f-1ef139e2a15d tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] [instance: 6012d3b5-739c-4762-9bb4-09c51171dcd7] Checking state {{(pid=61923) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 879.212822] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f712191d-1470-4df7-9322-8c6522bec86d {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.234838] env[61923]: DEBUG nova.scheduler.client.report [None req-eb5e01f1-148c-4608-96e7-93572113f79a tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 177, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 879.317445] env[61923]: DEBUG oslo_vmware.api [None req-12656253-9def-48f0-9500-888c645ef4dc tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Task: {'id': task-1377681, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.341183] env[61923]: DEBUG nova.compute.manager [None req-55b58374-2c35-46b9-a508-d4833264ecac tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] [instance: 4cb3c397-ef1a-4267-a300-26da48206f3f] Start spawning the instance on the hypervisor. {{(pid=61923) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 879.377935] env[61923]: DEBUG nova.virt.hardware [None req-55b58374-2c35-46b9-a508-d4833264ecac tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-29T20:07:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='7fc091dd47178c1508df3f609d01ee7d',container_format='bare',created_at=2024-10-29T20:15:10Z,direct_url=,disk_format='vmdk',id=9ddad7e6-f1e4-49aa-b956-1b8d3273318d,min_disk=1,min_ram=0,name='tempest-test-snap-2025648207',owner='7abfe95660904cd4b8824257795d8e8a',properties=ImageMetaProps,protected=,size=21334016,status='active',tags=,updated_at=2024-10-29T20:15:27Z,virtual_size=,visibility=), allow threads: False {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 879.378281] env[61923]: DEBUG nova.virt.hardware [None req-55b58374-2c35-46b9-a508-d4833264ecac tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Flavor limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 879.378457] env[61923]: DEBUG nova.virt.hardware [None req-55b58374-2c35-46b9-a508-d4833264ecac tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Image limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 879.378645] env[61923]: DEBUG nova.virt.hardware [None req-55b58374-2c35-46b9-a508-d4833264ecac tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Flavor pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 879.378826] env[61923]: DEBUG nova.virt.hardware [None req-55b58374-2c35-46b9-a508-d4833264ecac tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Image pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 879.378985] env[61923]: DEBUG nova.virt.hardware [None req-55b58374-2c35-46b9-a508-d4833264ecac tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 879.379255] env[61923]: DEBUG nova.virt.hardware [None req-55b58374-2c35-46b9-a508-d4833264ecac tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 879.379470] env[61923]: DEBUG nova.virt.hardware [None req-55b58374-2c35-46b9-a508-d4833264ecac tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 879.379656] env[61923]: DEBUG nova.virt.hardware [None req-55b58374-2c35-46b9-a508-d4833264ecac tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Got 1 possible topologies {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 879.379828] env[61923]: DEBUG nova.virt.hardware [None req-55b58374-2c35-46b9-a508-d4833264ecac tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 879.380012] env[61923]: DEBUG nova.virt.hardware [None req-55b58374-2c35-46b9-a508-d4833264ecac tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 879.381091] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3605d470-30ed-497c-8036-48dcb3b8cbe4 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.393011] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82413d56-5e55-4d27-8ed3-7ae9e440d4dc {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.497178] env[61923]: DEBUG oslo_vmware.api [None req-6d4df172-3edf-4dbb-b96f-afa60a7b76e1 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]5227ddfc-fdbd-71ab-867b-b2c717bee843, 'name': SearchDatastore_Task, 'duration_secs': 0.038851} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 879.497651] env[61923]: DEBUG oslo_concurrency.lockutils [None req-6d4df172-3edf-4dbb-b96f-afa60a7b76e1 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 879.498015] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-6d4df172-3edf-4dbb-b96f-afa60a7b76e1 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] [instance: a511392e-9ab6-42fb-b07c-f90e9786dcc1] Processing image 0f153f63-ae0a-45b1-b7f5-7f9b673c947f {{(pid=61923) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 879.498419] env[61923]: DEBUG oslo_concurrency.lockutils [None req-6d4df172-3edf-4dbb-b96f-afa60a7b76e1 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 879.498669] env[61923]: DEBUG oslo_concurrency.lockutils [None req-6d4df172-3edf-4dbb-b96f-afa60a7b76e1 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 879.498958] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-6d4df172-3edf-4dbb-b96f-afa60a7b76e1 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61923) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 879.499362] env[61923]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-8fac78a4-ac2d-4a23-b54c-ad05c0dece19 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.510133] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-6d4df172-3edf-4dbb-b96f-afa60a7b76e1 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61923) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 879.510444] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-6d4df172-3edf-4dbb-b96f-afa60a7b76e1 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61923) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 879.511548] env[61923]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f4fd1834-b8fb-49d3-afd7-9edc4b720c21 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.518914] env[61923]: DEBUG oslo_vmware.api [None req-6d4df172-3edf-4dbb-b96f-afa60a7b76e1 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Waiting for the task: (returnval){ [ 879.518914] env[61923]: value = "session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]5201cc83-1b65-1c12-e1d8-6ff8f555b489" [ 879.518914] env[61923]: _type = "Task" [ 879.518914] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 879.532039] env[61923]: DEBUG oslo_vmware.api [None req-6d4df172-3edf-4dbb-b96f-afa60a7b76e1 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]5201cc83-1b65-1c12-e1d8-6ff8f555b489, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.736402] env[61923]: INFO nova.compute.manager [None req-9887b108-13fe-4987-ad6f-1ef139e2a15d tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] [instance: 6012d3b5-739c-4762-9bb4-09c51171dcd7] Took 27.96 seconds to build instance. [ 879.742836] env[61923]: DEBUG oslo_concurrency.lockutils [None req-eb5e01f1-148c-4608-96e7-93572113f79a tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.427s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 879.743391] env[61923]: DEBUG nova.compute.manager [None req-eb5e01f1-148c-4608-96e7-93572113f79a tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: 4c2c973a-5d23-4b8a-8b99-6b634e337054] Start building networks asynchronously for instance. {{(pid=61923) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 879.751169] env[61923]: DEBUG oslo_concurrency.lockutils [None req-7e66bb89-e0a5-4005-a075-6cc387f412c9 tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 4.552s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 879.751482] env[61923]: DEBUG nova.objects.instance [None req-7e66bb89-e0a5-4005-a075-6cc387f412c9 tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Lazy-loading 'resources' on Instance uuid 444affa5-a7ed-4a17-9015-9fd5724aab64 {{(pid=61923) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 879.818581] env[61923]: DEBUG oslo_vmware.api [None req-12656253-9def-48f0-9500-888c645ef4dc tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Task: {'id': task-1377681, 'name': RemoveSnapshot_Task, 'duration_secs': 0.952505} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 879.818727] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-12656253-9def-48f0-9500-888c645ef4dc tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] [instance: 87ae37dc-d0d0-4b76-afdd-0ba3e8f6ce0b] Deleted Snapshot of the VM instance {{(pid=61923) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 879.819254] env[61923]: INFO nova.compute.manager [None req-12656253-9def-48f0-9500-888c645ef4dc tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] [instance: 87ae37dc-d0d0-4b76-afdd-0ba3e8f6ce0b] Took 15.48 seconds to snapshot the instance on the hypervisor. [ 880.030160] env[61923]: DEBUG oslo_vmware.api [None req-6d4df172-3edf-4dbb-b96f-afa60a7b76e1 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]5201cc83-1b65-1c12-e1d8-6ff8f555b489, 'name': SearchDatastore_Task, 'duration_secs': 0.014922} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 880.030961] env[61923]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d39c9a86-0b90-444e-aaaf-2734a635d89e {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.037390] env[61923]: DEBUG oslo_vmware.api [None req-6d4df172-3edf-4dbb-b96f-afa60a7b76e1 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Waiting for the task: (returnval){ [ 880.037390] env[61923]: value = "session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52e74026-ef54-6c98-5f50-8b962eb0bfb0" [ 880.037390] env[61923]: _type = "Task" [ 880.037390] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 880.045379] env[61923]: DEBUG oslo_vmware.api [None req-6d4df172-3edf-4dbb-b96f-afa60a7b76e1 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52e74026-ef54-6c98-5f50-8b962eb0bfb0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.240127] env[61923]: DEBUG oslo_concurrency.lockutils [None req-9887b108-13fe-4987-ad6f-1ef139e2a15d tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Lock "6012d3b5-739c-4762-9bb4-09c51171dcd7" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 31.610s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 880.250824] env[61923]: DEBUG nova.compute.utils [None req-eb5e01f1-148c-4608-96e7-93572113f79a tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Using /dev/sd instead of None {{(pid=61923) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 880.256194] env[61923]: DEBUG nova.compute.manager [None req-eb5e01f1-148c-4608-96e7-93572113f79a tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: 4c2c973a-5d23-4b8a-8b99-6b634e337054] Allocating IP information in the background. {{(pid=61923) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 880.256405] env[61923]: DEBUG nova.network.neutron [None req-eb5e01f1-148c-4608-96e7-93572113f79a tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: 4c2c973a-5d23-4b8a-8b99-6b634e337054] allocate_for_instance() {{(pid=61923) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 880.293619] env[61923]: DEBUG nova.compute.manager [req-9e7e16b6-3f25-42e3-b7dc-55eaaa8c5e44 req-7b8dae6e-2842-4374-a56a-a551a35cb177 service nova] [instance: 4cb3c397-ef1a-4267-a300-26da48206f3f] Received event network-vif-plugged-83e705c4-d432-4fdf-8486-d31b46e1e6a5 {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 880.293761] env[61923]: DEBUG oslo_concurrency.lockutils [req-9e7e16b6-3f25-42e3-b7dc-55eaaa8c5e44 req-7b8dae6e-2842-4374-a56a-a551a35cb177 service nova] Acquiring lock "4cb3c397-ef1a-4267-a300-26da48206f3f-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 880.293945] env[61923]: DEBUG oslo_concurrency.lockutils [req-9e7e16b6-3f25-42e3-b7dc-55eaaa8c5e44 req-7b8dae6e-2842-4374-a56a-a551a35cb177 service nova] Lock "4cb3c397-ef1a-4267-a300-26da48206f3f-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 880.294238] env[61923]: DEBUG oslo_concurrency.lockutils [req-9e7e16b6-3f25-42e3-b7dc-55eaaa8c5e44 req-7b8dae6e-2842-4374-a56a-a551a35cb177 service nova] Lock "4cb3c397-ef1a-4267-a300-26da48206f3f-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 880.294457] env[61923]: DEBUG nova.compute.manager [req-9e7e16b6-3f25-42e3-b7dc-55eaaa8c5e44 req-7b8dae6e-2842-4374-a56a-a551a35cb177 service nova] [instance: 4cb3c397-ef1a-4267-a300-26da48206f3f] No waiting events found dispatching network-vif-plugged-83e705c4-d432-4fdf-8486-d31b46e1e6a5 {{(pid=61923) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 880.294693] env[61923]: WARNING nova.compute.manager [req-9e7e16b6-3f25-42e3-b7dc-55eaaa8c5e44 req-7b8dae6e-2842-4374-a56a-a551a35cb177 service nova] [instance: 4cb3c397-ef1a-4267-a300-26da48206f3f] Received unexpected event network-vif-plugged-83e705c4-d432-4fdf-8486-d31b46e1e6a5 for instance with vm_state building and task_state spawning. [ 880.331391] env[61923]: DEBUG nova.policy [None req-eb5e01f1-148c-4608-96e7-93572113f79a tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5b8a639e99914e75857b4571f0d58a3c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '69cc941eb6dd4780ac12aa29656c37f7', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61923) authorize /opt/stack/nova/nova/policy.py:201}} [ 880.487019] env[61923]: DEBUG nova.network.neutron [None req-55b58374-2c35-46b9-a508-d4833264ecac tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] [instance: 4cb3c397-ef1a-4267-a300-26da48206f3f] Successfully updated port: 83e705c4-d432-4fdf-8486-d31b46e1e6a5 {{(pid=61923) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 880.509871] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd7c58cb-7e94-4d0a-be03-f9f4408b6531 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.518780] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f0a7f21-a88a-4800-b543-94aa89e870c7 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.559511] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bfc8e876-7a37-4832-9e35-f8eed4eec301 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.574063] env[61923]: DEBUG oslo_vmware.api [None req-6d4df172-3edf-4dbb-b96f-afa60a7b76e1 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52e74026-ef54-6c98-5f50-8b962eb0bfb0, 'name': SearchDatastore_Task, 'duration_secs': 0.044606} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 880.576783] env[61923]: DEBUG oslo_concurrency.lockutils [None req-6d4df172-3edf-4dbb-b96f-afa60a7b76e1 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 880.577116] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-6d4df172-3edf-4dbb-b96f-afa60a7b76e1 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk to [datastore2] a511392e-9ab6-42fb-b07c-f90e9786dcc1/a511392e-9ab6-42fb-b07c-f90e9786dcc1.vmdk {{(pid=61923) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 880.578436] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb9eddf9-92f2-40a3-9ab4-e710e270a153 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.582634] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-9887ed13-4e64-4d7a-962c-0f88ab8fdd0d {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.597611] env[61923]: DEBUG nova.compute.provider_tree [None req-7e66bb89-e0a5-4005-a075-6cc387f412c9 tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Updating inventory in ProviderTree for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 880.600339] env[61923]: DEBUG oslo_vmware.api [None req-6d4df172-3edf-4dbb-b96f-afa60a7b76e1 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Waiting for the task: (returnval){ [ 880.600339] env[61923]: value = "task-1377682" [ 880.600339] env[61923]: _type = "Task" [ 880.600339] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 880.604465] env[61923]: DEBUG nova.network.neutron [None req-eb5e01f1-148c-4608-96e7-93572113f79a tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: 4c2c973a-5d23-4b8a-8b99-6b634e337054] Successfully created port: 9584dacf-1833-4a40-bfbe-56d7d5ea9998 {{(pid=61923) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 880.613308] env[61923]: DEBUG oslo_vmware.api [None req-6d4df172-3edf-4dbb-b96f-afa60a7b76e1 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Task: {'id': task-1377682, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.761826] env[61923]: DEBUG nova.compute.manager [None req-eb5e01f1-148c-4608-96e7-93572113f79a tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: 4c2c973a-5d23-4b8a-8b99-6b634e337054] Start building block device mappings for instance. {{(pid=61923) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 880.993262] env[61923]: DEBUG oslo_concurrency.lockutils [None req-55b58374-2c35-46b9-a508-d4833264ecac tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Acquiring lock "refresh_cache-4cb3c397-ef1a-4267-a300-26da48206f3f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 880.993405] env[61923]: DEBUG oslo_concurrency.lockutils [None req-55b58374-2c35-46b9-a508-d4833264ecac tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Acquired lock "refresh_cache-4cb3c397-ef1a-4267-a300-26da48206f3f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 880.993557] env[61923]: DEBUG nova.network.neutron [None req-55b58374-2c35-46b9-a508-d4833264ecac tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] [instance: 4cb3c397-ef1a-4267-a300-26da48206f3f] Building network info cache for instance {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 881.115165] env[61923]: DEBUG oslo_vmware.api [None req-6d4df172-3edf-4dbb-b96f-afa60a7b76e1 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Task: {'id': task-1377682, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.138062] env[61923]: DEBUG nova.scheduler.client.report [None req-7e66bb89-e0a5-4005-a075-6cc387f412c9 tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Updated inventory for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 with generation 91 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 881.138359] env[61923]: DEBUG nova.compute.provider_tree [None req-7e66bb89-e0a5-4005-a075-6cc387f412c9 tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Updating resource provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 generation from 91 to 92 during operation: update_inventory {{(pid=61923) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 881.141092] env[61923]: DEBUG nova.compute.provider_tree [None req-7e66bb89-e0a5-4005-a075-6cc387f412c9 tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Updating inventory in ProviderTree for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 881.535878] env[61923]: DEBUG nova.network.neutron [None req-55b58374-2c35-46b9-a508-d4833264ecac tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] [instance: 4cb3c397-ef1a-4267-a300-26da48206f3f] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 881.625402] env[61923]: DEBUG oslo_vmware.api [None req-6d4df172-3edf-4dbb-b96f-afa60a7b76e1 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Task: {'id': task-1377682, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.658431} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 881.625814] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-6d4df172-3edf-4dbb-b96f-afa60a7b76e1 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk to [datastore2] a511392e-9ab6-42fb-b07c-f90e9786dcc1/a511392e-9ab6-42fb-b07c-f90e9786dcc1.vmdk {{(pid=61923) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 881.626071] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-6d4df172-3edf-4dbb-b96f-afa60a7b76e1 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] [instance: a511392e-9ab6-42fb-b07c-f90e9786dcc1] Extending root virtual disk to 1048576 {{(pid=61923) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 881.627135] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-182c8ee6-fd3f-43b1-b998-2e688b1b707a {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.640269] env[61923]: DEBUG oslo_vmware.api [None req-6d4df172-3edf-4dbb-b96f-afa60a7b76e1 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Waiting for the task: (returnval){ [ 881.640269] env[61923]: value = "task-1377683" [ 881.640269] env[61923]: _type = "Task" [ 881.640269] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 881.647060] env[61923]: DEBUG oslo_concurrency.lockutils [None req-7e66bb89-e0a5-4005-a075-6cc387f412c9 tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.895s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 881.649355] env[61923]: DEBUG oslo_concurrency.lockutils [None req-89f7a2ee-0580-4457-b803-a0236f2ecef8 tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 6.288s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 881.649603] env[61923]: DEBUG nova.objects.instance [None req-89f7a2ee-0580-4457-b803-a0236f2ecef8 tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Lazy-loading 'resources' on Instance uuid e413c45d-cd89-44d4-9102-3d188907e7bb {{(pid=61923) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 881.657346] env[61923]: DEBUG oslo_vmware.api [None req-6d4df172-3edf-4dbb-b96f-afa60a7b76e1 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Task: {'id': task-1377683, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.672999] env[61923]: INFO nova.scheduler.client.report [None req-7e66bb89-e0a5-4005-a075-6cc387f412c9 tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Deleted allocations for instance 444affa5-a7ed-4a17-9015-9fd5724aab64 [ 881.774641] env[61923]: DEBUG nova.compute.manager [None req-eb5e01f1-148c-4608-96e7-93572113f79a tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: 4c2c973a-5d23-4b8a-8b99-6b634e337054] Start spawning the instance on the hypervisor. {{(pid=61923) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 881.788290] env[61923]: DEBUG nova.network.neutron [None req-55b58374-2c35-46b9-a508-d4833264ecac tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] [instance: 4cb3c397-ef1a-4267-a300-26da48206f3f] Updating instance_info_cache with network_info: [{"id": "83e705c4-d432-4fdf-8486-d31b46e1e6a5", "address": "fa:16:3e:81:47:77", "network": {"id": "b8c2e837-0d64-4b07-9cc5-a66c8e8db1e9", "bridge": "br-int", "label": "tempest-ImagesTestJSON-133215580-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7abfe95660904cd4b8824257795d8e8a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f2c424c9-6446-4b2a-af8c-4d9c29117c39", "external-id": "nsx-vlan-transportzone-437", "segmentation_id": 437, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap83e705c4-d4", "ovs_interfaceid": "83e705c4-d432-4fdf-8486-d31b46e1e6a5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 881.809116] env[61923]: DEBUG nova.virt.hardware [None req-eb5e01f1-148c-4608-96e7-93572113f79a tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-29T20:07:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-29T20:07:35Z,direct_url=,disk_format='vmdk',id=0f153f63-ae0a-45b1-b7f5-7f9b673c947f,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='4e7b5e3b3c3443c8bd5c70f8a15739f5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-29T20:07:36Z,virtual_size=,visibility=), allow threads: False {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 881.809661] env[61923]: DEBUG nova.virt.hardware [None req-eb5e01f1-148c-4608-96e7-93572113f79a tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Flavor limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 881.809841] env[61923]: DEBUG nova.virt.hardware [None req-eb5e01f1-148c-4608-96e7-93572113f79a tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Image limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 881.810321] env[61923]: DEBUG nova.virt.hardware [None req-eb5e01f1-148c-4608-96e7-93572113f79a tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Flavor pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 881.810620] env[61923]: DEBUG nova.virt.hardware [None req-eb5e01f1-148c-4608-96e7-93572113f79a tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Image pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 881.810711] env[61923]: DEBUG nova.virt.hardware [None req-eb5e01f1-148c-4608-96e7-93572113f79a tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 881.810938] env[61923]: DEBUG nova.virt.hardware [None req-eb5e01f1-148c-4608-96e7-93572113f79a tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 881.811122] env[61923]: DEBUG nova.virt.hardware [None req-eb5e01f1-148c-4608-96e7-93572113f79a tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 881.811316] env[61923]: DEBUG nova.virt.hardware [None req-eb5e01f1-148c-4608-96e7-93572113f79a tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Got 1 possible topologies {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 881.811544] env[61923]: DEBUG nova.virt.hardware [None req-eb5e01f1-148c-4608-96e7-93572113f79a tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 881.811685] env[61923]: DEBUG nova.virt.hardware [None req-eb5e01f1-148c-4608-96e7-93572113f79a tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 881.813080] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bbc33925-0960-4db8-95b3-e563addc4f33 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.823020] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12f9b952-f3e8-4360-ab7d-36d0b9560dae {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.013870] env[61923]: DEBUG oslo_concurrency.lockutils [None req-a5e456d9-94ec-4a32-9f7d-b8da7dd3daa2 tempest-ServersTestManualDisk-2012652654 tempest-ServersTestManualDisk-2012652654-project-member] Acquiring lock "f1839f61-4314-48fe-8ab6-14b5e49d569d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 882.014199] env[61923]: DEBUG oslo_concurrency.lockutils [None req-a5e456d9-94ec-4a32-9f7d-b8da7dd3daa2 tempest-ServersTestManualDisk-2012652654 tempest-ServersTestManualDisk-2012652654-project-member] Lock "f1839f61-4314-48fe-8ab6-14b5e49d569d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 882.014426] env[61923]: DEBUG oslo_concurrency.lockutils [None req-a5e456d9-94ec-4a32-9f7d-b8da7dd3daa2 tempest-ServersTestManualDisk-2012652654 tempest-ServersTestManualDisk-2012652654-project-member] Acquiring lock "f1839f61-4314-48fe-8ab6-14b5e49d569d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 882.014616] env[61923]: DEBUG oslo_concurrency.lockutils [None req-a5e456d9-94ec-4a32-9f7d-b8da7dd3daa2 tempest-ServersTestManualDisk-2012652654 tempest-ServersTestManualDisk-2012652654-project-member] Lock "f1839f61-4314-48fe-8ab6-14b5e49d569d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 882.014787] env[61923]: DEBUG oslo_concurrency.lockutils [None req-a5e456d9-94ec-4a32-9f7d-b8da7dd3daa2 tempest-ServersTestManualDisk-2012652654 tempest-ServersTestManualDisk-2012652654-project-member] Lock "f1839f61-4314-48fe-8ab6-14b5e49d569d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 882.017210] env[61923]: INFO nova.compute.manager [None req-a5e456d9-94ec-4a32-9f7d-b8da7dd3daa2 tempest-ServersTestManualDisk-2012652654 tempest-ServersTestManualDisk-2012652654-project-member] [instance: f1839f61-4314-48fe-8ab6-14b5e49d569d] Terminating instance [ 882.019023] env[61923]: DEBUG nova.compute.manager [None req-a5e456d9-94ec-4a32-9f7d-b8da7dd3daa2 tempest-ServersTestManualDisk-2012652654 tempest-ServersTestManualDisk-2012652654-project-member] [instance: f1839f61-4314-48fe-8ab6-14b5e49d569d] Start destroying the instance on the hypervisor. {{(pid=61923) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 882.019295] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-a5e456d9-94ec-4a32-9f7d-b8da7dd3daa2 tempest-ServersTestManualDisk-2012652654 tempest-ServersTestManualDisk-2012652654-project-member] [instance: f1839f61-4314-48fe-8ab6-14b5e49d569d] Destroying instance {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 882.020310] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6219e911-23de-4cd4-9048-f003768cea5a {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.028586] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-a5e456d9-94ec-4a32-9f7d-b8da7dd3daa2 tempest-ServersTestManualDisk-2012652654 tempest-ServersTestManualDisk-2012652654-project-member] [instance: f1839f61-4314-48fe-8ab6-14b5e49d569d] Powering off the VM {{(pid=61923) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 882.029914] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-15c22ddb-c480-4b58-9f76-e456ef71e62d {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.036409] env[61923]: DEBUG oslo_vmware.api [None req-a5e456d9-94ec-4a32-9f7d-b8da7dd3daa2 tempest-ServersTestManualDisk-2012652654 tempest-ServersTestManualDisk-2012652654-project-member] Waiting for the task: (returnval){ [ 882.036409] env[61923]: value = "task-1377684" [ 882.036409] env[61923]: _type = "Task" [ 882.036409] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 882.045147] env[61923]: DEBUG oslo_vmware.api [None req-a5e456d9-94ec-4a32-9f7d-b8da7dd3daa2 tempest-ServersTestManualDisk-2012652654 tempest-ServersTestManualDisk-2012652654-project-member] Task: {'id': task-1377684, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.099554] env[61923]: DEBUG nova.compute.manager [req-137f3fe4-8f72-46c5-9bd6-30cebe085052 req-a5d9d8a4-3cb4-4a1c-ba94-698be723b3e3 service nova] [instance: 6012d3b5-739c-4762-9bb4-09c51171dcd7] Received event network-changed-269a41b0-75e7-46af-9497-8eae3102df11 {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 882.101112] env[61923]: DEBUG nova.compute.manager [req-137f3fe4-8f72-46c5-9bd6-30cebe085052 req-a5d9d8a4-3cb4-4a1c-ba94-698be723b3e3 service nova] [instance: 6012d3b5-739c-4762-9bb4-09c51171dcd7] Refreshing instance network info cache due to event network-changed-269a41b0-75e7-46af-9497-8eae3102df11. {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 882.101112] env[61923]: DEBUG oslo_concurrency.lockutils [req-137f3fe4-8f72-46c5-9bd6-30cebe085052 req-a5d9d8a4-3cb4-4a1c-ba94-698be723b3e3 service nova] Acquiring lock "refresh_cache-6012d3b5-739c-4762-9bb4-09c51171dcd7" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 882.101112] env[61923]: DEBUG oslo_concurrency.lockutils [req-137f3fe4-8f72-46c5-9bd6-30cebe085052 req-a5d9d8a4-3cb4-4a1c-ba94-698be723b3e3 service nova] Acquired lock "refresh_cache-6012d3b5-739c-4762-9bb4-09c51171dcd7" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 882.101112] env[61923]: DEBUG nova.network.neutron [req-137f3fe4-8f72-46c5-9bd6-30cebe085052 req-a5d9d8a4-3cb4-4a1c-ba94-698be723b3e3 service nova] [instance: 6012d3b5-739c-4762-9bb4-09c51171dcd7] Refreshing network info cache for port 269a41b0-75e7-46af-9497-8eae3102df11 {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 882.155083] env[61923]: DEBUG oslo_vmware.api [None req-6d4df172-3edf-4dbb-b96f-afa60a7b76e1 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Task: {'id': task-1377683, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.090267} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 882.155083] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-6d4df172-3edf-4dbb-b96f-afa60a7b76e1 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] [instance: a511392e-9ab6-42fb-b07c-f90e9786dcc1] Extended root virtual disk {{(pid=61923) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 882.156086] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef60f603-ae88-4389-a0da-9f6790d4d02d {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.181753] env[61923]: DEBUG nova.virt.vmwareapi.volumeops [None req-6d4df172-3edf-4dbb-b96f-afa60a7b76e1 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] [instance: a511392e-9ab6-42fb-b07c-f90e9786dcc1] Reconfiguring VM instance instance-0000004f to attach disk [datastore2] a511392e-9ab6-42fb-b07c-f90e9786dcc1/a511392e-9ab6-42fb-b07c-f90e9786dcc1.vmdk or device None with type sparse {{(pid=61923) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 882.186814] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c08a497f-b1f4-499b-ab6a-a744f14472a0 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.204090] env[61923]: DEBUG oslo_concurrency.lockutils [None req-7e66bb89-e0a5-4005-a075-6cc387f412c9 tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Lock "444affa5-a7ed-4a17-9015-9fd5724aab64" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 10.094s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 882.214189] env[61923]: DEBUG oslo_vmware.api [None req-6d4df172-3edf-4dbb-b96f-afa60a7b76e1 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Waiting for the task: (returnval){ [ 882.214189] env[61923]: value = "task-1377685" [ 882.214189] env[61923]: _type = "Task" [ 882.214189] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 882.229976] env[61923]: DEBUG oslo_vmware.api [None req-6d4df172-3edf-4dbb-b96f-afa60a7b76e1 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Task: {'id': task-1377685, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.262037] env[61923]: DEBUG nova.network.neutron [None req-eb5e01f1-148c-4608-96e7-93572113f79a tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: 4c2c973a-5d23-4b8a-8b99-6b634e337054] Successfully updated port: 9584dacf-1833-4a40-bfbe-56d7d5ea9998 {{(pid=61923) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 882.290956] env[61923]: DEBUG oslo_concurrency.lockutils [None req-55b58374-2c35-46b9-a508-d4833264ecac tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Releasing lock "refresh_cache-4cb3c397-ef1a-4267-a300-26da48206f3f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 882.291285] env[61923]: DEBUG nova.compute.manager [None req-55b58374-2c35-46b9-a508-d4833264ecac tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] [instance: 4cb3c397-ef1a-4267-a300-26da48206f3f] Instance network_info: |[{"id": "83e705c4-d432-4fdf-8486-d31b46e1e6a5", "address": "fa:16:3e:81:47:77", "network": {"id": "b8c2e837-0d64-4b07-9cc5-a66c8e8db1e9", "bridge": "br-int", "label": "tempest-ImagesTestJSON-133215580-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7abfe95660904cd4b8824257795d8e8a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f2c424c9-6446-4b2a-af8c-4d9c29117c39", "external-id": "nsx-vlan-transportzone-437", "segmentation_id": 437, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap83e705c4-d4", "ovs_interfaceid": "83e705c4-d432-4fdf-8486-d31b46e1e6a5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61923) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 882.291705] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-55b58374-2c35-46b9-a508-d4833264ecac tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] [instance: 4cb3c397-ef1a-4267-a300-26da48206f3f] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:81:47:77', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'f2c424c9-6446-4b2a-af8c-4d9c29117c39', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '83e705c4-d432-4fdf-8486-d31b46e1e6a5', 'vif_model': 'vmxnet3'}] {{(pid=61923) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 882.301355] env[61923]: DEBUG oslo.service.loopingcall [None req-55b58374-2c35-46b9-a508-d4833264ecac tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61923) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 882.305340] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4cb3c397-ef1a-4267-a300-26da48206f3f] Creating VM on the ESX host {{(pid=61923) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 882.311020] env[61923]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-686d1e99-0f8e-4999-98a9-08e722588c79 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.330583] env[61923]: DEBUG nova.compute.manager [req-208d0a7d-2f2d-46e5-8906-f7288e523826 req-deb0f534-1282-4140-b173-624dddd5a97d service nova] [instance: 4cb3c397-ef1a-4267-a300-26da48206f3f] Received event network-changed-83e705c4-d432-4fdf-8486-d31b46e1e6a5 {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 882.330829] env[61923]: DEBUG nova.compute.manager [req-208d0a7d-2f2d-46e5-8906-f7288e523826 req-deb0f534-1282-4140-b173-624dddd5a97d service nova] [instance: 4cb3c397-ef1a-4267-a300-26da48206f3f] Refreshing instance network info cache due to event network-changed-83e705c4-d432-4fdf-8486-d31b46e1e6a5. {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 882.331093] env[61923]: DEBUG oslo_concurrency.lockutils [req-208d0a7d-2f2d-46e5-8906-f7288e523826 req-deb0f534-1282-4140-b173-624dddd5a97d service nova] Acquiring lock "refresh_cache-4cb3c397-ef1a-4267-a300-26da48206f3f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 882.331283] env[61923]: DEBUG oslo_concurrency.lockutils [req-208d0a7d-2f2d-46e5-8906-f7288e523826 req-deb0f534-1282-4140-b173-624dddd5a97d service nova] Acquired lock "refresh_cache-4cb3c397-ef1a-4267-a300-26da48206f3f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 882.331513] env[61923]: DEBUG nova.network.neutron [req-208d0a7d-2f2d-46e5-8906-f7288e523826 req-deb0f534-1282-4140-b173-624dddd5a97d service nova] [instance: 4cb3c397-ef1a-4267-a300-26da48206f3f] Refreshing network info cache for port 83e705c4-d432-4fdf-8486-d31b46e1e6a5 {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 882.342105] env[61923]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 882.342105] env[61923]: value = "task-1377686" [ 882.342105] env[61923]: _type = "Task" [ 882.342105] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 882.352119] env[61923]: DEBUG oslo_vmware.api [-] Task: {'id': task-1377686, 'name': CreateVM_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.424967] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d867330-4c44-40f3-aa13-ae86bb0266ee {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.433740] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0954fe5e-5f1f-4941-8624-3b295226676a {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.466843] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0826450-6b2c-4195-a37a-71002e664363 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.475539] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0083c58c-74b8-41b7-8ac3-1c1795589d69 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.491472] env[61923]: DEBUG nova.compute.provider_tree [None req-89f7a2ee-0580-4457-b803-a0236f2ecef8 tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 882.547606] env[61923]: DEBUG oslo_vmware.api [None req-a5e456d9-94ec-4a32-9f7d-b8da7dd3daa2 tempest-ServersTestManualDisk-2012652654 tempest-ServersTestManualDisk-2012652654-project-member] Task: {'id': task-1377684, 'name': PowerOffVM_Task, 'duration_secs': 0.202808} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 882.547948] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-a5e456d9-94ec-4a32-9f7d-b8da7dd3daa2 tempest-ServersTestManualDisk-2012652654 tempest-ServersTestManualDisk-2012652654-project-member] [instance: f1839f61-4314-48fe-8ab6-14b5e49d569d] Powered off the VM {{(pid=61923) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 882.548161] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-a5e456d9-94ec-4a32-9f7d-b8da7dd3daa2 tempest-ServersTestManualDisk-2012652654 tempest-ServersTestManualDisk-2012652654-project-member] [instance: f1839f61-4314-48fe-8ab6-14b5e49d569d] Unregistering the VM {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 882.548427] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-728714bf-0a2b-4515-a72b-3bed35bc7980 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.624785] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-a5e456d9-94ec-4a32-9f7d-b8da7dd3daa2 tempest-ServersTestManualDisk-2012652654 tempest-ServersTestManualDisk-2012652654-project-member] [instance: f1839f61-4314-48fe-8ab6-14b5e49d569d] Unregistered the VM {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 882.625040] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-a5e456d9-94ec-4a32-9f7d-b8da7dd3daa2 tempest-ServersTestManualDisk-2012652654 tempest-ServersTestManualDisk-2012652654-project-member] [instance: f1839f61-4314-48fe-8ab6-14b5e49d569d] Deleting contents of the VM from datastore datastore1 {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 882.625232] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-a5e456d9-94ec-4a32-9f7d-b8da7dd3daa2 tempest-ServersTestManualDisk-2012652654 tempest-ServersTestManualDisk-2012652654-project-member] Deleting the datastore file [datastore1] f1839f61-4314-48fe-8ab6-14b5e49d569d {{(pid=61923) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 882.625512] env[61923]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e047047d-ba25-480c-99dd-3c38b3c05401 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.632563] env[61923]: DEBUG oslo_vmware.api [None req-a5e456d9-94ec-4a32-9f7d-b8da7dd3daa2 tempest-ServersTestManualDisk-2012652654 tempest-ServersTestManualDisk-2012652654-project-member] Waiting for the task: (returnval){ [ 882.632563] env[61923]: value = "task-1377688" [ 882.632563] env[61923]: _type = "Task" [ 882.632563] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 882.642700] env[61923]: DEBUG oslo_vmware.api [None req-a5e456d9-94ec-4a32-9f7d-b8da7dd3daa2 tempest-ServersTestManualDisk-2012652654 tempest-ServersTestManualDisk-2012652654-project-member] Task: {'id': task-1377688, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.715907] env[61923]: DEBUG nova.compute.manager [None req-4482c5e6-0d59-468c-9d9f-cee661c7eaed tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] [instance: 2719569b-8572-4199-8158-7bb367d17dc5] Checking state {{(pid=61923) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 882.716887] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9764d61-3681-4754-a351-9c01cf3de2fe {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.733691] env[61923]: DEBUG oslo_vmware.api [None req-6d4df172-3edf-4dbb-b96f-afa60a7b76e1 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Task: {'id': task-1377685, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.764147] env[61923]: DEBUG oslo_concurrency.lockutils [None req-eb5e01f1-148c-4608-96e7-93572113f79a tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Acquiring lock "refresh_cache-4c2c973a-5d23-4b8a-8b99-6b634e337054" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 882.764249] env[61923]: DEBUG oslo_concurrency.lockutils [None req-eb5e01f1-148c-4608-96e7-93572113f79a tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Acquired lock "refresh_cache-4c2c973a-5d23-4b8a-8b99-6b634e337054" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 882.764399] env[61923]: DEBUG nova.network.neutron [None req-eb5e01f1-148c-4608-96e7-93572113f79a tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: 4c2c973a-5d23-4b8a-8b99-6b634e337054] Building network info cache for instance {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 882.809735] env[61923]: DEBUG oslo_concurrency.lockutils [None req-02b5a1b8-a821-49de-bb63-1c89123e0e10 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Acquiring lock "interface-422d06d6-6932-46c3-bb25-841e6f65c028-b395b183-1c30-42de-929f-690ea10d9d75" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 882.810225] env[61923]: DEBUG oslo_concurrency.lockutils [None req-02b5a1b8-a821-49de-bb63-1c89123e0e10 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Lock "interface-422d06d6-6932-46c3-bb25-841e6f65c028-b395b183-1c30-42de-929f-690ea10d9d75" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.001s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 882.810847] env[61923]: DEBUG nova.objects.instance [None req-02b5a1b8-a821-49de-bb63-1c89123e0e10 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Lazy-loading 'flavor' on Instance uuid 422d06d6-6932-46c3-bb25-841e6f65c028 {{(pid=61923) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 882.843205] env[61923]: DEBUG nova.network.neutron [req-137f3fe4-8f72-46c5-9bd6-30cebe085052 req-a5d9d8a4-3cb4-4a1c-ba94-698be723b3e3 service nova] [instance: 6012d3b5-739c-4762-9bb4-09c51171dcd7] Updated VIF entry in instance network info cache for port 269a41b0-75e7-46af-9497-8eae3102df11. {{(pid=61923) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 882.843980] env[61923]: DEBUG nova.network.neutron [req-137f3fe4-8f72-46c5-9bd6-30cebe085052 req-a5d9d8a4-3cb4-4a1c-ba94-698be723b3e3 service nova] [instance: 6012d3b5-739c-4762-9bb4-09c51171dcd7] Updating instance_info_cache with network_info: [{"id": "269a41b0-75e7-46af-9497-8eae3102df11", "address": "fa:16:3e:e1:02:b8", "network": {"id": "a8b7e1b4-1ec6-4226-bdd6-535dbce450d2", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1928758125-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "912300a09b80452e85e3cd13c4a644cc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7806fe18-2b89-4386-87b1-f22876f82af2", "external-id": "nsx-vlan-transportzone-727", "segmentation_id": 727, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap269a41b0-75", "ovs_interfaceid": "269a41b0-75e7-46af-9497-8eae3102df11", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 882.856198] env[61923]: DEBUG oslo_vmware.api [-] Task: {'id': task-1377686, 'name': CreateVM_Task} progress is 99%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.995694] env[61923]: DEBUG nova.scheduler.client.report [None req-89f7a2ee-0580-4457-b803-a0236f2ecef8 tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 883.143159] env[61923]: DEBUG oslo_vmware.api [None req-a5e456d9-94ec-4a32-9f7d-b8da7dd3daa2 tempest-ServersTestManualDisk-2012652654 tempest-ServersTestManualDisk-2012652654-project-member] Task: {'id': task-1377688, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.165794} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 883.143480] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-a5e456d9-94ec-4a32-9f7d-b8da7dd3daa2 tempest-ServersTestManualDisk-2012652654 tempest-ServersTestManualDisk-2012652654-project-member] Deleted the datastore file {{(pid=61923) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 883.143669] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-a5e456d9-94ec-4a32-9f7d-b8da7dd3daa2 tempest-ServersTestManualDisk-2012652654 tempest-ServersTestManualDisk-2012652654-project-member] [instance: f1839f61-4314-48fe-8ab6-14b5e49d569d] Deleted contents of the VM from datastore datastore1 {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 883.143849] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-a5e456d9-94ec-4a32-9f7d-b8da7dd3daa2 tempest-ServersTestManualDisk-2012652654 tempest-ServersTestManualDisk-2012652654-project-member] [instance: f1839f61-4314-48fe-8ab6-14b5e49d569d] Instance destroyed {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 883.144038] env[61923]: INFO nova.compute.manager [None req-a5e456d9-94ec-4a32-9f7d-b8da7dd3daa2 tempest-ServersTestManualDisk-2012652654 tempest-ServersTestManualDisk-2012652654-project-member] [instance: f1839f61-4314-48fe-8ab6-14b5e49d569d] Took 1.12 seconds to destroy the instance on the hypervisor. [ 883.144293] env[61923]: DEBUG oslo.service.loopingcall [None req-a5e456d9-94ec-4a32-9f7d-b8da7dd3daa2 tempest-ServersTestManualDisk-2012652654 tempest-ServersTestManualDisk-2012652654-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61923) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 883.144844] env[61923]: DEBUG nova.compute.manager [-] [instance: f1839f61-4314-48fe-8ab6-14b5e49d569d] Deallocating network for instance {{(pid=61923) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 883.144953] env[61923]: DEBUG nova.network.neutron [-] [instance: f1839f61-4314-48fe-8ab6-14b5e49d569d] deallocate_for_instance() {{(pid=61923) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 883.226305] env[61923]: DEBUG oslo_vmware.api [None req-6d4df172-3edf-4dbb-b96f-afa60a7b76e1 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Task: {'id': task-1377685, 'name': ReconfigVM_Task, 'duration_secs': 0.769601} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 883.226547] env[61923]: DEBUG nova.virt.vmwareapi.volumeops [None req-6d4df172-3edf-4dbb-b96f-afa60a7b76e1 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] [instance: a511392e-9ab6-42fb-b07c-f90e9786dcc1] Reconfigured VM instance instance-0000004f to attach disk [datastore2] a511392e-9ab6-42fb-b07c-f90e9786dcc1/a511392e-9ab6-42fb-b07c-f90e9786dcc1.vmdk or device None with type sparse {{(pid=61923) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 883.227221] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-725bd8ff-69ac-498a-b2db-94c0ee662544 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.237014] env[61923]: INFO nova.compute.manager [None req-4482c5e6-0d59-468c-9d9f-cee661c7eaed tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] [instance: 2719569b-8572-4199-8158-7bb367d17dc5] instance snapshotting [ 883.237184] env[61923]: DEBUG oslo_vmware.api [None req-6d4df172-3edf-4dbb-b96f-afa60a7b76e1 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Waiting for the task: (returnval){ [ 883.237184] env[61923]: value = "task-1377689" [ 883.237184] env[61923]: _type = "Task" [ 883.237184] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 883.238235] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89ef43c0-fa01-444f-a299-761e6de2812f {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.264162] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1dd663f7-7127-4a8b-a90f-d0af487e04ff {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.271730] env[61923]: DEBUG oslo_vmware.api [None req-6d4df172-3edf-4dbb-b96f-afa60a7b76e1 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Task: {'id': task-1377689, 'name': Rename_Task} progress is 14%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.312397] env[61923]: DEBUG nova.network.neutron [req-208d0a7d-2f2d-46e5-8906-f7288e523826 req-deb0f534-1282-4140-b173-624dddd5a97d service nova] [instance: 4cb3c397-ef1a-4267-a300-26da48206f3f] Updated VIF entry in instance network info cache for port 83e705c4-d432-4fdf-8486-d31b46e1e6a5. {{(pid=61923) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 883.312748] env[61923]: DEBUG nova.network.neutron [req-208d0a7d-2f2d-46e5-8906-f7288e523826 req-deb0f534-1282-4140-b173-624dddd5a97d service nova] [instance: 4cb3c397-ef1a-4267-a300-26da48206f3f] Updating instance_info_cache with network_info: [{"id": "83e705c4-d432-4fdf-8486-d31b46e1e6a5", "address": "fa:16:3e:81:47:77", "network": {"id": "b8c2e837-0d64-4b07-9cc5-a66c8e8db1e9", "bridge": "br-int", "label": "tempest-ImagesTestJSON-133215580-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7abfe95660904cd4b8824257795d8e8a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f2c424c9-6446-4b2a-af8c-4d9c29117c39", "external-id": "nsx-vlan-transportzone-437", "segmentation_id": 437, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap83e705c4-d4", "ovs_interfaceid": "83e705c4-d432-4fdf-8486-d31b46e1e6a5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 883.321351] env[61923]: DEBUG nova.network.neutron [None req-eb5e01f1-148c-4608-96e7-93572113f79a tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: 4c2c973a-5d23-4b8a-8b99-6b634e337054] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 883.350586] env[61923]: DEBUG oslo_concurrency.lockutils [req-137f3fe4-8f72-46c5-9bd6-30cebe085052 req-a5d9d8a4-3cb4-4a1c-ba94-698be723b3e3 service nova] Releasing lock "refresh_cache-6012d3b5-739c-4762-9bb4-09c51171dcd7" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 883.357397] env[61923]: DEBUG oslo_vmware.api [-] Task: {'id': task-1377686, 'name': CreateVM_Task, 'duration_secs': 0.620307} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 883.357575] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4cb3c397-ef1a-4267-a300-26da48206f3f] Created VM on the ESX host {{(pid=61923) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 883.358239] env[61923]: DEBUG oslo_concurrency.lockutils [None req-55b58374-2c35-46b9-a508-d4833264ecac tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9ddad7e6-f1e4-49aa-b956-1b8d3273318d" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 883.358404] env[61923]: DEBUG oslo_concurrency.lockutils [None req-55b58374-2c35-46b9-a508-d4833264ecac tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9ddad7e6-f1e4-49aa-b956-1b8d3273318d" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 883.358772] env[61923]: DEBUG oslo_concurrency.lockutils [None req-55b58374-2c35-46b9-a508-d4833264ecac tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/9ddad7e6-f1e4-49aa-b956-1b8d3273318d" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 883.359352] env[61923]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5fc44210-d780-440d-aa42-ad37edbf0551 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.365304] env[61923]: DEBUG oslo_vmware.api [None req-55b58374-2c35-46b9-a508-d4833264ecac tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Waiting for the task: (returnval){ [ 883.365304] env[61923]: value = "session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52edfe84-fe97-063c-00cd-599a48839a0a" [ 883.365304] env[61923]: _type = "Task" [ 883.365304] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 883.376759] env[61923]: DEBUG oslo_vmware.api [None req-55b58374-2c35-46b9-a508-d4833264ecac tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52edfe84-fe97-063c-00cd-599a48839a0a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.501856] env[61923]: DEBUG oslo_concurrency.lockutils [None req-89f7a2ee-0580-4457-b803-a0236f2ecef8 tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.852s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 883.506654] env[61923]: DEBUG oslo_concurrency.lockutils [None req-59b20242-d3b6-4451-b636-b9e52b91d9f2 tempest-ServerAddressesTestJSON-1846886542 tempest-ServerAddressesTestJSON-1846886542-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 7.465s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 883.507014] env[61923]: DEBUG nova.objects.instance [None req-59b20242-d3b6-4451-b636-b9e52b91d9f2 tempest-ServerAddressesTestJSON-1846886542 tempest-ServerAddressesTestJSON-1846886542-project-member] Lazy-loading 'resources' on Instance uuid 5b73e07a-d6ef-4dcf-bdae-eea91d2aeb6f {{(pid=61923) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 883.533678] env[61923]: INFO nova.scheduler.client.report [None req-89f7a2ee-0580-4457-b803-a0236f2ecef8 tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Deleted allocations for instance e413c45d-cd89-44d4-9102-3d188907e7bb [ 883.557092] env[61923]: DEBUG nova.objects.instance [None req-02b5a1b8-a821-49de-bb63-1c89123e0e10 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Lazy-loading 'pci_requests' on Instance uuid 422d06d6-6932-46c3-bb25-841e6f65c028 {{(pid=61923) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 883.609126] env[61923]: DEBUG nova.network.neutron [None req-eb5e01f1-148c-4608-96e7-93572113f79a tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: 4c2c973a-5d23-4b8a-8b99-6b634e337054] Updating instance_info_cache with network_info: [{"id": "9584dacf-1833-4a40-bfbe-56d7d5ea9998", "address": "fa:16:3e:48:9c:20", "network": {"id": "d1c38272-0461-43e5-93ae-3964346f77d0", "bridge": "br-int", "label": "tempest-ServersTestJSON-151839165-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "69cc941eb6dd4780ac12aa29656c37f7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8ba3bd22-c936-470e-89bd-b3a5587e87a0", "external-id": "nsx-vlan-transportzone-605", "segmentation_id": 605, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9584dacf-18", "ovs_interfaceid": "9584dacf-1833-4a40-bfbe-56d7d5ea9998", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 883.747063] env[61923]: DEBUG oslo_vmware.api [None req-6d4df172-3edf-4dbb-b96f-afa60a7b76e1 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Task: {'id': task-1377689, 'name': Rename_Task} progress is 99%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.781187] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-4482c5e6-0d59-468c-9d9f-cee661c7eaed tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] [instance: 2719569b-8572-4199-8158-7bb367d17dc5] Creating Snapshot of the VM instance {{(pid=61923) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 883.781478] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-ecbe66f0-8b51-4f16-accd-974e256e47de {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.790808] env[61923]: DEBUG oslo_vmware.api [None req-4482c5e6-0d59-468c-9d9f-cee661c7eaed tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Waiting for the task: (returnval){ [ 883.790808] env[61923]: value = "task-1377690" [ 883.790808] env[61923]: _type = "Task" [ 883.790808] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 883.801480] env[61923]: DEBUG oslo_vmware.api [None req-4482c5e6-0d59-468c-9d9f-cee661c7eaed tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Task: {'id': task-1377690, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.819614] env[61923]: DEBUG oslo_concurrency.lockutils [req-208d0a7d-2f2d-46e5-8906-f7288e523826 req-deb0f534-1282-4140-b173-624dddd5a97d service nova] Releasing lock "refresh_cache-4cb3c397-ef1a-4267-a300-26da48206f3f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 883.819904] env[61923]: DEBUG nova.compute.manager [req-208d0a7d-2f2d-46e5-8906-f7288e523826 req-deb0f534-1282-4140-b173-624dddd5a97d service nova] [instance: 422d06d6-6932-46c3-bb25-841e6f65c028] Received event network-changed-97caa1b8-427c-46b0-be7a-d59b932900c3 {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 883.820119] env[61923]: DEBUG nova.compute.manager [req-208d0a7d-2f2d-46e5-8906-f7288e523826 req-deb0f534-1282-4140-b173-624dddd5a97d service nova] [instance: 422d06d6-6932-46c3-bb25-841e6f65c028] Refreshing instance network info cache due to event network-changed-97caa1b8-427c-46b0-be7a-d59b932900c3. {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 883.820340] env[61923]: DEBUG oslo_concurrency.lockutils [req-208d0a7d-2f2d-46e5-8906-f7288e523826 req-deb0f534-1282-4140-b173-624dddd5a97d service nova] Acquiring lock "refresh_cache-422d06d6-6932-46c3-bb25-841e6f65c028" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 883.820483] env[61923]: DEBUG oslo_concurrency.lockutils [req-208d0a7d-2f2d-46e5-8906-f7288e523826 req-deb0f534-1282-4140-b173-624dddd5a97d service nova] Acquired lock "refresh_cache-422d06d6-6932-46c3-bb25-841e6f65c028" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 883.820644] env[61923]: DEBUG nova.network.neutron [req-208d0a7d-2f2d-46e5-8906-f7288e523826 req-deb0f534-1282-4140-b173-624dddd5a97d service nova] [instance: 422d06d6-6932-46c3-bb25-841e6f65c028] Refreshing network info cache for port 97caa1b8-427c-46b0-be7a-d59b932900c3 {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 883.876316] env[61923]: DEBUG oslo_concurrency.lockutils [None req-55b58374-2c35-46b9-a508-d4833264ecac tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9ddad7e6-f1e4-49aa-b956-1b8d3273318d" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 883.876596] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-55b58374-2c35-46b9-a508-d4833264ecac tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] [instance: 4cb3c397-ef1a-4267-a300-26da48206f3f] Processing image 9ddad7e6-f1e4-49aa-b956-1b8d3273318d {{(pid=61923) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 883.876836] env[61923]: DEBUG oslo_concurrency.lockutils [None req-55b58374-2c35-46b9-a508-d4833264ecac tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/9ddad7e6-f1e4-49aa-b956-1b8d3273318d/9ddad7e6-f1e4-49aa-b956-1b8d3273318d.vmdk" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 883.876997] env[61923]: DEBUG oslo_concurrency.lockutils [None req-55b58374-2c35-46b9-a508-d4833264ecac tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Acquired lock "[datastore2] devstack-image-cache_base/9ddad7e6-f1e4-49aa-b956-1b8d3273318d/9ddad7e6-f1e4-49aa-b956-1b8d3273318d.vmdk" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 883.877259] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-55b58374-2c35-46b9-a508-d4833264ecac tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61923) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 883.877545] env[61923]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-dee5588a-4c64-4902-97ad-591965a5da40 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.893450] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-55b58374-2c35-46b9-a508-d4833264ecac tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61923) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 883.893936] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-55b58374-2c35-46b9-a508-d4833264ecac tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61923) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 883.894472] env[61923]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7c092f85-7d2d-42dd-bad8-bc74b41e13b6 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.901213] env[61923]: DEBUG oslo_vmware.api [None req-55b58374-2c35-46b9-a508-d4833264ecac tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Waiting for the task: (returnval){ [ 883.901213] env[61923]: value = "session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]5232b51d-6722-0b02-15d4-1de3952a53ae" [ 883.901213] env[61923]: _type = "Task" [ 883.901213] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 883.909492] env[61923]: DEBUG oslo_vmware.api [None req-55b58374-2c35-46b9-a508-d4833264ecac tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]5232b51d-6722-0b02-15d4-1de3952a53ae, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.046682] env[61923]: DEBUG oslo_concurrency.lockutils [None req-89f7a2ee-0580-4457-b803-a0236f2ecef8 tempest-MultipleCreateTestJSON-1907070631 tempest-MultipleCreateTestJSON-1907070631-project-member] Lock "e413c45d-cd89-44d4-9102-3d188907e7bb" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 11.825s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 884.059662] env[61923]: DEBUG nova.objects.base [None req-02b5a1b8-a821-49de-bb63-1c89123e0e10 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Object Instance<422d06d6-6932-46c3-bb25-841e6f65c028> lazy-loaded attributes: flavor,pci_requests {{(pid=61923) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 884.059898] env[61923]: DEBUG nova.network.neutron [None req-02b5a1b8-a821-49de-bb63-1c89123e0e10 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] [instance: 422d06d6-6932-46c3-bb25-841e6f65c028] allocate_for_instance() {{(pid=61923) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 884.112251] env[61923]: DEBUG oslo_concurrency.lockutils [None req-eb5e01f1-148c-4608-96e7-93572113f79a tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Releasing lock "refresh_cache-4c2c973a-5d23-4b8a-8b99-6b634e337054" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 884.112573] env[61923]: DEBUG nova.compute.manager [None req-eb5e01f1-148c-4608-96e7-93572113f79a tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: 4c2c973a-5d23-4b8a-8b99-6b634e337054] Instance network_info: |[{"id": "9584dacf-1833-4a40-bfbe-56d7d5ea9998", "address": "fa:16:3e:48:9c:20", "network": {"id": "d1c38272-0461-43e5-93ae-3964346f77d0", "bridge": "br-int", "label": "tempest-ServersTestJSON-151839165-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "69cc941eb6dd4780ac12aa29656c37f7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8ba3bd22-c936-470e-89bd-b3a5587e87a0", "external-id": "nsx-vlan-transportzone-605", "segmentation_id": 605, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9584dacf-18", "ovs_interfaceid": "9584dacf-1833-4a40-bfbe-56d7d5ea9998", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61923) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 884.112966] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-eb5e01f1-148c-4608-96e7-93572113f79a tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: 4c2c973a-5d23-4b8a-8b99-6b634e337054] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:48:9c:20', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '8ba3bd22-c936-470e-89bd-b3a5587e87a0', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '9584dacf-1833-4a40-bfbe-56d7d5ea9998', 'vif_model': 'vmxnet3'}] {{(pid=61923) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 884.121111] env[61923]: DEBUG oslo.service.loopingcall [None req-eb5e01f1-148c-4608-96e7-93572113f79a tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61923) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 884.123641] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4c2c973a-5d23-4b8a-8b99-6b634e337054] Creating VM on the ESX host {{(pid=61923) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 884.124081] env[61923]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-3c737819-4932-4675-86e3-7575aec679f1 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.143266] env[61923]: DEBUG nova.policy [None req-02b5a1b8-a821-49de-bb63-1c89123e0e10 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'cd5039bf8f4344e490df828e643d2779', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '912300a09b80452e85e3cd13c4a644cc', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61923) authorize /opt/stack/nova/nova/policy.py:201}} [ 884.146615] env[61923]: DEBUG nova.compute.manager [req-79aa1fc8-f53c-4c0d-a13b-9bcb67526c91 req-a2ab7a4f-f911-41a2-911d-b8c40e2401d3 service nova] [instance: 422d06d6-6932-46c3-bb25-841e6f65c028] Received event network-changed-97caa1b8-427c-46b0-be7a-d59b932900c3 {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 884.146795] env[61923]: DEBUG nova.compute.manager [req-79aa1fc8-f53c-4c0d-a13b-9bcb67526c91 req-a2ab7a4f-f911-41a2-911d-b8c40e2401d3 service nova] [instance: 422d06d6-6932-46c3-bb25-841e6f65c028] Refreshing instance network info cache due to event network-changed-97caa1b8-427c-46b0-be7a-d59b932900c3. {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 884.146983] env[61923]: DEBUG oslo_concurrency.lockutils [req-79aa1fc8-f53c-4c0d-a13b-9bcb67526c91 req-a2ab7a4f-f911-41a2-911d-b8c40e2401d3 service nova] Acquiring lock "refresh_cache-422d06d6-6932-46c3-bb25-841e6f65c028" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 884.155076] env[61923]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 884.155076] env[61923]: value = "task-1377691" [ 884.155076] env[61923]: _type = "Task" [ 884.155076] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 884.168993] env[61923]: DEBUG oslo_vmware.api [-] Task: {'id': task-1377691, 'name': CreateVM_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.174621] env[61923]: DEBUG nova.network.neutron [-] [instance: f1839f61-4314-48fe-8ab6-14b5e49d569d] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 884.207024] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-471035bb-8202-4597-aa1a-bbbf19a418de {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.214786] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f95cb44b-ffbe-4bcb-9abe-56cbe2821fbb {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.252628] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-871cbdaf-0e10-4e8a-89ad-a38484837ba6 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.261714] env[61923]: DEBUG oslo_vmware.api [None req-6d4df172-3edf-4dbb-b96f-afa60a7b76e1 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Task: {'id': task-1377689, 'name': Rename_Task} progress is 99%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.264958] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dcfaeaab-0095-476c-ac1c-06bc8b8119b5 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.280953] env[61923]: DEBUG nova.compute.provider_tree [None req-59b20242-d3b6-4451-b636-b9e52b91d9f2 tempest-ServerAddressesTestJSON-1846886542 tempest-ServerAddressesTestJSON-1846886542-project-member] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 884.301259] env[61923]: DEBUG oslo_vmware.api [None req-4482c5e6-0d59-468c-9d9f-cee661c7eaed tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Task: {'id': task-1377690, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.414348] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-55b58374-2c35-46b9-a508-d4833264ecac tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] [instance: 4cb3c397-ef1a-4267-a300-26da48206f3f] Preparing fetch location {{(pid=61923) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 884.415671] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-55b58374-2c35-46b9-a508-d4833264ecac tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] [instance: 4cb3c397-ef1a-4267-a300-26da48206f3f] Fetch image to [datastore2] OSTACK_IMG_345122e9-0048-4f77-a160-7345d2ae6537/OSTACK_IMG_345122e9-0048-4f77-a160-7345d2ae6537.vmdk {{(pid=61923) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 884.415671] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-55b58374-2c35-46b9-a508-d4833264ecac tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] [instance: 4cb3c397-ef1a-4267-a300-26da48206f3f] Downloading stream optimized image 9ddad7e6-f1e4-49aa-b956-1b8d3273318d to [datastore2] OSTACK_IMG_345122e9-0048-4f77-a160-7345d2ae6537/OSTACK_IMG_345122e9-0048-4f77-a160-7345d2ae6537.vmdk on the data store datastore2 as vApp {{(pid=61923) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 884.415671] env[61923]: DEBUG nova.virt.vmwareapi.images [None req-55b58374-2c35-46b9-a508-d4833264ecac tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] [instance: 4cb3c397-ef1a-4267-a300-26da48206f3f] Downloading image file data 9ddad7e6-f1e4-49aa-b956-1b8d3273318d to the ESX as VM named 'OSTACK_IMG_345122e9-0048-4f77-a160-7345d2ae6537' {{(pid=61923) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 884.510029] env[61923]: DEBUG oslo_vmware.rw_handles [None req-55b58374-2c35-46b9-a508-d4833264ecac tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 884.510029] env[61923]: value = "resgroup-9" [ 884.510029] env[61923]: _type = "ResourcePool" [ 884.510029] env[61923]: }. {{(pid=61923) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 884.512130] env[61923]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-140965c7-d648-44ee-ad54-a1928aa02aed {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.538145] env[61923]: DEBUG oslo_vmware.rw_handles [None req-55b58374-2c35-46b9-a508-d4833264ecac tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Lease: (returnval){ [ 884.538145] env[61923]: value = "session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]5261d83c-9297-e08f-05d4-23664224eca3" [ 884.538145] env[61923]: _type = "HttpNfcLease" [ 884.538145] env[61923]: } obtained for vApp import into resource pool (val){ [ 884.538145] env[61923]: value = "resgroup-9" [ 884.538145] env[61923]: _type = "ResourcePool" [ 884.538145] env[61923]: }. {{(pid=61923) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 884.538495] env[61923]: DEBUG oslo_vmware.api [None req-55b58374-2c35-46b9-a508-d4833264ecac tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Waiting for the lease: (returnval){ [ 884.538495] env[61923]: value = "session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]5261d83c-9297-e08f-05d4-23664224eca3" [ 884.538495] env[61923]: _type = "HttpNfcLease" [ 884.538495] env[61923]: } to be ready. {{(pid=61923) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 884.545861] env[61923]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 884.545861] env[61923]: value = "session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]5261d83c-9297-e08f-05d4-23664224eca3" [ 884.545861] env[61923]: _type = "HttpNfcLease" [ 884.545861] env[61923]: } is initializing. {{(pid=61923) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 884.665630] env[61923]: DEBUG oslo_vmware.api [-] Task: {'id': task-1377691, 'name': CreateVM_Task} progress is 25%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.677765] env[61923]: INFO nova.compute.manager [-] [instance: f1839f61-4314-48fe-8ab6-14b5e49d569d] Took 1.53 seconds to deallocate network for instance. [ 884.759621] env[61923]: DEBUG oslo_vmware.api [None req-6d4df172-3edf-4dbb-b96f-afa60a7b76e1 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Task: {'id': task-1377689, 'name': Rename_Task, 'duration_secs': 1.302151} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 884.759956] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-6d4df172-3edf-4dbb-b96f-afa60a7b76e1 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] [instance: a511392e-9ab6-42fb-b07c-f90e9786dcc1] Powering on the VM {{(pid=61923) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 884.760371] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d2bf8ffb-3d24-4298-9e61-970e57804f6b {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.769997] env[61923]: DEBUG oslo_vmware.api [None req-6d4df172-3edf-4dbb-b96f-afa60a7b76e1 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Waiting for the task: (returnval){ [ 884.769997] env[61923]: value = "task-1377693" [ 884.769997] env[61923]: _type = "Task" [ 884.769997] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 884.781066] env[61923]: DEBUG oslo_vmware.api [None req-6d4df172-3edf-4dbb-b96f-afa60a7b76e1 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Task: {'id': task-1377693, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.783186] env[61923]: DEBUG nova.scheduler.client.report [None req-59b20242-d3b6-4451-b636-b9e52b91d9f2 tempest-ServerAddressesTestJSON-1846886542 tempest-ServerAddressesTestJSON-1846886542-project-member] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 884.802757] env[61923]: DEBUG oslo_vmware.api [None req-4482c5e6-0d59-468c-9d9f-cee661c7eaed tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Task: {'id': task-1377690, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.817032] env[61923]: DEBUG nova.network.neutron [req-208d0a7d-2f2d-46e5-8906-f7288e523826 req-deb0f534-1282-4140-b173-624dddd5a97d service nova] [instance: 422d06d6-6932-46c3-bb25-841e6f65c028] Updated VIF entry in instance network info cache for port 97caa1b8-427c-46b0-be7a-d59b932900c3. {{(pid=61923) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 884.817497] env[61923]: DEBUG nova.network.neutron [req-208d0a7d-2f2d-46e5-8906-f7288e523826 req-deb0f534-1282-4140-b173-624dddd5a97d service nova] [instance: 422d06d6-6932-46c3-bb25-841e6f65c028] Updating instance_info_cache with network_info: [{"id": "97caa1b8-427c-46b0-be7a-d59b932900c3", "address": "fa:16:3e:fa:f8:d1", "network": {"id": "a8b7e1b4-1ec6-4226-bdd6-535dbce450d2", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1928758125-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.146", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "912300a09b80452e85e3cd13c4a644cc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7806fe18-2b89-4386-87b1-f22876f82af2", "external-id": "nsx-vlan-transportzone-727", "segmentation_id": 727, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap97caa1b8-42", "ovs_interfaceid": "97caa1b8-427c-46b0-be7a-d59b932900c3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 885.047388] env[61923]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 885.047388] env[61923]: value = "session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]5261d83c-9297-e08f-05d4-23664224eca3" [ 885.047388] env[61923]: _type = "HttpNfcLease" [ 885.047388] env[61923]: } is initializing. {{(pid=61923) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 885.169234] env[61923]: DEBUG oslo_vmware.api [-] Task: {'id': task-1377691, 'name': CreateVM_Task, 'duration_secs': 0.680914} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 885.169336] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4c2c973a-5d23-4b8a-8b99-6b634e337054] Created VM on the ESX host {{(pid=61923) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 885.170112] env[61923]: DEBUG oslo_concurrency.lockutils [None req-eb5e01f1-148c-4608-96e7-93572113f79a tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 885.170213] env[61923]: DEBUG oslo_concurrency.lockutils [None req-eb5e01f1-148c-4608-96e7-93572113f79a tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 885.170508] env[61923]: DEBUG oslo_concurrency.lockutils [None req-eb5e01f1-148c-4608-96e7-93572113f79a tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 885.170735] env[61923]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7581fa14-5b45-43b5-a4c2-4982d11dfa9c {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.176139] env[61923]: DEBUG oslo_vmware.api [None req-eb5e01f1-148c-4608-96e7-93572113f79a tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Waiting for the task: (returnval){ [ 885.176139] env[61923]: value = "session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52059526-e54f-9513-8fe8-60a368c7a202" [ 885.176139] env[61923]: _type = "Task" [ 885.176139] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 885.184698] env[61923]: DEBUG oslo_concurrency.lockutils [None req-a5e456d9-94ec-4a32-9f7d-b8da7dd3daa2 tempest-ServersTestManualDisk-2012652654 tempest-ServersTestManualDisk-2012652654-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 885.184964] env[61923]: DEBUG oslo_vmware.api [None req-eb5e01f1-148c-4608-96e7-93572113f79a tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52059526-e54f-9513-8fe8-60a368c7a202, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.280388] env[61923]: DEBUG oslo_vmware.api [None req-6d4df172-3edf-4dbb-b96f-afa60a7b76e1 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Task: {'id': task-1377693, 'name': PowerOnVM_Task, 'duration_secs': 0.496494} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 885.280693] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-6d4df172-3edf-4dbb-b96f-afa60a7b76e1 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] [instance: a511392e-9ab6-42fb-b07c-f90e9786dcc1] Powered on the VM {{(pid=61923) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 885.280901] env[61923]: INFO nova.compute.manager [None req-6d4df172-3edf-4dbb-b96f-afa60a7b76e1 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] [instance: a511392e-9ab6-42fb-b07c-f90e9786dcc1] Took 9.22 seconds to spawn the instance on the hypervisor. [ 885.281095] env[61923]: DEBUG nova.compute.manager [None req-6d4df172-3edf-4dbb-b96f-afa60a7b76e1 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] [instance: a511392e-9ab6-42fb-b07c-f90e9786dcc1] Checking state {{(pid=61923) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 885.281877] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0468d5c-9c98-4a71-ba74-233b9d93b69c {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.292310] env[61923]: DEBUG oslo_concurrency.lockutils [None req-59b20242-d3b6-4451-b636-b9e52b91d9f2 tempest-ServerAddressesTestJSON-1846886542 tempest-ServerAddressesTestJSON-1846886542-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.786s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 885.295186] env[61923]: DEBUG oslo_concurrency.lockutils [None req-3192dd2d-7513-4a93-8a8c-c00e6e67889c tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 8.990s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 885.296592] env[61923]: INFO nova.compute.claims [None req-3192dd2d-7513-4a93-8a8c-c00e6e67889c tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] [instance: 263f0866-49af-4ab6-8132-d993642ebedb] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 885.308700] env[61923]: DEBUG oslo_vmware.api [None req-4482c5e6-0d59-468c-9d9f-cee661c7eaed tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Task: {'id': task-1377690, 'name': CreateSnapshot_Task, 'duration_secs': 1.311772} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 885.309676] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-4482c5e6-0d59-468c-9d9f-cee661c7eaed tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] [instance: 2719569b-8572-4199-8158-7bb367d17dc5] Created Snapshot of the VM instance {{(pid=61923) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 885.311153] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa4d3135-3580-44c7-b3c4-a737a9be861f {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.321932] env[61923]: DEBUG oslo_concurrency.lockutils [req-208d0a7d-2f2d-46e5-8906-f7288e523826 req-deb0f534-1282-4140-b173-624dddd5a97d service nova] Releasing lock "refresh_cache-422d06d6-6932-46c3-bb25-841e6f65c028" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 885.322178] env[61923]: DEBUG nova.compute.manager [req-208d0a7d-2f2d-46e5-8906-f7288e523826 req-deb0f534-1282-4140-b173-624dddd5a97d service nova] [instance: 6012d3b5-739c-4762-9bb4-09c51171dcd7] Received event network-changed-269a41b0-75e7-46af-9497-8eae3102df11 {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 885.322360] env[61923]: DEBUG nova.compute.manager [req-208d0a7d-2f2d-46e5-8906-f7288e523826 req-deb0f534-1282-4140-b173-624dddd5a97d service nova] [instance: 6012d3b5-739c-4762-9bb4-09c51171dcd7] Refreshing instance network info cache due to event network-changed-269a41b0-75e7-46af-9497-8eae3102df11. {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 885.322604] env[61923]: DEBUG oslo_concurrency.lockutils [req-208d0a7d-2f2d-46e5-8906-f7288e523826 req-deb0f534-1282-4140-b173-624dddd5a97d service nova] Acquiring lock "refresh_cache-6012d3b5-739c-4762-9bb4-09c51171dcd7" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 885.322792] env[61923]: DEBUG oslo_concurrency.lockutils [req-208d0a7d-2f2d-46e5-8906-f7288e523826 req-deb0f534-1282-4140-b173-624dddd5a97d service nova] Acquired lock "refresh_cache-6012d3b5-739c-4762-9bb4-09c51171dcd7" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 885.322977] env[61923]: DEBUG nova.network.neutron [req-208d0a7d-2f2d-46e5-8906-f7288e523826 req-deb0f534-1282-4140-b173-624dddd5a97d service nova] [instance: 6012d3b5-739c-4762-9bb4-09c51171dcd7] Refreshing network info cache for port 269a41b0-75e7-46af-9497-8eae3102df11 {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 885.328217] env[61923]: DEBUG oslo_concurrency.lockutils [req-79aa1fc8-f53c-4c0d-a13b-9bcb67526c91 req-a2ab7a4f-f911-41a2-911d-b8c40e2401d3 service nova] Acquired lock "refresh_cache-422d06d6-6932-46c3-bb25-841e6f65c028" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 885.328473] env[61923]: DEBUG nova.network.neutron [req-79aa1fc8-f53c-4c0d-a13b-9bcb67526c91 req-a2ab7a4f-f911-41a2-911d-b8c40e2401d3 service nova] [instance: 422d06d6-6932-46c3-bb25-841e6f65c028] Refreshing network info cache for port 97caa1b8-427c-46b0-be7a-d59b932900c3 {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 885.331073] env[61923]: INFO nova.scheduler.client.report [None req-59b20242-d3b6-4451-b636-b9e52b91d9f2 tempest-ServerAddressesTestJSON-1846886542 tempest-ServerAddressesTestJSON-1846886542-project-member] Deleted allocations for instance 5b73e07a-d6ef-4dcf-bdae-eea91d2aeb6f [ 885.548493] env[61923]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 885.548493] env[61923]: value = "session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]5261d83c-9297-e08f-05d4-23664224eca3" [ 885.548493] env[61923]: _type = "HttpNfcLease" [ 885.548493] env[61923]: } is ready. {{(pid=61923) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 885.549385] env[61923]: DEBUG oslo_vmware.rw_handles [None req-55b58374-2c35-46b9-a508-d4833264ecac tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 885.549385] env[61923]: value = "session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]5261d83c-9297-e08f-05d4-23664224eca3" [ 885.549385] env[61923]: _type = "HttpNfcLease" [ 885.549385] env[61923]: }. {{(pid=61923) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 885.550142] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8871210-56a8-4797-b950-d7790a12181c {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.558577] env[61923]: DEBUG oslo_vmware.rw_handles [None req-55b58374-2c35-46b9-a508-d4833264ecac tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52425bb0-dffe-722e-9ac5-23834c47972f/disk-0.vmdk from lease info. {{(pid=61923) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 885.558841] env[61923]: DEBUG oslo_vmware.rw_handles [None req-55b58374-2c35-46b9-a508-d4833264ecac tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Creating HTTP connection to write to file with size = 21334016 and URL = https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52425bb0-dffe-722e-9ac5-23834c47972f/disk-0.vmdk. {{(pid=61923) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 885.633649] env[61923]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-84ba305f-8160-4839-a2e9-f29c24f80093 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.687753] env[61923]: DEBUG oslo_vmware.api [None req-eb5e01f1-148c-4608-96e7-93572113f79a tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52059526-e54f-9513-8fe8-60a368c7a202, 'name': SearchDatastore_Task, 'duration_secs': 0.020814} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 885.688205] env[61923]: DEBUG oslo_concurrency.lockutils [None req-eb5e01f1-148c-4608-96e7-93572113f79a tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 885.688525] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-eb5e01f1-148c-4608-96e7-93572113f79a tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: 4c2c973a-5d23-4b8a-8b99-6b634e337054] Processing image 0f153f63-ae0a-45b1-b7f5-7f9b673c947f {{(pid=61923) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 885.688860] env[61923]: DEBUG oslo_concurrency.lockutils [None req-eb5e01f1-148c-4608-96e7-93572113f79a tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 885.689163] env[61923]: DEBUG oslo_concurrency.lockutils [None req-eb5e01f1-148c-4608-96e7-93572113f79a tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 885.689432] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-eb5e01f1-148c-4608-96e7-93572113f79a tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61923) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 885.689887] env[61923]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-9f856777-0397-4593-b174-19a7cf04c4cb {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.700529] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-eb5e01f1-148c-4608-96e7-93572113f79a tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61923) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 885.700754] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-eb5e01f1-148c-4608-96e7-93572113f79a tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61923) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 885.701549] env[61923]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5557e2f6-332a-47a9-a857-707343d977a6 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.707459] env[61923]: DEBUG oslo_vmware.api [None req-eb5e01f1-148c-4608-96e7-93572113f79a tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Waiting for the task: (returnval){ [ 885.707459] env[61923]: value = "session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52d3e875-34e5-a1dc-3b63-6c04fe08298c" [ 885.707459] env[61923]: _type = "Task" [ 885.707459] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 885.716336] env[61923]: DEBUG oslo_vmware.api [None req-eb5e01f1-148c-4608-96e7-93572113f79a tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52d3e875-34e5-a1dc-3b63-6c04fe08298c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.804396] env[61923]: INFO nova.compute.manager [None req-6d4df172-3edf-4dbb-b96f-afa60a7b76e1 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] [instance: a511392e-9ab6-42fb-b07c-f90e9786dcc1] Took 14.17 seconds to build instance. [ 885.840445] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-4482c5e6-0d59-468c-9d9f-cee661c7eaed tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] [instance: 2719569b-8572-4199-8158-7bb367d17dc5] Creating linked-clone VM from snapshot {{(pid=61923) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 885.846821] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-0170ad53-9634-416f-94b0-3e42d94076c9 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.853587] env[61923]: DEBUG oslo_concurrency.lockutils [None req-59b20242-d3b6-4451-b636-b9e52b91d9f2 tempest-ServerAddressesTestJSON-1846886542 tempest-ServerAddressesTestJSON-1846886542-project-member] Lock "5b73e07a-d6ef-4dcf-bdae-eea91d2aeb6f" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 13.396s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 885.867285] env[61923]: DEBUG oslo_vmware.api [None req-4482c5e6-0d59-468c-9d9f-cee661c7eaed tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Waiting for the task: (returnval){ [ 885.867285] env[61923]: value = "task-1377694" [ 885.867285] env[61923]: _type = "Task" [ 885.867285] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 885.876022] env[61923]: DEBUG nova.network.neutron [None req-02b5a1b8-a821-49de-bb63-1c89123e0e10 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] [instance: 422d06d6-6932-46c3-bb25-841e6f65c028] Successfully updated port: b395b183-1c30-42de-929f-690ea10d9d75 {{(pid=61923) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 885.882396] env[61923]: DEBUG oslo_vmware.api [None req-4482c5e6-0d59-468c-9d9f-cee661c7eaed tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Task: {'id': task-1377694, 'name': CloneVM_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.222252] env[61923]: DEBUG oslo_vmware.api [None req-eb5e01f1-148c-4608-96e7-93572113f79a tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52d3e875-34e5-a1dc-3b63-6c04fe08298c, 'name': SearchDatastore_Task, 'duration_secs': 0.009876} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 886.223131] env[61923]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7b662ea6-b1de-496f-9c9d-2cfa4aec7813 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.229784] env[61923]: DEBUG oslo_vmware.api [None req-eb5e01f1-148c-4608-96e7-93572113f79a tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Waiting for the task: (returnval){ [ 886.229784] env[61923]: value = "session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52427120-b448-37f5-7358-776548a06e34" [ 886.229784] env[61923]: _type = "Task" [ 886.229784] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 886.238797] env[61923]: DEBUG oslo_vmware.api [None req-eb5e01f1-148c-4608-96e7-93572113f79a tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52427120-b448-37f5-7358-776548a06e34, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.309839] env[61923]: DEBUG oslo_concurrency.lockutils [None req-6d4df172-3edf-4dbb-b96f-afa60a7b76e1 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Lock "a511392e-9ab6-42fb-b07c-f90e9786dcc1" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.690s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 886.378349] env[61923]: DEBUG oslo_vmware.api [None req-4482c5e6-0d59-468c-9d9f-cee661c7eaed tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Task: {'id': task-1377694, 'name': CloneVM_Task} progress is 94%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.382798] env[61923]: DEBUG oslo_concurrency.lockutils [None req-02b5a1b8-a821-49de-bb63-1c89123e0e10 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Acquiring lock "refresh_cache-422d06d6-6932-46c3-bb25-841e6f65c028" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 886.529567] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97949008-321e-4455-ba39-a85273543724 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.541212] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5009a7b5-e044-4b9c-b57a-a93ada432d8c {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.552293] env[61923]: DEBUG nova.network.neutron [req-208d0a7d-2f2d-46e5-8906-f7288e523826 req-deb0f534-1282-4140-b173-624dddd5a97d service nova] [instance: 6012d3b5-739c-4762-9bb4-09c51171dcd7] Updated VIF entry in instance network info cache for port 269a41b0-75e7-46af-9497-8eae3102df11. {{(pid=61923) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 886.552810] env[61923]: DEBUG nova.network.neutron [req-208d0a7d-2f2d-46e5-8906-f7288e523826 req-deb0f534-1282-4140-b173-624dddd5a97d service nova] [instance: 6012d3b5-739c-4762-9bb4-09c51171dcd7] Updating instance_info_cache with network_info: [{"id": "269a41b0-75e7-46af-9497-8eae3102df11", "address": "fa:16:3e:e1:02:b8", "network": {"id": "a8b7e1b4-1ec6-4226-bdd6-535dbce450d2", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1928758125-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "912300a09b80452e85e3cd13c4a644cc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7806fe18-2b89-4386-87b1-f22876f82af2", "external-id": "nsx-vlan-transportzone-727", "segmentation_id": 727, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap269a41b0-75", "ovs_interfaceid": "269a41b0-75e7-46af-9497-8eae3102df11", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 886.588933] env[61923]: DEBUG nova.network.neutron [req-79aa1fc8-f53c-4c0d-a13b-9bcb67526c91 req-a2ab7a4f-f911-41a2-911d-b8c40e2401d3 service nova] [instance: 422d06d6-6932-46c3-bb25-841e6f65c028] Updated VIF entry in instance network info cache for port 97caa1b8-427c-46b0-be7a-d59b932900c3. {{(pid=61923) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 886.588933] env[61923]: DEBUG nova.network.neutron [req-79aa1fc8-f53c-4c0d-a13b-9bcb67526c91 req-a2ab7a4f-f911-41a2-911d-b8c40e2401d3 service nova] [instance: 422d06d6-6932-46c3-bb25-841e6f65c028] Updating instance_info_cache with network_info: [{"id": "97caa1b8-427c-46b0-be7a-d59b932900c3", "address": "fa:16:3e:fa:f8:d1", "network": {"id": "a8b7e1b4-1ec6-4226-bdd6-535dbce450d2", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1928758125-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.146", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "912300a09b80452e85e3cd13c4a644cc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7806fe18-2b89-4386-87b1-f22876f82af2", "external-id": "nsx-vlan-transportzone-727", "segmentation_id": 727, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap97caa1b8-42", "ovs_interfaceid": "97caa1b8-427c-46b0-be7a-d59b932900c3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 886.591380] env[61923]: DEBUG oslo_concurrency.lockutils [req-208d0a7d-2f2d-46e5-8906-f7288e523826 req-deb0f534-1282-4140-b173-624dddd5a97d service nova] Releasing lock "refresh_cache-6012d3b5-739c-4762-9bb4-09c51171dcd7" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 886.592576] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-321110d9-2909-4d0e-b868-e8926c4a095a {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.609263] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94b90e5c-19e1-4a3a-abea-811444faa823 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.615186] env[61923]: DEBUG nova.compute.manager [req-b3b2f993-a16d-4f8a-a12b-1f3ab369a5bd req-854b1971-437f-471f-bcdd-fa87191b01c2 service nova] [instance: 422d06d6-6932-46c3-bb25-841e6f65c028] Received event network-vif-plugged-b395b183-1c30-42de-929f-690ea10d9d75 {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 886.615334] env[61923]: DEBUG oslo_concurrency.lockutils [req-b3b2f993-a16d-4f8a-a12b-1f3ab369a5bd req-854b1971-437f-471f-bcdd-fa87191b01c2 service nova] Acquiring lock "422d06d6-6932-46c3-bb25-841e6f65c028-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 886.616308] env[61923]: DEBUG oslo_concurrency.lockutils [req-b3b2f993-a16d-4f8a-a12b-1f3ab369a5bd req-854b1971-437f-471f-bcdd-fa87191b01c2 service nova] Lock "422d06d6-6932-46c3-bb25-841e6f65c028-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 886.616308] env[61923]: DEBUG oslo_concurrency.lockutils [req-b3b2f993-a16d-4f8a-a12b-1f3ab369a5bd req-854b1971-437f-471f-bcdd-fa87191b01c2 service nova] Lock "422d06d6-6932-46c3-bb25-841e6f65c028-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 886.616308] env[61923]: DEBUG nova.compute.manager [req-b3b2f993-a16d-4f8a-a12b-1f3ab369a5bd req-854b1971-437f-471f-bcdd-fa87191b01c2 service nova] [instance: 422d06d6-6932-46c3-bb25-841e6f65c028] No waiting events found dispatching network-vif-plugged-b395b183-1c30-42de-929f-690ea10d9d75 {{(pid=61923) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 886.616308] env[61923]: WARNING nova.compute.manager [req-b3b2f993-a16d-4f8a-a12b-1f3ab369a5bd req-854b1971-437f-471f-bcdd-fa87191b01c2 service nova] [instance: 422d06d6-6932-46c3-bb25-841e6f65c028] Received unexpected event network-vif-plugged-b395b183-1c30-42de-929f-690ea10d9d75 for instance with vm_state active and task_state None. [ 886.616491] env[61923]: DEBUG nova.compute.manager [req-b3b2f993-a16d-4f8a-a12b-1f3ab369a5bd req-854b1971-437f-471f-bcdd-fa87191b01c2 service nova] [instance: 422d06d6-6932-46c3-bb25-841e6f65c028] Received event network-changed-b395b183-1c30-42de-929f-690ea10d9d75 {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 886.616631] env[61923]: DEBUG nova.compute.manager [req-b3b2f993-a16d-4f8a-a12b-1f3ab369a5bd req-854b1971-437f-471f-bcdd-fa87191b01c2 service nova] [instance: 422d06d6-6932-46c3-bb25-841e6f65c028] Refreshing instance network info cache due to event network-changed-b395b183-1c30-42de-929f-690ea10d9d75. {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 886.616817] env[61923]: DEBUG oslo_concurrency.lockutils [req-b3b2f993-a16d-4f8a-a12b-1f3ab369a5bd req-854b1971-437f-471f-bcdd-fa87191b01c2 service nova] Acquiring lock "refresh_cache-422d06d6-6932-46c3-bb25-841e6f65c028" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 886.632657] env[61923]: DEBUG nova.compute.provider_tree [None req-3192dd2d-7513-4a93-8a8c-c00e6e67889c tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 886.742971] env[61923]: DEBUG oslo_vmware.api [None req-eb5e01f1-148c-4608-96e7-93572113f79a tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52427120-b448-37f5-7358-776548a06e34, 'name': SearchDatastore_Task, 'duration_secs': 0.011089} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 886.742971] env[61923]: DEBUG oslo_concurrency.lockutils [None req-eb5e01f1-148c-4608-96e7-93572113f79a tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 886.743363] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-eb5e01f1-148c-4608-96e7-93572113f79a tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk to [datastore1] 4c2c973a-5d23-4b8a-8b99-6b634e337054/4c2c973a-5d23-4b8a-8b99-6b634e337054.vmdk {{(pid=61923) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 886.743475] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-b8db493b-9e28-40b1-a35f-71a3284983f8 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.752771] env[61923]: DEBUG oslo_vmware.api [None req-eb5e01f1-148c-4608-96e7-93572113f79a tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Waiting for the task: (returnval){ [ 886.752771] env[61923]: value = "task-1377695" [ 886.752771] env[61923]: _type = "Task" [ 886.752771] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 886.762456] env[61923]: DEBUG oslo_vmware.api [None req-eb5e01f1-148c-4608-96e7-93572113f79a tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Task: {'id': task-1377695, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.875584] env[61923]: DEBUG oslo_vmware.rw_handles [None req-55b58374-2c35-46b9-a508-d4833264ecac tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Completed reading data from the image iterator. {{(pid=61923) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 886.876211] env[61923]: DEBUG oslo_vmware.rw_handles [None req-55b58374-2c35-46b9-a508-d4833264ecac tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52425bb0-dffe-722e-9ac5-23834c47972f/disk-0.vmdk. {{(pid=61923) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 886.877590] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a7737ad-b8f5-48b1-b094-c3f9eb13ab80 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.886656] env[61923]: DEBUG oslo_vmware.rw_handles [None req-55b58374-2c35-46b9-a508-d4833264ecac tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52425bb0-dffe-722e-9ac5-23834c47972f/disk-0.vmdk is in state: ready. {{(pid=61923) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 886.886656] env[61923]: DEBUG oslo_vmware.rw_handles [None req-55b58374-2c35-46b9-a508-d4833264ecac tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Releasing lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52425bb0-dffe-722e-9ac5-23834c47972f/disk-0.vmdk. {{(pid=61923) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 886.891614] env[61923]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-51852ea7-8ba0-49a0-a297-931304f67562 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.893718] env[61923]: DEBUG oslo_vmware.api [None req-4482c5e6-0d59-468c-9d9f-cee661c7eaed tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Task: {'id': task-1377694, 'name': CloneVM_Task} progress is 94%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.099188] env[61923]: DEBUG oslo_concurrency.lockutils [req-79aa1fc8-f53c-4c0d-a13b-9bcb67526c91 req-a2ab7a4f-f911-41a2-911d-b8c40e2401d3 service nova] Releasing lock "refresh_cache-422d06d6-6932-46c3-bb25-841e6f65c028" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 887.099188] env[61923]: DEBUG nova.compute.manager [req-79aa1fc8-f53c-4c0d-a13b-9bcb67526c91 req-a2ab7a4f-f911-41a2-911d-b8c40e2401d3 service nova] [instance: 4c2c973a-5d23-4b8a-8b99-6b634e337054] Received event network-vif-plugged-9584dacf-1833-4a40-bfbe-56d7d5ea9998 {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 887.099188] env[61923]: DEBUG oslo_concurrency.lockutils [req-79aa1fc8-f53c-4c0d-a13b-9bcb67526c91 req-a2ab7a4f-f911-41a2-911d-b8c40e2401d3 service nova] Acquiring lock "4c2c973a-5d23-4b8a-8b99-6b634e337054-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 887.099188] env[61923]: DEBUG oslo_concurrency.lockutils [req-79aa1fc8-f53c-4c0d-a13b-9bcb67526c91 req-a2ab7a4f-f911-41a2-911d-b8c40e2401d3 service nova] Lock "4c2c973a-5d23-4b8a-8b99-6b634e337054-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 887.099473] env[61923]: DEBUG oslo_concurrency.lockutils [req-79aa1fc8-f53c-4c0d-a13b-9bcb67526c91 req-a2ab7a4f-f911-41a2-911d-b8c40e2401d3 service nova] Lock "4c2c973a-5d23-4b8a-8b99-6b634e337054-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 887.099737] env[61923]: DEBUG nova.compute.manager [req-79aa1fc8-f53c-4c0d-a13b-9bcb67526c91 req-a2ab7a4f-f911-41a2-911d-b8c40e2401d3 service nova] [instance: 4c2c973a-5d23-4b8a-8b99-6b634e337054] No waiting events found dispatching network-vif-plugged-9584dacf-1833-4a40-bfbe-56d7d5ea9998 {{(pid=61923) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 887.100044] env[61923]: WARNING nova.compute.manager [req-79aa1fc8-f53c-4c0d-a13b-9bcb67526c91 req-a2ab7a4f-f911-41a2-911d-b8c40e2401d3 service nova] [instance: 4c2c973a-5d23-4b8a-8b99-6b634e337054] Received unexpected event network-vif-plugged-9584dacf-1833-4a40-bfbe-56d7d5ea9998 for instance with vm_state building and task_state spawning. [ 887.100352] env[61923]: DEBUG nova.compute.manager [req-79aa1fc8-f53c-4c0d-a13b-9bcb67526c91 req-a2ab7a4f-f911-41a2-911d-b8c40e2401d3 service nova] [instance: 4c2c973a-5d23-4b8a-8b99-6b634e337054] Received event network-changed-9584dacf-1833-4a40-bfbe-56d7d5ea9998 {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 887.100640] env[61923]: DEBUG nova.compute.manager [req-79aa1fc8-f53c-4c0d-a13b-9bcb67526c91 req-a2ab7a4f-f911-41a2-911d-b8c40e2401d3 service nova] [instance: 4c2c973a-5d23-4b8a-8b99-6b634e337054] Refreshing instance network info cache due to event network-changed-9584dacf-1833-4a40-bfbe-56d7d5ea9998. {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 887.100951] env[61923]: DEBUG oslo_concurrency.lockutils [req-79aa1fc8-f53c-4c0d-a13b-9bcb67526c91 req-a2ab7a4f-f911-41a2-911d-b8c40e2401d3 service nova] Acquiring lock "refresh_cache-4c2c973a-5d23-4b8a-8b99-6b634e337054" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 887.101194] env[61923]: DEBUG oslo_concurrency.lockutils [req-79aa1fc8-f53c-4c0d-a13b-9bcb67526c91 req-a2ab7a4f-f911-41a2-911d-b8c40e2401d3 service nova] Acquired lock "refresh_cache-4c2c973a-5d23-4b8a-8b99-6b634e337054" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 887.101458] env[61923]: DEBUG nova.network.neutron [req-79aa1fc8-f53c-4c0d-a13b-9bcb67526c91 req-a2ab7a4f-f911-41a2-911d-b8c40e2401d3 service nova] [instance: 4c2c973a-5d23-4b8a-8b99-6b634e337054] Refreshing network info cache for port 9584dacf-1833-4a40-bfbe-56d7d5ea9998 {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 887.103224] env[61923]: DEBUG oslo_concurrency.lockutils [None req-02b5a1b8-a821-49de-bb63-1c89123e0e10 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Acquired lock "refresh_cache-422d06d6-6932-46c3-bb25-841e6f65c028" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 887.103505] env[61923]: DEBUG nova.network.neutron [None req-02b5a1b8-a821-49de-bb63-1c89123e0e10 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] [instance: 422d06d6-6932-46c3-bb25-841e6f65c028] Building network info cache for instance {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 887.141209] env[61923]: DEBUG nova.scheduler.client.report [None req-3192dd2d-7513-4a93-8a8c-c00e6e67889c tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 887.211824] env[61923]: DEBUG oslo_vmware.rw_handles [None req-55b58374-2c35-46b9-a508-d4833264ecac tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Closed VMDK write handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52425bb0-dffe-722e-9ac5-23834c47972f/disk-0.vmdk. {{(pid=61923) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 887.212047] env[61923]: INFO nova.virt.vmwareapi.images [None req-55b58374-2c35-46b9-a508-d4833264ecac tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] [instance: 4cb3c397-ef1a-4267-a300-26da48206f3f] Downloaded image file data 9ddad7e6-f1e4-49aa-b956-1b8d3273318d [ 887.213041] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e10c1e6-fd17-4fe5-b450-19a7d76befd7 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.233795] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-6d5ab94c-4027-4109-8db5-99f6d476d981 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.262513] env[61923]: INFO nova.virt.vmwareapi.images [None req-55b58374-2c35-46b9-a508-d4833264ecac tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] [instance: 4cb3c397-ef1a-4267-a300-26da48206f3f] The imported VM was unregistered [ 887.269711] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-55b58374-2c35-46b9-a508-d4833264ecac tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] [instance: 4cb3c397-ef1a-4267-a300-26da48206f3f] Caching image {{(pid=61923) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 887.270250] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-55b58374-2c35-46b9-a508-d4833264ecac tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Creating directory with path [datastore2] devstack-image-cache_base/9ddad7e6-f1e4-49aa-b956-1b8d3273318d {{(pid=61923) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 887.274043] env[61923]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-40232c42-b882-4b46-bd9f-986b63f4c2aa {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.275726] env[61923]: DEBUG oslo_vmware.api [None req-eb5e01f1-148c-4608-96e7-93572113f79a tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Task: {'id': task-1377695, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.290124] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-55b58374-2c35-46b9-a508-d4833264ecac tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Created directory with path [datastore2] devstack-image-cache_base/9ddad7e6-f1e4-49aa-b956-1b8d3273318d {{(pid=61923) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 887.290397] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-55b58374-2c35-46b9-a508-d4833264ecac tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Moving virtual disk from [datastore2] OSTACK_IMG_345122e9-0048-4f77-a160-7345d2ae6537/OSTACK_IMG_345122e9-0048-4f77-a160-7345d2ae6537.vmdk to [datastore2] devstack-image-cache_base/9ddad7e6-f1e4-49aa-b956-1b8d3273318d/9ddad7e6-f1e4-49aa-b956-1b8d3273318d.vmdk. {{(pid=61923) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 887.290687] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-bd7cba89-f712-4d61-a6a8-5b8c1f5c43e7 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.300816] env[61923]: DEBUG oslo_vmware.api [None req-55b58374-2c35-46b9-a508-d4833264ecac tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Waiting for the task: (returnval){ [ 887.300816] env[61923]: value = "task-1377697" [ 887.300816] env[61923]: _type = "Task" [ 887.300816] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 887.312464] env[61923]: DEBUG oslo_vmware.api [None req-55b58374-2c35-46b9-a508-d4833264ecac tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Task: {'id': task-1377697, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.380386] env[61923]: DEBUG oslo_vmware.api [None req-4482c5e6-0d59-468c-9d9f-cee661c7eaed tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Task: {'id': task-1377694, 'name': CloneVM_Task} progress is 100%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.644925] env[61923]: DEBUG oslo_concurrency.lockutils [None req-3192dd2d-7513-4a93-8a8c-c00e6e67889c tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.350s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 887.648455] env[61923]: DEBUG nova.compute.manager [None req-3192dd2d-7513-4a93-8a8c-c00e6e67889c tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] [instance: 263f0866-49af-4ab6-8132-d993642ebedb] Start building networks asynchronously for instance. {{(pid=61923) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 887.651169] env[61923]: DEBUG oslo_concurrency.lockutils [None req-a5e456d9-94ec-4a32-9f7d-b8da7dd3daa2 tempest-ServersTestManualDisk-2012652654 tempest-ServersTestManualDisk-2012652654-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.466s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 887.651403] env[61923]: DEBUG nova.objects.instance [None req-a5e456d9-94ec-4a32-9f7d-b8da7dd3daa2 tempest-ServersTestManualDisk-2012652654 tempest-ServersTestManualDisk-2012652654-project-member] Lazy-loading 'resources' on Instance uuid f1839f61-4314-48fe-8ab6-14b5e49d569d {{(pid=61923) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 887.697781] env[61923]: WARNING nova.network.neutron [None req-02b5a1b8-a821-49de-bb63-1c89123e0e10 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] [instance: 422d06d6-6932-46c3-bb25-841e6f65c028] a8b7e1b4-1ec6-4226-bdd6-535dbce450d2 already exists in list: networks containing: ['a8b7e1b4-1ec6-4226-bdd6-535dbce450d2']. ignoring it [ 887.770293] env[61923]: DEBUG oslo_vmware.api [None req-eb5e01f1-148c-4608-96e7-93572113f79a tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Task: {'id': task-1377695, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.622768} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 887.775178] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-eb5e01f1-148c-4608-96e7-93572113f79a tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk to [datastore1] 4c2c973a-5d23-4b8a-8b99-6b634e337054/4c2c973a-5d23-4b8a-8b99-6b634e337054.vmdk {{(pid=61923) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 887.775813] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-eb5e01f1-148c-4608-96e7-93572113f79a tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: 4c2c973a-5d23-4b8a-8b99-6b634e337054] Extending root virtual disk to 1048576 {{(pid=61923) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 887.775813] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-fb2b108f-37fd-4555-9ff1-ced6740fd71e {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.785395] env[61923]: DEBUG oslo_vmware.api [None req-eb5e01f1-148c-4608-96e7-93572113f79a tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Waiting for the task: (returnval){ [ 887.785395] env[61923]: value = "task-1377698" [ 887.785395] env[61923]: _type = "Task" [ 887.785395] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 887.799768] env[61923]: DEBUG oslo_vmware.api [None req-eb5e01f1-148c-4608-96e7-93572113f79a tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Task: {'id': task-1377698, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.812923] env[61923]: DEBUG oslo_vmware.api [None req-55b58374-2c35-46b9-a508-d4833264ecac tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Task: {'id': task-1377697, 'name': MoveVirtualDisk_Task} progress is 12%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.881630] env[61923]: DEBUG oslo_vmware.api [None req-4482c5e6-0d59-468c-9d9f-cee661c7eaed tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Task: {'id': task-1377694, 'name': CloneVM_Task, 'duration_secs': 1.566089} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 887.881998] env[61923]: INFO nova.virt.vmwareapi.vmops [None req-4482c5e6-0d59-468c-9d9f-cee661c7eaed tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] [instance: 2719569b-8572-4199-8158-7bb367d17dc5] Created linked-clone VM from snapshot [ 887.883069] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a83c268-e26f-4c22-ba87-5f3a4c1bca7b {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.892823] env[61923]: DEBUG nova.virt.vmwareapi.images [None req-4482c5e6-0d59-468c-9d9f-cee661c7eaed tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] [instance: 2719569b-8572-4199-8158-7bb367d17dc5] Uploading image b1a1622a-32c0-45e0-b073-5adb9b6d5da6 {{(pid=61923) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 887.931549] env[61923]: DEBUG oslo_vmware.rw_handles [None req-4482c5e6-0d59-468c-9d9f-cee661c7eaed tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 887.931549] env[61923]: value = "vm-292704" [ 887.931549] env[61923]: _type = "VirtualMachine" [ 887.931549] env[61923]: }. {{(pid=61923) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 887.932366] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-482ead9b-bf0c-4bd9-bf2b-4c36a94c7096 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.942065] env[61923]: DEBUG oslo_vmware.rw_handles [None req-4482c5e6-0d59-468c-9d9f-cee661c7eaed tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Lease: (returnval){ [ 887.942065] env[61923]: value = "session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]520e5bb6-e351-c878-37ab-a46918fc3b10" [ 887.942065] env[61923]: _type = "HttpNfcLease" [ 887.942065] env[61923]: } obtained for exporting VM: (result){ [ 887.942065] env[61923]: value = "vm-292704" [ 887.942065] env[61923]: _type = "VirtualMachine" [ 887.942065] env[61923]: }. {{(pid=61923) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 887.942715] env[61923]: DEBUG oslo_vmware.api [None req-4482c5e6-0d59-468c-9d9f-cee661c7eaed tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Waiting for the lease: (returnval){ [ 887.942715] env[61923]: value = "session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]520e5bb6-e351-c878-37ab-a46918fc3b10" [ 887.942715] env[61923]: _type = "HttpNfcLease" [ 887.942715] env[61923]: } to be ready. {{(pid=61923) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 887.958822] env[61923]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 887.958822] env[61923]: value = "session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]520e5bb6-e351-c878-37ab-a46918fc3b10" [ 887.958822] env[61923]: _type = "HttpNfcLease" [ 887.958822] env[61923]: } is initializing. {{(pid=61923) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 888.119035] env[61923]: DEBUG nova.network.neutron [req-79aa1fc8-f53c-4c0d-a13b-9bcb67526c91 req-a2ab7a4f-f911-41a2-911d-b8c40e2401d3 service nova] [instance: 4c2c973a-5d23-4b8a-8b99-6b634e337054] Updated VIF entry in instance network info cache for port 9584dacf-1833-4a40-bfbe-56d7d5ea9998. {{(pid=61923) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 888.119921] env[61923]: DEBUG nova.network.neutron [req-79aa1fc8-f53c-4c0d-a13b-9bcb67526c91 req-a2ab7a4f-f911-41a2-911d-b8c40e2401d3 service nova] [instance: 4c2c973a-5d23-4b8a-8b99-6b634e337054] Updating instance_info_cache with network_info: [{"id": "9584dacf-1833-4a40-bfbe-56d7d5ea9998", "address": "fa:16:3e:48:9c:20", "network": {"id": "d1c38272-0461-43e5-93ae-3964346f77d0", "bridge": "br-int", "label": "tempest-ServersTestJSON-151839165-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "69cc941eb6dd4780ac12aa29656c37f7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8ba3bd22-c936-470e-89bd-b3a5587e87a0", "external-id": "nsx-vlan-transportzone-605", "segmentation_id": 605, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9584dacf-18", "ovs_interfaceid": "9584dacf-1833-4a40-bfbe-56d7d5ea9998", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 888.149611] env[61923]: DEBUG oslo_concurrency.lockutils [None req-6cea46fb-73b5-4047-8297-b9cdc5a261d0 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Acquiring lock "bac71328-3796-4a65-aa41-ccda93bc7926" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 888.150460] env[61923]: DEBUG oslo_concurrency.lockutils [None req-6cea46fb-73b5-4047-8297-b9cdc5a261d0 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Lock "bac71328-3796-4a65-aa41-ccda93bc7926" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 888.154111] env[61923]: DEBUG nova.compute.utils [None req-3192dd2d-7513-4a93-8a8c-c00e6e67889c tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] Using /dev/sd instead of None {{(pid=61923) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 888.158524] env[61923]: DEBUG nova.compute.manager [None req-3192dd2d-7513-4a93-8a8c-c00e6e67889c tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] [instance: 263f0866-49af-4ab6-8132-d993642ebedb] Allocating IP information in the background. {{(pid=61923) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 888.158710] env[61923]: DEBUG nova.network.neutron [None req-3192dd2d-7513-4a93-8a8c-c00e6e67889c tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] [instance: 263f0866-49af-4ab6-8132-d993642ebedb] allocate_for_instance() {{(pid=61923) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 888.237719] env[61923]: DEBUG nova.policy [None req-3192dd2d-7513-4a93-8a8c-c00e6e67889c tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e423cb2e47934db39ae2c17a22ee0ceb', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c304b5b5d4df485fbda55a570494a62f', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61923) authorize /opt/stack/nova/nova/policy.py:201}} [ 888.304242] env[61923]: DEBUG oslo_vmware.api [None req-eb5e01f1-148c-4608-96e7-93572113f79a tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Task: {'id': task-1377698, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.091876} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 888.307524] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-eb5e01f1-148c-4608-96e7-93572113f79a tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: 4c2c973a-5d23-4b8a-8b99-6b634e337054] Extended root virtual disk {{(pid=61923) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 888.312244] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b768d4df-14f1-4714-b0ae-6ba4eba8759d {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.334578] env[61923]: DEBUG oslo_vmware.api [None req-55b58374-2c35-46b9-a508-d4833264ecac tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Task: {'id': task-1377697, 'name': MoveVirtualDisk_Task} progress is 35%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.345015] env[61923]: DEBUG nova.virt.vmwareapi.volumeops [None req-eb5e01f1-148c-4608-96e7-93572113f79a tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: 4c2c973a-5d23-4b8a-8b99-6b634e337054] Reconfiguring VM instance instance-00000051 to attach disk [datastore1] 4c2c973a-5d23-4b8a-8b99-6b634e337054/4c2c973a-5d23-4b8a-8b99-6b634e337054.vmdk or device None with type sparse {{(pid=61923) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 888.348440] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-200c93ce-7603-459a-9f01-b2b98e12daab {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.366576] env[61923]: DEBUG nova.network.neutron [None req-02b5a1b8-a821-49de-bb63-1c89123e0e10 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] [instance: 422d06d6-6932-46c3-bb25-841e6f65c028] Updating instance_info_cache with network_info: [{"id": "97caa1b8-427c-46b0-be7a-d59b932900c3", "address": "fa:16:3e:fa:f8:d1", "network": {"id": "a8b7e1b4-1ec6-4226-bdd6-535dbce450d2", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1928758125-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.146", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "912300a09b80452e85e3cd13c4a644cc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7806fe18-2b89-4386-87b1-f22876f82af2", "external-id": "nsx-vlan-transportzone-727", "segmentation_id": 727, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap97caa1b8-42", "ovs_interfaceid": "97caa1b8-427c-46b0-be7a-d59b932900c3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "b395b183-1c30-42de-929f-690ea10d9d75", "address": "fa:16:3e:7a:d5:2a", "network": {"id": "a8b7e1b4-1ec6-4226-bdd6-535dbce450d2", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1928758125-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "912300a09b80452e85e3cd13c4a644cc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7806fe18-2b89-4386-87b1-f22876f82af2", "external-id": "nsx-vlan-transportzone-727", "segmentation_id": 727, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb395b183-1c", "ovs_interfaceid": "b395b183-1c30-42de-929f-690ea10d9d75", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 888.375333] env[61923]: DEBUG oslo_vmware.api [None req-eb5e01f1-148c-4608-96e7-93572113f79a tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Waiting for the task: (returnval){ [ 888.375333] env[61923]: value = "task-1377700" [ 888.375333] env[61923]: _type = "Task" [ 888.375333] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 888.389016] env[61923]: DEBUG oslo_vmware.api [None req-eb5e01f1-148c-4608-96e7-93572113f79a tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Task: {'id': task-1377700, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.460440] env[61923]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 888.460440] env[61923]: value = "session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]520e5bb6-e351-c878-37ab-a46918fc3b10" [ 888.460440] env[61923]: _type = "HttpNfcLease" [ 888.460440] env[61923]: } is ready. {{(pid=61923) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 888.460801] env[61923]: DEBUG oslo_vmware.rw_handles [None req-4482c5e6-0d59-468c-9d9f-cee661c7eaed tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 888.460801] env[61923]: value = "session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]520e5bb6-e351-c878-37ab-a46918fc3b10" [ 888.460801] env[61923]: _type = "HttpNfcLease" [ 888.460801] env[61923]: }. {{(pid=61923) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 888.461655] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e004674-f119-4fd1-b0ff-30e066ef9f90 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.466603] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b74aa194-bd09-4316-8ba4-2f54152b67ac {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.475439] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bae9992b-640d-4628-80c3-caa5076a42f6 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.482362] env[61923]: DEBUG oslo_vmware.rw_handles [None req-4482c5e6-0d59-468c-9d9f-cee661c7eaed tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52c8888b-9c22-85ef-0fcf-b5c96df0cb46/disk-0.vmdk from lease info. {{(pid=61923) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 888.482645] env[61923]: DEBUG oslo_vmware.rw_handles [None req-4482c5e6-0d59-468c-9d9f-cee661c7eaed tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52c8888b-9c22-85ef-0fcf-b5c96df0cb46/disk-0.vmdk for reading. {{(pid=61923) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 888.579496] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec4105c9-3626-484f-b7fe-5529c639c5b4 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.587117] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50d48bb5-6a6a-4eb4-b496-ee9a0cfed33d {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.604870] env[61923]: DEBUG nova.compute.provider_tree [None req-a5e456d9-94ec-4a32-9f7d-b8da7dd3daa2 tempest-ServersTestManualDisk-2012652654 tempest-ServersTestManualDisk-2012652654-project-member] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 888.623978] env[61923]: DEBUG oslo_concurrency.lockutils [req-79aa1fc8-f53c-4c0d-a13b-9bcb67526c91 req-a2ab7a4f-f911-41a2-911d-b8c40e2401d3 service nova] Releasing lock "refresh_cache-4c2c973a-5d23-4b8a-8b99-6b634e337054" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 888.624464] env[61923]: DEBUG nova.compute.manager [req-79aa1fc8-f53c-4c0d-a13b-9bcb67526c91 req-a2ab7a4f-f911-41a2-911d-b8c40e2401d3 service nova] [instance: f1839f61-4314-48fe-8ab6-14b5e49d569d] Received event network-vif-deleted-f36077f4-1c49-48e7-9c41-0f45a58883e7 {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 888.624538] env[61923]: INFO nova.compute.manager [req-79aa1fc8-f53c-4c0d-a13b-9bcb67526c91 req-a2ab7a4f-f911-41a2-911d-b8c40e2401d3 service nova] [instance: f1839f61-4314-48fe-8ab6-14b5e49d569d] Neutron deleted interface f36077f4-1c49-48e7-9c41-0f45a58883e7; detaching it from the instance and deleting it from the info cache [ 888.624706] env[61923]: DEBUG nova.network.neutron [req-79aa1fc8-f53c-4c0d-a13b-9bcb67526c91 req-a2ab7a4f-f911-41a2-911d-b8c40e2401d3 service nova] [instance: f1839f61-4314-48fe-8ab6-14b5e49d569d] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 888.632225] env[61923]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-79b8593a-b31c-46ac-b6ba-47dc84d6d640 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.640400] env[61923]: DEBUG nova.network.neutron [None req-3192dd2d-7513-4a93-8a8c-c00e6e67889c tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] [instance: 263f0866-49af-4ab6-8132-d993642ebedb] Successfully created port: 99825a7a-e38f-449b-8633-318da7f11ab5 {{(pid=61923) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 888.652612] env[61923]: DEBUG nova.compute.manager [None req-6cea46fb-73b5-4047-8297-b9cdc5a261d0 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] [instance: bac71328-3796-4a65-aa41-ccda93bc7926] Starting instance... {{(pid=61923) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 888.660125] env[61923]: DEBUG nova.compute.manager [None req-3192dd2d-7513-4a93-8a8c-c00e6e67889c tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] [instance: 263f0866-49af-4ab6-8132-d993642ebedb] Start building block device mappings for instance. {{(pid=61923) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 888.818147] env[61923]: DEBUG oslo_vmware.api [None req-55b58374-2c35-46b9-a508-d4833264ecac tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Task: {'id': task-1377697, 'name': MoveVirtualDisk_Task} progress is 52%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.870476] env[61923]: DEBUG oslo_concurrency.lockutils [None req-02b5a1b8-a821-49de-bb63-1c89123e0e10 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Releasing lock "refresh_cache-422d06d6-6932-46c3-bb25-841e6f65c028" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 888.871476] env[61923]: DEBUG oslo_concurrency.lockutils [None req-02b5a1b8-a821-49de-bb63-1c89123e0e10 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Acquiring lock "422d06d6-6932-46c3-bb25-841e6f65c028" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 888.871635] env[61923]: DEBUG oslo_concurrency.lockutils [None req-02b5a1b8-a821-49de-bb63-1c89123e0e10 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Acquired lock "422d06d6-6932-46c3-bb25-841e6f65c028" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 888.871858] env[61923]: DEBUG oslo_concurrency.lockutils [req-b3b2f993-a16d-4f8a-a12b-1f3ab369a5bd req-854b1971-437f-471f-bcdd-fa87191b01c2 service nova] Acquired lock "refresh_cache-422d06d6-6932-46c3-bb25-841e6f65c028" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 888.872143] env[61923]: DEBUG nova.network.neutron [req-b3b2f993-a16d-4f8a-a12b-1f3ab369a5bd req-854b1971-437f-471f-bcdd-fa87191b01c2 service nova] [instance: 422d06d6-6932-46c3-bb25-841e6f65c028] Refreshing network info cache for port b395b183-1c30-42de-929f-690ea10d9d75 {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 888.874390] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24443e2e-51c1-49ea-9609-9ae52e53ca61 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.904393] env[61923]: DEBUG nova.virt.hardware [None req-02b5a1b8-a821-49de-bb63-1c89123e0e10 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-29T20:07:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 888.904609] env[61923]: DEBUG nova.virt.hardware [None req-02b5a1b8-a821-49de-bb63-1c89123e0e10 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Flavor limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 888.904762] env[61923]: DEBUG nova.virt.hardware [None req-02b5a1b8-a821-49de-bb63-1c89123e0e10 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Image limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 888.904948] env[61923]: DEBUG nova.virt.hardware [None req-02b5a1b8-a821-49de-bb63-1c89123e0e10 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Flavor pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 888.905115] env[61923]: DEBUG nova.virt.hardware [None req-02b5a1b8-a821-49de-bb63-1c89123e0e10 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Image pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 888.905274] env[61923]: DEBUG nova.virt.hardware [None req-02b5a1b8-a821-49de-bb63-1c89123e0e10 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 888.905491] env[61923]: DEBUG nova.virt.hardware [None req-02b5a1b8-a821-49de-bb63-1c89123e0e10 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 888.905922] env[61923]: DEBUG nova.virt.hardware [None req-02b5a1b8-a821-49de-bb63-1c89123e0e10 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 888.906242] env[61923]: DEBUG nova.virt.hardware [None req-02b5a1b8-a821-49de-bb63-1c89123e0e10 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Got 1 possible topologies {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 888.906428] env[61923]: DEBUG nova.virt.hardware [None req-02b5a1b8-a821-49de-bb63-1c89123e0e10 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 888.906608] env[61923]: DEBUG nova.virt.hardware [None req-02b5a1b8-a821-49de-bb63-1c89123e0e10 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 888.913664] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-02b5a1b8-a821-49de-bb63-1c89123e0e10 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] [instance: 422d06d6-6932-46c3-bb25-841e6f65c028] Reconfiguring VM to attach interface {{(pid=61923) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1923}} [ 888.921292] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8de0797e-0627-48db-9d96-41c7c635a0a8 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.937270] env[61923]: DEBUG oslo_vmware.api [None req-eb5e01f1-148c-4608-96e7-93572113f79a tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Task: {'id': task-1377700, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.951040] env[61923]: DEBUG oslo_vmware.api [None req-02b5a1b8-a821-49de-bb63-1c89123e0e10 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Waiting for the task: (returnval){ [ 888.951040] env[61923]: value = "task-1377701" [ 888.951040] env[61923]: _type = "Task" [ 888.951040] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 888.964828] env[61923]: DEBUG oslo_vmware.api [None req-02b5a1b8-a821-49de-bb63-1c89123e0e10 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Task: {'id': task-1377701, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.003626] env[61923]: DEBUG oslo_concurrency.lockutils [None req-b7a2d28f-39dc-4a35-bb16-94846b77bc41 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Acquiring lock "c98c96a5-d18e-4519-996a-949ded16225f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 889.004049] env[61923]: DEBUG oslo_concurrency.lockutils [None req-b7a2d28f-39dc-4a35-bb16-94846b77bc41 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Lock "c98c96a5-d18e-4519-996a-949ded16225f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 889.110536] env[61923]: DEBUG nova.scheduler.client.report [None req-a5e456d9-94ec-4a32-9f7d-b8da7dd3daa2 tempest-ServersTestManualDisk-2012652654 tempest-ServersTestManualDisk-2012652654-project-member] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 889.128446] env[61923]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-b029e1ff-c6dc-4e5a-90bc-5b10efc8656c {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.155445] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a9459a6-e4b2-4766-a5f9-2d4c50a79ebd {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.170366] env[61923]: DEBUG nova.network.neutron [None req-3192dd2d-7513-4a93-8a8c-c00e6e67889c tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] [instance: 263f0866-49af-4ab6-8132-d993642ebedb] Successfully created port: a16defc4-1374-44de-a7c1-e2d75f27e954 {{(pid=61923) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 889.209055] env[61923]: DEBUG nova.compute.manager [req-79aa1fc8-f53c-4c0d-a13b-9bcb67526c91 req-a2ab7a4f-f911-41a2-911d-b8c40e2401d3 service nova] [instance: f1839f61-4314-48fe-8ab6-14b5e49d569d] Detach interface failed, port_id=f36077f4-1c49-48e7-9c41-0f45a58883e7, reason: Instance f1839f61-4314-48fe-8ab6-14b5e49d569d could not be found. {{(pid=61923) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 889.210727] env[61923]: DEBUG oslo_concurrency.lockutils [None req-6cea46fb-73b5-4047-8297-b9cdc5a261d0 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 889.318604] env[61923]: DEBUG oslo_vmware.api [None req-55b58374-2c35-46b9-a508-d4833264ecac tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Task: {'id': task-1377697, 'name': MoveVirtualDisk_Task} progress is 71%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.399724] env[61923]: DEBUG oslo_vmware.api [None req-eb5e01f1-148c-4608-96e7-93572113f79a tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Task: {'id': task-1377700, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.468582] env[61923]: DEBUG oslo_vmware.api [None req-02b5a1b8-a821-49de-bb63-1c89123e0e10 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Task: {'id': task-1377701, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.507038] env[61923]: DEBUG nova.compute.manager [None req-b7a2d28f-39dc-4a35-bb16-94846b77bc41 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] [instance: c98c96a5-d18e-4519-996a-949ded16225f] Starting instance... {{(pid=61923) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 889.617132] env[61923]: DEBUG oslo_concurrency.lockutils [None req-a5e456d9-94ec-4a32-9f7d-b8da7dd3daa2 tempest-ServersTestManualDisk-2012652654 tempest-ServersTestManualDisk-2012652654-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.966s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 889.619719] env[61923]: DEBUG oslo_concurrency.lockutils [None req-6cea46fb-73b5-4047-8297-b9cdc5a261d0 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.409s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 889.622446] env[61923]: INFO nova.compute.claims [None req-6cea46fb-73b5-4047-8297-b9cdc5a261d0 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] [instance: bac71328-3796-4a65-aa41-ccda93bc7926] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 889.642778] env[61923]: INFO nova.scheduler.client.report [None req-a5e456d9-94ec-4a32-9f7d-b8da7dd3daa2 tempest-ServersTestManualDisk-2012652654 tempest-ServersTestManualDisk-2012652654-project-member] Deleted allocations for instance f1839f61-4314-48fe-8ab6-14b5e49d569d [ 889.683890] env[61923]: DEBUG nova.compute.manager [None req-3192dd2d-7513-4a93-8a8c-c00e6e67889c tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] [instance: 263f0866-49af-4ab6-8132-d993642ebedb] Start spawning the instance on the hypervisor. {{(pid=61923) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 889.716996] env[61923]: DEBUG nova.virt.hardware [None req-3192dd2d-7513-4a93-8a8c-c00e6e67889c tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-29T20:07:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-29T20:07:35Z,direct_url=,disk_format='vmdk',id=0f153f63-ae0a-45b1-b7f5-7f9b673c947f,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='4e7b5e3b3c3443c8bd5c70f8a15739f5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-29T20:07:36Z,virtual_size=,visibility=), allow threads: False {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 889.717519] env[61923]: DEBUG nova.virt.hardware [None req-3192dd2d-7513-4a93-8a8c-c00e6e67889c tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] Flavor limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 889.717698] env[61923]: DEBUG nova.virt.hardware [None req-3192dd2d-7513-4a93-8a8c-c00e6e67889c tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] Image limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 889.718064] env[61923]: DEBUG nova.virt.hardware [None req-3192dd2d-7513-4a93-8a8c-c00e6e67889c tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] Flavor pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 889.718241] env[61923]: DEBUG nova.virt.hardware [None req-3192dd2d-7513-4a93-8a8c-c00e6e67889c tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] Image pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 889.718649] env[61923]: DEBUG nova.virt.hardware [None req-3192dd2d-7513-4a93-8a8c-c00e6e67889c tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 889.718979] env[61923]: DEBUG nova.virt.hardware [None req-3192dd2d-7513-4a93-8a8c-c00e6e67889c tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 889.719304] env[61923]: DEBUG nova.virt.hardware [None req-3192dd2d-7513-4a93-8a8c-c00e6e67889c tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 889.719552] env[61923]: DEBUG nova.virt.hardware [None req-3192dd2d-7513-4a93-8a8c-c00e6e67889c tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] Got 1 possible topologies {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 889.719905] env[61923]: DEBUG nova.virt.hardware [None req-3192dd2d-7513-4a93-8a8c-c00e6e67889c tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 889.720228] env[61923]: DEBUG nova.virt.hardware [None req-3192dd2d-7513-4a93-8a8c-c00e6e67889c tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 889.721638] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5cb0acef-20cb-4782-89e1-ca34ca7ae0cb {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.738408] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4b37371-fab8-4cc4-a8e8-8ffa412c6721 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.765888] env[61923]: DEBUG oslo_concurrency.lockutils [None req-11f7570e-7eca-4d2f-9bb4-0fd5b38bff3d tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] Acquiring lock "af3c317a-4007-4cea-a060-1e7dde5ce49e" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 889.766193] env[61923]: DEBUG oslo_concurrency.lockutils [None req-11f7570e-7eca-4d2f-9bb4-0fd5b38bff3d tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] Lock "af3c317a-4007-4cea-a060-1e7dde5ce49e" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.001s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 889.771122] env[61923]: DEBUG nova.network.neutron [req-b3b2f993-a16d-4f8a-a12b-1f3ab369a5bd req-854b1971-437f-471f-bcdd-fa87191b01c2 service nova] [instance: 422d06d6-6932-46c3-bb25-841e6f65c028] Updated VIF entry in instance network info cache for port b395b183-1c30-42de-929f-690ea10d9d75. {{(pid=61923) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 889.771729] env[61923]: DEBUG nova.network.neutron [req-b3b2f993-a16d-4f8a-a12b-1f3ab369a5bd req-854b1971-437f-471f-bcdd-fa87191b01c2 service nova] [instance: 422d06d6-6932-46c3-bb25-841e6f65c028] Updating instance_info_cache with network_info: [{"id": "97caa1b8-427c-46b0-be7a-d59b932900c3", "address": "fa:16:3e:fa:f8:d1", "network": {"id": "a8b7e1b4-1ec6-4226-bdd6-535dbce450d2", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1928758125-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.146", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "912300a09b80452e85e3cd13c4a644cc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7806fe18-2b89-4386-87b1-f22876f82af2", "external-id": "nsx-vlan-transportzone-727", "segmentation_id": 727, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap97caa1b8-42", "ovs_interfaceid": "97caa1b8-427c-46b0-be7a-d59b932900c3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "b395b183-1c30-42de-929f-690ea10d9d75", "address": "fa:16:3e:7a:d5:2a", "network": {"id": "a8b7e1b4-1ec6-4226-bdd6-535dbce450d2", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1928758125-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "912300a09b80452e85e3cd13c4a644cc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7806fe18-2b89-4386-87b1-f22876f82af2", "external-id": "nsx-vlan-transportzone-727", "segmentation_id": 727, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb395b183-1c", "ovs_interfaceid": "b395b183-1c30-42de-929f-690ea10d9d75", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 889.820302] env[61923]: DEBUG oslo_vmware.api [None req-55b58374-2c35-46b9-a508-d4833264ecac tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Task: {'id': task-1377697, 'name': MoveVirtualDisk_Task} progress is 94%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.894528] env[61923]: DEBUG oslo_concurrency.lockutils [None req-e1b75445-6f25-4dcb-9846-d40076ac5b42 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Acquiring lock "624e9207-7f24-4f45-8f52-0fc7a60c1e84" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 889.894934] env[61923]: DEBUG oslo_concurrency.lockutils [None req-e1b75445-6f25-4dcb-9846-d40076ac5b42 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Lock "624e9207-7f24-4f45-8f52-0fc7a60c1e84" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 889.902552] env[61923]: DEBUG oslo_vmware.api [None req-eb5e01f1-148c-4608-96e7-93572113f79a tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Task: {'id': task-1377700, 'name': ReconfigVM_Task, 'duration_secs': 1.49512} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 889.903054] env[61923]: DEBUG nova.virt.vmwareapi.volumeops [None req-eb5e01f1-148c-4608-96e7-93572113f79a tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: 4c2c973a-5d23-4b8a-8b99-6b634e337054] Reconfigured VM instance instance-00000051 to attach disk [datastore1] 4c2c973a-5d23-4b8a-8b99-6b634e337054/4c2c973a-5d23-4b8a-8b99-6b634e337054.vmdk or device None with type sparse {{(pid=61923) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 889.904054] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-cd6bdcea-6bac-4094-9f03-8107b8bd3961 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.916964] env[61923]: DEBUG oslo_vmware.api [None req-eb5e01f1-148c-4608-96e7-93572113f79a tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Waiting for the task: (returnval){ [ 889.916964] env[61923]: value = "task-1377702" [ 889.916964] env[61923]: _type = "Task" [ 889.916964] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 889.931510] env[61923]: DEBUG oslo_vmware.api [None req-eb5e01f1-148c-4608-96e7-93572113f79a tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Task: {'id': task-1377702, 'name': Rename_Task} progress is 5%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.964868] env[61923]: DEBUG oslo_vmware.api [None req-02b5a1b8-a821-49de-bb63-1c89123e0e10 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Task: {'id': task-1377701, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.033386] env[61923]: DEBUG oslo_concurrency.lockutils [None req-b7a2d28f-39dc-4a35-bb16-94846b77bc41 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 890.156250] env[61923]: DEBUG oslo_concurrency.lockutils [None req-a5e456d9-94ec-4a32-9f7d-b8da7dd3daa2 tempest-ServersTestManualDisk-2012652654 tempest-ServersTestManualDisk-2012652654-project-member] Lock "f1839f61-4314-48fe-8ab6-14b5e49d569d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 8.142s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 890.270126] env[61923]: DEBUG nova.compute.utils [None req-11f7570e-7eca-4d2f-9bb4-0fd5b38bff3d tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] Using /dev/sd instead of None {{(pid=61923) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 890.274587] env[61923]: DEBUG oslo_concurrency.lockutils [req-b3b2f993-a16d-4f8a-a12b-1f3ab369a5bd req-854b1971-437f-471f-bcdd-fa87191b01c2 service nova] Releasing lock "refresh_cache-422d06d6-6932-46c3-bb25-841e6f65c028" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 890.316777] env[61923]: DEBUG oslo_vmware.api [None req-55b58374-2c35-46b9-a508-d4833264ecac tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Task: {'id': task-1377697, 'name': MoveVirtualDisk_Task, 'duration_secs': 2.724222} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 890.317192] env[61923]: INFO nova.virt.vmwareapi.ds_util [None req-55b58374-2c35-46b9-a508-d4833264ecac tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Moved virtual disk from [datastore2] OSTACK_IMG_345122e9-0048-4f77-a160-7345d2ae6537/OSTACK_IMG_345122e9-0048-4f77-a160-7345d2ae6537.vmdk to [datastore2] devstack-image-cache_base/9ddad7e6-f1e4-49aa-b956-1b8d3273318d/9ddad7e6-f1e4-49aa-b956-1b8d3273318d.vmdk. [ 890.317486] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-55b58374-2c35-46b9-a508-d4833264ecac tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] [instance: 4cb3c397-ef1a-4267-a300-26da48206f3f] Cleaning up location [datastore2] OSTACK_IMG_345122e9-0048-4f77-a160-7345d2ae6537 {{(pid=61923) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 890.317722] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-55b58374-2c35-46b9-a508-d4833264ecac tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Deleting the datastore file [datastore2] OSTACK_IMG_345122e9-0048-4f77-a160-7345d2ae6537 {{(pid=61923) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 890.318056] env[61923]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-beee01bf-5a84-4a6c-98c1-7c3e862e6086 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.325199] env[61923]: DEBUG oslo_vmware.api [None req-55b58374-2c35-46b9-a508-d4833264ecac tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Waiting for the task: (returnval){ [ 890.325199] env[61923]: value = "task-1377703" [ 890.325199] env[61923]: _type = "Task" [ 890.325199] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 890.333428] env[61923]: DEBUG oslo_vmware.api [None req-55b58374-2c35-46b9-a508-d4833264ecac tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Task: {'id': task-1377703, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.397584] env[61923]: DEBUG nova.compute.manager [None req-e1b75445-6f25-4dcb-9846-d40076ac5b42 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] [instance: 624e9207-7f24-4f45-8f52-0fc7a60c1e84] Starting instance... {{(pid=61923) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 890.427328] env[61923]: DEBUG oslo_vmware.api [None req-eb5e01f1-148c-4608-96e7-93572113f79a tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Task: {'id': task-1377702, 'name': Rename_Task, 'duration_secs': 0.36072} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 890.427691] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-eb5e01f1-148c-4608-96e7-93572113f79a tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: 4c2c973a-5d23-4b8a-8b99-6b634e337054] Powering on the VM {{(pid=61923) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 890.428609] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c302b963-20e5-4772-958a-9c0aeea0939c {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.438018] env[61923]: DEBUG oslo_vmware.api [None req-eb5e01f1-148c-4608-96e7-93572113f79a tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Waiting for the task: (returnval){ [ 890.438018] env[61923]: value = "task-1377704" [ 890.438018] env[61923]: _type = "Task" [ 890.438018] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 890.446851] env[61923]: DEBUG oslo_vmware.api [None req-eb5e01f1-148c-4608-96e7-93572113f79a tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Task: {'id': task-1377704, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.462508] env[61923]: DEBUG oslo_vmware.api [None req-02b5a1b8-a821-49de-bb63-1c89123e0e10 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Task: {'id': task-1377701, 'name': ReconfigVM_Task, 'duration_secs': 1.188963} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 890.463617] env[61923]: DEBUG oslo_concurrency.lockutils [None req-02b5a1b8-a821-49de-bb63-1c89123e0e10 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Releasing lock "422d06d6-6932-46c3-bb25-841e6f65c028" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 890.463617] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-02b5a1b8-a821-49de-bb63-1c89123e0e10 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] [instance: 422d06d6-6932-46c3-bb25-841e6f65c028] Reconfigured VM to attach interface {{(pid=61923) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1937}} [ 890.773649] env[61923]: DEBUG oslo_concurrency.lockutils [None req-11f7570e-7eca-4d2f-9bb4-0fd5b38bff3d tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] Lock "af3c317a-4007-4cea-a060-1e7dde5ce49e" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.007s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 890.836832] env[61923]: DEBUG oslo_vmware.api [None req-55b58374-2c35-46b9-a508-d4833264ecac tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Task: {'id': task-1377703, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.035265} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 890.837258] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-55b58374-2c35-46b9-a508-d4833264ecac tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Deleted the datastore file {{(pid=61923) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 890.837515] env[61923]: DEBUG oslo_concurrency.lockutils [None req-55b58374-2c35-46b9-a508-d4833264ecac tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Releasing lock "[datastore2] devstack-image-cache_base/9ddad7e6-f1e4-49aa-b956-1b8d3273318d/9ddad7e6-f1e4-49aa-b956-1b8d3273318d.vmdk" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 890.837799] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-55b58374-2c35-46b9-a508-d4833264ecac tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/9ddad7e6-f1e4-49aa-b956-1b8d3273318d/9ddad7e6-f1e4-49aa-b956-1b8d3273318d.vmdk to [datastore2] 4cb3c397-ef1a-4267-a300-26da48206f3f/4cb3c397-ef1a-4267-a300-26da48206f3f.vmdk {{(pid=61923) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 890.838100] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-b4d349e0-cad7-4564-ad73-2c6bc18712d1 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.847261] env[61923]: DEBUG oslo_vmware.api [None req-55b58374-2c35-46b9-a508-d4833264ecac tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Waiting for the task: (returnval){ [ 890.847261] env[61923]: value = "task-1377705" [ 890.847261] env[61923]: _type = "Task" [ 890.847261] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 890.853395] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-edffb146-db86-4ff0-a5bd-db77299d1e43 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.859478] env[61923]: DEBUG oslo_vmware.api [None req-55b58374-2c35-46b9-a508-d4833264ecac tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Task: {'id': task-1377705, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.865085] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19ff8104-6312-4565-80f0-ea68c4b845e1 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.900845] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b11dbb6-6abe-443a-91d9-af7ed1a7ce3d {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.912488] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-589885e3-e652-4b40-bfc0-650132c96d15 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.929915] env[61923]: DEBUG oslo_concurrency.lockutils [None req-e1b75445-6f25-4dcb-9846-d40076ac5b42 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 890.931769] env[61923]: DEBUG nova.compute.provider_tree [None req-6cea46fb-73b5-4047-8297-b9cdc5a261d0 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Updating inventory in ProviderTree for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 177, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 890.950669] env[61923]: DEBUG oslo_vmware.api [None req-eb5e01f1-148c-4608-96e7-93572113f79a tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Task: {'id': task-1377704, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.962037] env[61923]: DEBUG nova.network.neutron [None req-3192dd2d-7513-4a93-8a8c-c00e6e67889c tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] [instance: 263f0866-49af-4ab6-8132-d993642ebedb] Successfully updated port: 99825a7a-e38f-449b-8633-318da7f11ab5 {{(pid=61923) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 890.968756] env[61923]: DEBUG oslo_concurrency.lockutils [None req-02b5a1b8-a821-49de-bb63-1c89123e0e10 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Lock "interface-422d06d6-6932-46c3-bb25-841e6f65c028-b395b183-1c30-42de-929f-690ea10d9d75" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 8.158s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 891.362870] env[61923]: DEBUG oslo_vmware.api [None req-55b58374-2c35-46b9-a508-d4833264ecac tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Task: {'id': task-1377705, 'name': CopyVirtualDisk_Task} progress is 21%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.450938] env[61923]: DEBUG oslo_vmware.api [None req-eb5e01f1-148c-4608-96e7-93572113f79a tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Task: {'id': task-1377704, 'name': PowerOnVM_Task, 'duration_secs': 0.973191} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 891.451441] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-eb5e01f1-148c-4608-96e7-93572113f79a tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: 4c2c973a-5d23-4b8a-8b99-6b634e337054] Powered on the VM {{(pid=61923) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 891.451706] env[61923]: INFO nova.compute.manager [None req-eb5e01f1-148c-4608-96e7-93572113f79a tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: 4c2c973a-5d23-4b8a-8b99-6b634e337054] Took 9.68 seconds to spawn the instance on the hypervisor. [ 891.451910] env[61923]: DEBUG nova.compute.manager [None req-eb5e01f1-148c-4608-96e7-93572113f79a tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: 4c2c973a-5d23-4b8a-8b99-6b634e337054] Checking state {{(pid=61923) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 891.452738] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36a82132-31d4-41eb-89e5-ef24c49dfe73 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.456961] env[61923]: ERROR nova.scheduler.client.report [None req-6cea46fb-73b5-4047-8297-b9cdc5a261d0 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] [req-4e826a49-9d76-4314-be45-07f8e064e6f9] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 177, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID f81803f2-f7f8-4939-a757-a77d34a1d0a2. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-4e826a49-9d76-4314-be45-07f8e064e6f9"}]} [ 891.475995] env[61923]: DEBUG nova.scheduler.client.report [None req-6cea46fb-73b5-4047-8297-b9cdc5a261d0 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Refreshing inventories for resource provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 891.494083] env[61923]: DEBUG nova.scheduler.client.report [None req-6cea46fb-73b5-4047-8297-b9cdc5a261d0 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Updating ProviderTree inventory for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 891.494446] env[61923]: DEBUG nova.compute.provider_tree [None req-6cea46fb-73b5-4047-8297-b9cdc5a261d0 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Updating inventory in ProviderTree for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 891.513117] env[61923]: DEBUG nova.scheduler.client.report [None req-6cea46fb-73b5-4047-8297-b9cdc5a261d0 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Refreshing aggregate associations for resource provider f81803f2-f7f8-4939-a757-a77d34a1d0a2, aggregates: None {{(pid=61923) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 891.549903] env[61923]: DEBUG nova.scheduler.client.report [None req-6cea46fb-73b5-4047-8297-b9cdc5a261d0 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Refreshing trait associations for resource provider f81803f2-f7f8-4939-a757-a77d34a1d0a2, traits: COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NODE,HW_ARCH_X86_64,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_IMAGE_TYPE_ISO {{(pid=61923) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 891.827178] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d00216a-e693-4cb3-aabb-4a51acabd7c9 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.836967] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c89e6255-1adf-4416-b7ea-02cd0503e780 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.869846] env[61923]: DEBUG oslo_concurrency.lockutils [None req-11f7570e-7eca-4d2f-9bb4-0fd5b38bff3d tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] Acquiring lock "af3c317a-4007-4cea-a060-1e7dde5ce49e" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 891.870143] env[61923]: DEBUG oslo_concurrency.lockutils [None req-11f7570e-7eca-4d2f-9bb4-0fd5b38bff3d tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] Lock "af3c317a-4007-4cea-a060-1e7dde5ce49e" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 891.870402] env[61923]: INFO nova.compute.manager [None req-11f7570e-7eca-4d2f-9bb4-0fd5b38bff3d tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] [instance: af3c317a-4007-4cea-a060-1e7dde5ce49e] Attaching volume cddb3090-62b4-4639-850d-4eee7f2a2900 to /dev/sdb [ 891.876147] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb0ac0eb-369f-40be-add6-74d32d19af6c {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.887484] env[61923]: DEBUG oslo_vmware.api [None req-55b58374-2c35-46b9-a508-d4833264ecac tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Task: {'id': task-1377705, 'name': CopyVirtualDisk_Task} progress is 40%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.891054] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a7e11e1-5be9-45a5-b5ed-a94455332502 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.913466] env[61923]: DEBUG nova.compute.provider_tree [None req-6cea46fb-73b5-4047-8297-b9cdc5a261d0 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Updating inventory in ProviderTree for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 177, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 891.917499] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-954ecb41-8381-4d05-be23-1df86caadec0 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.927219] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b43aa268-2d08-435a-ad71-53acd57ba077 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.947218] env[61923]: DEBUG nova.virt.block_device [None req-11f7570e-7eca-4d2f-9bb4-0fd5b38bff3d tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] [instance: af3c317a-4007-4cea-a060-1e7dde5ce49e] Updating existing volume attachment record: 692d7ee9-1a6a-4e16-a8c0-09239f2a3eb1 {{(pid=61923) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 891.978035] env[61923]: INFO nova.compute.manager [None req-eb5e01f1-148c-4608-96e7-93572113f79a tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: 4c2c973a-5d23-4b8a-8b99-6b634e337054] Took 17.69 seconds to build instance. [ 892.084020] env[61923]: DEBUG nova.compute.manager [req-824b5acd-3bcc-4b6f-bd12-f7e84ccbdb11 req-47abc76a-1e53-4304-92ee-25656511c507 service nova] [instance: 263f0866-49af-4ab6-8132-d993642ebedb] Received event network-vif-plugged-99825a7a-e38f-449b-8633-318da7f11ab5 {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 892.084377] env[61923]: DEBUG oslo_concurrency.lockutils [req-824b5acd-3bcc-4b6f-bd12-f7e84ccbdb11 req-47abc76a-1e53-4304-92ee-25656511c507 service nova] Acquiring lock "263f0866-49af-4ab6-8132-d993642ebedb-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 892.084524] env[61923]: DEBUG oslo_concurrency.lockutils [req-824b5acd-3bcc-4b6f-bd12-f7e84ccbdb11 req-47abc76a-1e53-4304-92ee-25656511c507 service nova] Lock "263f0866-49af-4ab6-8132-d993642ebedb-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 892.084731] env[61923]: DEBUG oslo_concurrency.lockutils [req-824b5acd-3bcc-4b6f-bd12-f7e84ccbdb11 req-47abc76a-1e53-4304-92ee-25656511c507 service nova] Lock "263f0866-49af-4ab6-8132-d993642ebedb-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 892.084913] env[61923]: DEBUG nova.compute.manager [req-824b5acd-3bcc-4b6f-bd12-f7e84ccbdb11 req-47abc76a-1e53-4304-92ee-25656511c507 service nova] [instance: 263f0866-49af-4ab6-8132-d993642ebedb] No waiting events found dispatching network-vif-plugged-99825a7a-e38f-449b-8633-318da7f11ab5 {{(pid=61923) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 892.085098] env[61923]: WARNING nova.compute.manager [req-824b5acd-3bcc-4b6f-bd12-f7e84ccbdb11 req-47abc76a-1e53-4304-92ee-25656511c507 service nova] [instance: 263f0866-49af-4ab6-8132-d993642ebedb] Received unexpected event network-vif-plugged-99825a7a-e38f-449b-8633-318da7f11ab5 for instance with vm_state building and task_state spawning. [ 892.382071] env[61923]: DEBUG oslo_vmware.api [None req-55b58374-2c35-46b9-a508-d4833264ecac tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Task: {'id': task-1377705, 'name': CopyVirtualDisk_Task} progress is 63%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.464013] env[61923]: DEBUG nova.scheduler.client.report [None req-6cea46fb-73b5-4047-8297-b9cdc5a261d0 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Updated inventory for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 with generation 95 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 177, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 892.464013] env[61923]: DEBUG nova.compute.provider_tree [None req-6cea46fb-73b5-4047-8297-b9cdc5a261d0 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Updating resource provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 generation from 95 to 96 during operation: update_inventory {{(pid=61923) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 892.464013] env[61923]: DEBUG nova.compute.provider_tree [None req-6cea46fb-73b5-4047-8297-b9cdc5a261d0 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Updating inventory in ProviderTree for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 177, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 892.482863] env[61923]: DEBUG oslo_concurrency.lockutils [None req-eb5e01f1-148c-4608-96e7-93572113f79a tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Lock "4c2c973a-5d23-4b8a-8b99-6b634e337054" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 19.208s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 892.721811] env[61923]: DEBUG oslo_concurrency.lockutils [None req-214ded71-bc45-4cf6-9e81-e6c214aa77f2 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Acquiring lock "interface-422d06d6-6932-46c3-bb25-841e6f65c028-b395b183-1c30-42de-929f-690ea10d9d75" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 892.723143] env[61923]: DEBUG oslo_concurrency.lockutils [None req-214ded71-bc45-4cf6-9e81-e6c214aa77f2 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Lock "interface-422d06d6-6932-46c3-bb25-841e6f65c028-b395b183-1c30-42de-929f-690ea10d9d75" acquired by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: waited 0.001s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 892.883211] env[61923]: DEBUG oslo_vmware.api [None req-55b58374-2c35-46b9-a508-d4833264ecac tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Task: {'id': task-1377705, 'name': CopyVirtualDisk_Task} progress is 83%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.969494] env[61923]: DEBUG oslo_concurrency.lockutils [None req-6cea46fb-73b5-4047-8297-b9cdc5a261d0 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.350s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 892.970860] env[61923]: DEBUG nova.compute.manager [None req-6cea46fb-73b5-4047-8297-b9cdc5a261d0 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] [instance: bac71328-3796-4a65-aa41-ccda93bc7926] Start building networks asynchronously for instance. {{(pid=61923) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 892.976056] env[61923]: DEBUG oslo_concurrency.lockutils [None req-b7a2d28f-39dc-4a35-bb16-94846b77bc41 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.941s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 892.976056] env[61923]: INFO nova.compute.claims [None req-b7a2d28f-39dc-4a35-bb16-94846b77bc41 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] [instance: c98c96a5-d18e-4519-996a-949ded16225f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 893.098787] env[61923]: DEBUG oslo_concurrency.lockutils [None req-cb709b57-999a-41a8-b165-7029c0613fc5 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Acquiring lock "03c1e8ea-a322-46fa-9b07-e283080a7871" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 893.098999] env[61923]: DEBUG oslo_concurrency.lockutils [None req-cb709b57-999a-41a8-b165-7029c0613fc5 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Lock "03c1e8ea-a322-46fa-9b07-e283080a7871" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 893.225659] env[61923]: DEBUG oslo_concurrency.lockutils [None req-214ded71-bc45-4cf6-9e81-e6c214aa77f2 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Acquiring lock "422d06d6-6932-46c3-bb25-841e6f65c028" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 893.225900] env[61923]: DEBUG oslo_concurrency.lockutils [None req-214ded71-bc45-4cf6-9e81-e6c214aa77f2 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Acquired lock "422d06d6-6932-46c3-bb25-841e6f65c028" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 893.226801] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a4b2c01-09a4-4994-93c8-a2221874c61a {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.248175] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79d2cf6c-e7de-4afb-b8a8-bd621ac74cad {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.276187] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-214ded71-bc45-4cf6-9e81-e6c214aa77f2 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] [instance: 422d06d6-6932-46c3-bb25-841e6f65c028] Reconfiguring VM to detach interface {{(pid=61923) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1968}} [ 893.276532] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-27ae9f14-3b91-4a9a-b5bf-c98e444aead6 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.296404] env[61923]: DEBUG oslo_vmware.api [None req-214ded71-bc45-4cf6-9e81-e6c214aa77f2 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Waiting for the task: (returnval){ [ 893.296404] env[61923]: value = "task-1377709" [ 893.296404] env[61923]: _type = "Task" [ 893.296404] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 893.305505] env[61923]: DEBUG oslo_vmware.api [None req-214ded71-bc45-4cf6-9e81-e6c214aa77f2 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Task: {'id': task-1377709, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.320319] env[61923]: DEBUG nova.network.neutron [None req-3192dd2d-7513-4a93-8a8c-c00e6e67889c tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] [instance: 263f0866-49af-4ab6-8132-d993642ebedb] Successfully updated port: a16defc4-1374-44de-a7c1-e2d75f27e954 {{(pid=61923) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 893.380456] env[61923]: DEBUG oslo_vmware.api [None req-55b58374-2c35-46b9-a508-d4833264ecac tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Task: {'id': task-1377705, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.376437} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 893.380786] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-55b58374-2c35-46b9-a508-d4833264ecac tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/9ddad7e6-f1e4-49aa-b956-1b8d3273318d/9ddad7e6-f1e4-49aa-b956-1b8d3273318d.vmdk to [datastore2] 4cb3c397-ef1a-4267-a300-26da48206f3f/4cb3c397-ef1a-4267-a300-26da48206f3f.vmdk {{(pid=61923) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 893.381601] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71127822-f909-4b7a-bf45-75b10a258e99 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.404995] env[61923]: DEBUG nova.virt.vmwareapi.volumeops [None req-55b58374-2c35-46b9-a508-d4833264ecac tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] [instance: 4cb3c397-ef1a-4267-a300-26da48206f3f] Reconfiguring VM instance instance-00000050 to attach disk [datastore2] 4cb3c397-ef1a-4267-a300-26da48206f3f/4cb3c397-ef1a-4267-a300-26da48206f3f.vmdk or device None with type streamOptimized {{(pid=61923) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 893.405672] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4aad1ab0-7103-4d6a-9c69-795c44481ebe {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.426500] env[61923]: DEBUG oslo_vmware.api [None req-55b58374-2c35-46b9-a508-d4833264ecac tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Waiting for the task: (returnval){ [ 893.426500] env[61923]: value = "task-1377710" [ 893.426500] env[61923]: _type = "Task" [ 893.426500] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 893.435015] env[61923]: DEBUG oslo_vmware.api [None req-55b58374-2c35-46b9-a508-d4833264ecac tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Task: {'id': task-1377710, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.486736] env[61923]: DEBUG nova.compute.utils [None req-6cea46fb-73b5-4047-8297-b9cdc5a261d0 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Using /dev/sd instead of None {{(pid=61923) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 893.491323] env[61923]: DEBUG nova.compute.manager [None req-6cea46fb-73b5-4047-8297-b9cdc5a261d0 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] [instance: bac71328-3796-4a65-aa41-ccda93bc7926] Allocating IP information in the background. {{(pid=61923) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 893.491323] env[61923]: DEBUG nova.network.neutron [None req-6cea46fb-73b5-4047-8297-b9cdc5a261d0 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] [instance: bac71328-3796-4a65-aa41-ccda93bc7926] allocate_for_instance() {{(pid=61923) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 893.545099] env[61923]: DEBUG nova.policy [None req-6cea46fb-73b5-4047-8297-b9cdc5a261d0 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '03056b1a9c2143efa96ebb46ee1fda1a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0d24f4a47d1f4b79a7d999a4dacb99ba', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61923) authorize /opt/stack/nova/nova/policy.py:201}} [ 893.601569] env[61923]: DEBUG nova.compute.manager [None req-cb709b57-999a-41a8-b165-7029c0613fc5 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: 03c1e8ea-a322-46fa-9b07-e283080a7871] Starting instance... {{(pid=61923) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 893.817522] env[61923]: DEBUG oslo_vmware.api [None req-214ded71-bc45-4cf6-9e81-e6c214aa77f2 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Task: {'id': task-1377709, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.822974] env[61923]: DEBUG oslo_concurrency.lockutils [None req-3192dd2d-7513-4a93-8a8c-c00e6e67889c tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] Acquiring lock "refresh_cache-263f0866-49af-4ab6-8132-d993642ebedb" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 893.823424] env[61923]: DEBUG oslo_concurrency.lockutils [None req-3192dd2d-7513-4a93-8a8c-c00e6e67889c tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] Acquired lock "refresh_cache-263f0866-49af-4ab6-8132-d993642ebedb" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 893.823424] env[61923]: DEBUG nova.network.neutron [None req-3192dd2d-7513-4a93-8a8c-c00e6e67889c tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] [instance: 263f0866-49af-4ab6-8132-d993642ebedb] Building network info cache for instance {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 893.903306] env[61923]: DEBUG nova.network.neutron [None req-6cea46fb-73b5-4047-8297-b9cdc5a261d0 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] [instance: bac71328-3796-4a65-aa41-ccda93bc7926] Successfully created port: 923d9288-8294-4410-b6e7-529fefc2c6a7 {{(pid=61923) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 893.939911] env[61923]: DEBUG oslo_vmware.api [None req-55b58374-2c35-46b9-a508-d4833264ecac tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Task: {'id': task-1377710, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.995935] env[61923]: DEBUG nova.compute.manager [None req-6cea46fb-73b5-4047-8297-b9cdc5a261d0 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] [instance: bac71328-3796-4a65-aa41-ccda93bc7926] Start building block device mappings for instance. {{(pid=61923) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 894.126563] env[61923]: DEBUG oslo_concurrency.lockutils [None req-cb709b57-999a-41a8-b165-7029c0613fc5 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 894.199477] env[61923]: DEBUG nova.compute.manager [req-8b5791a1-8429-443c-99f8-8e8538fc7c2f req-78f93423-5ed3-4d85-b1e0-24b05804f3c8 service nova] [instance: 263f0866-49af-4ab6-8132-d993642ebedb] Received event network-changed-99825a7a-e38f-449b-8633-318da7f11ab5 {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 894.199670] env[61923]: DEBUG nova.compute.manager [req-8b5791a1-8429-443c-99f8-8e8538fc7c2f req-78f93423-5ed3-4d85-b1e0-24b05804f3c8 service nova] [instance: 263f0866-49af-4ab6-8132-d993642ebedb] Refreshing instance network info cache due to event network-changed-99825a7a-e38f-449b-8633-318da7f11ab5. {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 894.199898] env[61923]: DEBUG oslo_concurrency.lockutils [req-8b5791a1-8429-443c-99f8-8e8538fc7c2f req-78f93423-5ed3-4d85-b1e0-24b05804f3c8 service nova] Acquiring lock "refresh_cache-263f0866-49af-4ab6-8132-d993642ebedb" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 894.305896] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4772ad57-8bf8-40ac-a088-7c8f089dc8be {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.313166] env[61923]: DEBUG oslo_vmware.api [None req-214ded71-bc45-4cf6-9e81-e6c214aa77f2 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Task: {'id': task-1377709, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.318519] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fdbed099-59d9-4410-98a7-678864051754 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.353866] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-053d719c-7517-4436-8e9f-dcee26a33c27 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.362970] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd45a3fc-5ae1-4ca2-a3ec-d74de3454679 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.379491] env[61923]: DEBUG nova.compute.provider_tree [None req-b7a2d28f-39dc-4a35-bb16-94846b77bc41 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 894.391803] env[61923]: DEBUG nova.network.neutron [None req-3192dd2d-7513-4a93-8a8c-c00e6e67889c tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] [instance: 263f0866-49af-4ab6-8132-d993642ebedb] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 894.443457] env[61923]: DEBUG oslo_vmware.api [None req-55b58374-2c35-46b9-a508-d4833264ecac tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Task: {'id': task-1377710, 'name': ReconfigVM_Task, 'duration_secs': 0.750694} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 894.443457] env[61923]: DEBUG nova.virt.vmwareapi.volumeops [None req-55b58374-2c35-46b9-a508-d4833264ecac tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] [instance: 4cb3c397-ef1a-4267-a300-26da48206f3f] Reconfigured VM instance instance-00000050 to attach disk [datastore2] 4cb3c397-ef1a-4267-a300-26da48206f3f/4cb3c397-ef1a-4267-a300-26da48206f3f.vmdk or device None with type streamOptimized {{(pid=61923) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 894.443457] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f53c4d5f-29f7-4282-bd15-c9efff640caf {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.451673] env[61923]: DEBUG oslo_vmware.api [None req-55b58374-2c35-46b9-a508-d4833264ecac tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Waiting for the task: (returnval){ [ 894.451673] env[61923]: value = "task-1377712" [ 894.451673] env[61923]: _type = "Task" [ 894.451673] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 894.461544] env[61923]: DEBUG oslo_vmware.api [None req-55b58374-2c35-46b9-a508-d4833264ecac tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Task: {'id': task-1377712, 'name': Rename_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.698376] env[61923]: DEBUG nova.network.neutron [None req-3192dd2d-7513-4a93-8a8c-c00e6e67889c tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] [instance: 263f0866-49af-4ab6-8132-d993642ebedb] Updating instance_info_cache with network_info: [{"id": "99825a7a-e38f-449b-8633-318da7f11ab5", "address": "fa:16:3e:28:17:ff", "network": {"id": "83abfb75-61d1-4c48-9d06-077d2ae7c85b", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1481658242", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.143", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c304b5b5d4df485fbda55a570494a62f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fb68953b-dee5-4d9d-b47b-277336ba76dc", "external-id": "nsx-vlan-transportzone-168", "segmentation_id": 168, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap99825a7a-e3", "ovs_interfaceid": "99825a7a-e38f-449b-8633-318da7f11ab5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "a16defc4-1374-44de-a7c1-e2d75f27e954", "address": "fa:16:3e:a2:ea:09", "network": {"id": "6e4efac0-4973-49c2-bc09-71eb0339e4a3", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-558573136", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "c304b5b5d4df485fbda55a570494a62f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "713e54d5-283f-493d-b003-f13182deaf7b", "external-id": "cl2-zone-703", "segmentation_id": 703, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa16defc4-13", "ovs_interfaceid": "a16defc4-1374-44de-a7c1-e2d75f27e954", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 894.810854] env[61923]: DEBUG oslo_vmware.api [None req-214ded71-bc45-4cf6-9e81-e6c214aa77f2 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Task: {'id': task-1377709, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.883669] env[61923]: DEBUG nova.scheduler.client.report [None req-b7a2d28f-39dc-4a35-bb16-94846b77bc41 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 177, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 894.963027] env[61923]: DEBUG oslo_vmware.api [None req-55b58374-2c35-46b9-a508-d4833264ecac tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Task: {'id': task-1377712, 'name': Rename_Task, 'duration_secs': 0.270465} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 894.963319] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-55b58374-2c35-46b9-a508-d4833264ecac tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] [instance: 4cb3c397-ef1a-4267-a300-26da48206f3f] Powering on the VM {{(pid=61923) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 894.963573] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a33f9e9f-504b-4b4e-afaf-6d21f76fdb27 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.970922] env[61923]: DEBUG oslo_vmware.api [None req-55b58374-2c35-46b9-a508-d4833264ecac tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Waiting for the task: (returnval){ [ 894.970922] env[61923]: value = "task-1377713" [ 894.970922] env[61923]: _type = "Task" [ 894.970922] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 894.979426] env[61923]: DEBUG oslo_vmware.api [None req-55b58374-2c35-46b9-a508-d4833264ecac tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Task: {'id': task-1377713, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.010911] env[61923]: DEBUG nova.compute.manager [None req-6cea46fb-73b5-4047-8297-b9cdc5a261d0 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] [instance: bac71328-3796-4a65-aa41-ccda93bc7926] Start spawning the instance on the hypervisor. {{(pid=61923) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 895.036360] env[61923]: DEBUG nova.virt.hardware [None req-6cea46fb-73b5-4047-8297-b9cdc5a261d0 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-29T20:07:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-29T20:07:35Z,direct_url=,disk_format='vmdk',id=0f153f63-ae0a-45b1-b7f5-7f9b673c947f,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='4e7b5e3b3c3443c8bd5c70f8a15739f5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-29T20:07:36Z,virtual_size=,visibility=), allow threads: False {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 895.036623] env[61923]: DEBUG nova.virt.hardware [None req-6cea46fb-73b5-4047-8297-b9cdc5a261d0 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Flavor limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 895.036783] env[61923]: DEBUG nova.virt.hardware [None req-6cea46fb-73b5-4047-8297-b9cdc5a261d0 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Image limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 895.036965] env[61923]: DEBUG nova.virt.hardware [None req-6cea46fb-73b5-4047-8297-b9cdc5a261d0 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Flavor pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 895.037151] env[61923]: DEBUG nova.virt.hardware [None req-6cea46fb-73b5-4047-8297-b9cdc5a261d0 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Image pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 895.037286] env[61923]: DEBUG nova.virt.hardware [None req-6cea46fb-73b5-4047-8297-b9cdc5a261d0 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 895.037496] env[61923]: DEBUG nova.virt.hardware [None req-6cea46fb-73b5-4047-8297-b9cdc5a261d0 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 895.037655] env[61923]: DEBUG nova.virt.hardware [None req-6cea46fb-73b5-4047-8297-b9cdc5a261d0 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 895.037825] env[61923]: DEBUG nova.virt.hardware [None req-6cea46fb-73b5-4047-8297-b9cdc5a261d0 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Got 1 possible topologies {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 895.037992] env[61923]: DEBUG nova.virt.hardware [None req-6cea46fb-73b5-4047-8297-b9cdc5a261d0 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 895.038182] env[61923]: DEBUG nova.virt.hardware [None req-6cea46fb-73b5-4047-8297-b9cdc5a261d0 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 895.039039] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4bd03df2-da3a-47a4-a5fd-9a1f3dd31bd9 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.047682] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bbe19554-7e2e-4c49-a19d-0c596bb02333 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.201289] env[61923]: DEBUG oslo_concurrency.lockutils [None req-3192dd2d-7513-4a93-8a8c-c00e6e67889c tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] Releasing lock "refresh_cache-263f0866-49af-4ab6-8132-d993642ebedb" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 895.201789] env[61923]: DEBUG nova.compute.manager [None req-3192dd2d-7513-4a93-8a8c-c00e6e67889c tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] [instance: 263f0866-49af-4ab6-8132-d993642ebedb] Instance network_info: |[{"id": "99825a7a-e38f-449b-8633-318da7f11ab5", "address": "fa:16:3e:28:17:ff", "network": {"id": "83abfb75-61d1-4c48-9d06-077d2ae7c85b", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1481658242", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.143", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c304b5b5d4df485fbda55a570494a62f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fb68953b-dee5-4d9d-b47b-277336ba76dc", "external-id": "nsx-vlan-transportzone-168", "segmentation_id": 168, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap99825a7a-e3", "ovs_interfaceid": "99825a7a-e38f-449b-8633-318da7f11ab5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "a16defc4-1374-44de-a7c1-e2d75f27e954", "address": "fa:16:3e:a2:ea:09", "network": {"id": "6e4efac0-4973-49c2-bc09-71eb0339e4a3", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-558573136", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "c304b5b5d4df485fbda55a570494a62f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "713e54d5-283f-493d-b003-f13182deaf7b", "external-id": "cl2-zone-703", "segmentation_id": 703, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa16defc4-13", "ovs_interfaceid": "a16defc4-1374-44de-a7c1-e2d75f27e954", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61923) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 895.202089] env[61923]: DEBUG oslo_concurrency.lockutils [req-8b5791a1-8429-443c-99f8-8e8538fc7c2f req-78f93423-5ed3-4d85-b1e0-24b05804f3c8 service nova] Acquired lock "refresh_cache-263f0866-49af-4ab6-8132-d993642ebedb" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 895.202268] env[61923]: DEBUG nova.network.neutron [req-8b5791a1-8429-443c-99f8-8e8538fc7c2f req-78f93423-5ed3-4d85-b1e0-24b05804f3c8 service nova] [instance: 263f0866-49af-4ab6-8132-d993642ebedb] Refreshing network info cache for port 99825a7a-e38f-449b-8633-318da7f11ab5 {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 895.203527] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-3192dd2d-7513-4a93-8a8c-c00e6e67889c tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] [instance: 263f0866-49af-4ab6-8132-d993642ebedb] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:28:17:ff', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'fb68953b-dee5-4d9d-b47b-277336ba76dc', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '99825a7a-e38f-449b-8633-318da7f11ab5', 'vif_model': 'vmxnet3'}, {'network_name': 'br-int', 'mac_address': 'fa:16:3e:a2:ea:09', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '713e54d5-283f-493d-b003-f13182deaf7b', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'a16defc4-1374-44de-a7c1-e2d75f27e954', 'vif_model': 'vmxnet3'}] {{(pid=61923) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 895.220417] env[61923]: DEBUG oslo.service.loopingcall [None req-3192dd2d-7513-4a93-8a8c-c00e6e67889c tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61923) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 895.223790] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 263f0866-49af-4ab6-8132-d993642ebedb] Creating VM on the ESX host {{(pid=61923) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 895.224322] env[61923]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-6b1ebbd7-3a89-4244-836f-0b78fd00590d {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.251830] env[61923]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 895.251830] env[61923]: value = "task-1377714" [ 895.251830] env[61923]: _type = "Task" [ 895.251830] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 895.263096] env[61923]: DEBUG oslo_vmware.api [-] Task: {'id': task-1377714, 'name': CreateVM_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.311624] env[61923]: DEBUG oslo_vmware.api [None req-214ded71-bc45-4cf6-9e81-e6c214aa77f2 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Task: {'id': task-1377709, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.389345] env[61923]: DEBUG oslo_concurrency.lockutils [None req-b7a2d28f-39dc-4a35-bb16-94846b77bc41 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.415s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 895.390319] env[61923]: DEBUG nova.compute.manager [None req-b7a2d28f-39dc-4a35-bb16-94846b77bc41 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] [instance: c98c96a5-d18e-4519-996a-949ded16225f] Start building networks asynchronously for instance. {{(pid=61923) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 895.393120] env[61923]: DEBUG oslo_concurrency.lockutils [None req-e1b75445-6f25-4dcb-9846-d40076ac5b42 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.463s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 895.394352] env[61923]: INFO nova.compute.claims [None req-e1b75445-6f25-4dcb-9846-d40076ac5b42 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] [instance: 624e9207-7f24-4f45-8f52-0fc7a60c1e84] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 895.485364] env[61923]: DEBUG oslo_vmware.api [None req-55b58374-2c35-46b9-a508-d4833264ecac tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Task: {'id': task-1377713, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.613082] env[61923]: DEBUG nova.compute.manager [req-22a19f49-763d-400f-bb56-b48a88890814 req-c8e0ec46-ca77-402b-b2e6-b40dcc3acd9b service nova] [instance: bac71328-3796-4a65-aa41-ccda93bc7926] Received event network-vif-plugged-923d9288-8294-4410-b6e7-529fefc2c6a7 {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 895.613394] env[61923]: DEBUG oslo_concurrency.lockutils [req-22a19f49-763d-400f-bb56-b48a88890814 req-c8e0ec46-ca77-402b-b2e6-b40dcc3acd9b service nova] Acquiring lock "bac71328-3796-4a65-aa41-ccda93bc7926-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 895.613642] env[61923]: DEBUG oslo_concurrency.lockutils [req-22a19f49-763d-400f-bb56-b48a88890814 req-c8e0ec46-ca77-402b-b2e6-b40dcc3acd9b service nova] Lock "bac71328-3796-4a65-aa41-ccda93bc7926-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 895.613861] env[61923]: DEBUG oslo_concurrency.lockutils [req-22a19f49-763d-400f-bb56-b48a88890814 req-c8e0ec46-ca77-402b-b2e6-b40dcc3acd9b service nova] Lock "bac71328-3796-4a65-aa41-ccda93bc7926-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 895.614293] env[61923]: DEBUG nova.compute.manager [req-22a19f49-763d-400f-bb56-b48a88890814 req-c8e0ec46-ca77-402b-b2e6-b40dcc3acd9b service nova] [instance: bac71328-3796-4a65-aa41-ccda93bc7926] No waiting events found dispatching network-vif-plugged-923d9288-8294-4410-b6e7-529fefc2c6a7 {{(pid=61923) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 895.614763] env[61923]: WARNING nova.compute.manager [req-22a19f49-763d-400f-bb56-b48a88890814 req-c8e0ec46-ca77-402b-b2e6-b40dcc3acd9b service nova] [instance: bac71328-3796-4a65-aa41-ccda93bc7926] Received unexpected event network-vif-plugged-923d9288-8294-4410-b6e7-529fefc2c6a7 for instance with vm_state building and task_state spawning. [ 895.633140] env[61923]: DEBUG nova.network.neutron [req-8b5791a1-8429-443c-99f8-8e8538fc7c2f req-78f93423-5ed3-4d85-b1e0-24b05804f3c8 service nova] [instance: 263f0866-49af-4ab6-8132-d993642ebedb] Updated VIF entry in instance network info cache for port 99825a7a-e38f-449b-8633-318da7f11ab5. {{(pid=61923) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 895.633717] env[61923]: DEBUG nova.network.neutron [req-8b5791a1-8429-443c-99f8-8e8538fc7c2f req-78f93423-5ed3-4d85-b1e0-24b05804f3c8 service nova] [instance: 263f0866-49af-4ab6-8132-d993642ebedb] Updating instance_info_cache with network_info: [{"id": "99825a7a-e38f-449b-8633-318da7f11ab5", "address": "fa:16:3e:28:17:ff", "network": {"id": "83abfb75-61d1-4c48-9d06-077d2ae7c85b", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1481658242", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.143", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c304b5b5d4df485fbda55a570494a62f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fb68953b-dee5-4d9d-b47b-277336ba76dc", "external-id": "nsx-vlan-transportzone-168", "segmentation_id": 168, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap99825a7a-e3", "ovs_interfaceid": "99825a7a-e38f-449b-8633-318da7f11ab5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "a16defc4-1374-44de-a7c1-e2d75f27e954", "address": "fa:16:3e:a2:ea:09", "network": {"id": "6e4efac0-4973-49c2-bc09-71eb0339e4a3", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-558573136", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "c304b5b5d4df485fbda55a570494a62f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "713e54d5-283f-493d-b003-f13182deaf7b", "external-id": "cl2-zone-703", "segmentation_id": 703, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa16defc4-13", "ovs_interfaceid": "a16defc4-1374-44de-a7c1-e2d75f27e954", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 895.762809] env[61923]: DEBUG oslo_vmware.api [-] Task: {'id': task-1377714, 'name': CreateVM_Task, 'duration_secs': 0.462028} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 895.763063] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 263f0866-49af-4ab6-8132-d993642ebedb] Created VM on the ESX host {{(pid=61923) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 895.763811] env[61923]: DEBUG oslo_concurrency.lockutils [None req-3192dd2d-7513-4a93-8a8c-c00e6e67889c tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 895.763994] env[61923]: DEBUG oslo_concurrency.lockutils [None req-3192dd2d-7513-4a93-8a8c-c00e6e67889c tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 895.764341] env[61923]: DEBUG oslo_concurrency.lockutils [None req-3192dd2d-7513-4a93-8a8c-c00e6e67889c tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 895.764617] env[61923]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8098f520-9a65-4a98-b7c4-19574475d731 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.770475] env[61923]: DEBUG oslo_vmware.api [None req-3192dd2d-7513-4a93-8a8c-c00e6e67889c tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] Waiting for the task: (returnval){ [ 895.770475] env[61923]: value = "session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52de6185-8398-24a1-dd8a-d99b76a9d57c" [ 895.770475] env[61923]: _type = "Task" [ 895.770475] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 895.780130] env[61923]: DEBUG oslo_vmware.api [None req-3192dd2d-7513-4a93-8a8c-c00e6e67889c tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52de6185-8398-24a1-dd8a-d99b76a9d57c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.809772] env[61923]: DEBUG oslo_vmware.api [None req-214ded71-bc45-4cf6-9e81-e6c214aa77f2 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Task: {'id': task-1377709, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.899449] env[61923]: DEBUG nova.compute.utils [None req-b7a2d28f-39dc-4a35-bb16-94846b77bc41 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Using /dev/sd instead of None {{(pid=61923) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 895.901214] env[61923]: DEBUG nova.compute.manager [None req-b7a2d28f-39dc-4a35-bb16-94846b77bc41 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] [instance: c98c96a5-d18e-4519-996a-949ded16225f] Not allocating networking since 'none' was specified. {{(pid=61923) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 895.982356] env[61923]: DEBUG oslo_vmware.api [None req-55b58374-2c35-46b9-a508-d4833264ecac tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Task: {'id': task-1377713, 'name': PowerOnVM_Task, 'duration_secs': 0.852691} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 895.982746] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-55b58374-2c35-46b9-a508-d4833264ecac tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] [instance: 4cb3c397-ef1a-4267-a300-26da48206f3f] Powered on the VM {{(pid=61923) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 895.982746] env[61923]: INFO nova.compute.manager [None req-55b58374-2c35-46b9-a508-d4833264ecac tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] [instance: 4cb3c397-ef1a-4267-a300-26da48206f3f] Took 16.64 seconds to spawn the instance on the hypervisor. [ 895.982922] env[61923]: DEBUG nova.compute.manager [None req-55b58374-2c35-46b9-a508-d4833264ecac tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] [instance: 4cb3c397-ef1a-4267-a300-26da48206f3f] Checking state {{(pid=61923) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 895.983702] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d265dea-7b34-4bd8-8ee7-c2e49daca836 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.136480] env[61923]: DEBUG oslo_concurrency.lockutils [req-8b5791a1-8429-443c-99f8-8e8538fc7c2f req-78f93423-5ed3-4d85-b1e0-24b05804f3c8 service nova] Releasing lock "refresh_cache-263f0866-49af-4ab6-8132-d993642ebedb" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 896.136768] env[61923]: DEBUG nova.compute.manager [req-8b5791a1-8429-443c-99f8-8e8538fc7c2f req-78f93423-5ed3-4d85-b1e0-24b05804f3c8 service nova] [instance: 263f0866-49af-4ab6-8132-d993642ebedb] Received event network-vif-plugged-a16defc4-1374-44de-a7c1-e2d75f27e954 {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 896.136966] env[61923]: DEBUG oslo_concurrency.lockutils [req-8b5791a1-8429-443c-99f8-8e8538fc7c2f req-78f93423-5ed3-4d85-b1e0-24b05804f3c8 service nova] Acquiring lock "263f0866-49af-4ab6-8132-d993642ebedb-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 896.137231] env[61923]: DEBUG oslo_concurrency.lockutils [req-8b5791a1-8429-443c-99f8-8e8538fc7c2f req-78f93423-5ed3-4d85-b1e0-24b05804f3c8 service nova] Lock "263f0866-49af-4ab6-8132-d993642ebedb-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 896.137468] env[61923]: DEBUG oslo_concurrency.lockutils [req-8b5791a1-8429-443c-99f8-8e8538fc7c2f req-78f93423-5ed3-4d85-b1e0-24b05804f3c8 service nova] Lock "263f0866-49af-4ab6-8132-d993642ebedb-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 896.137694] env[61923]: DEBUG nova.compute.manager [req-8b5791a1-8429-443c-99f8-8e8538fc7c2f req-78f93423-5ed3-4d85-b1e0-24b05804f3c8 service nova] [instance: 263f0866-49af-4ab6-8132-d993642ebedb] No waiting events found dispatching network-vif-plugged-a16defc4-1374-44de-a7c1-e2d75f27e954 {{(pid=61923) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 896.137960] env[61923]: WARNING nova.compute.manager [req-8b5791a1-8429-443c-99f8-8e8538fc7c2f req-78f93423-5ed3-4d85-b1e0-24b05804f3c8 service nova] [instance: 263f0866-49af-4ab6-8132-d993642ebedb] Received unexpected event network-vif-plugged-a16defc4-1374-44de-a7c1-e2d75f27e954 for instance with vm_state building and task_state spawning. [ 896.138162] env[61923]: DEBUG nova.compute.manager [req-8b5791a1-8429-443c-99f8-8e8538fc7c2f req-78f93423-5ed3-4d85-b1e0-24b05804f3c8 service nova] [instance: 263f0866-49af-4ab6-8132-d993642ebedb] Received event network-changed-a16defc4-1374-44de-a7c1-e2d75f27e954 {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 896.138328] env[61923]: DEBUG nova.compute.manager [req-8b5791a1-8429-443c-99f8-8e8538fc7c2f req-78f93423-5ed3-4d85-b1e0-24b05804f3c8 service nova] [instance: 263f0866-49af-4ab6-8132-d993642ebedb] Refreshing instance network info cache due to event network-changed-a16defc4-1374-44de-a7c1-e2d75f27e954. {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 896.138525] env[61923]: DEBUG oslo_concurrency.lockutils [req-8b5791a1-8429-443c-99f8-8e8538fc7c2f req-78f93423-5ed3-4d85-b1e0-24b05804f3c8 service nova] Acquiring lock "refresh_cache-263f0866-49af-4ab6-8132-d993642ebedb" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 896.138666] env[61923]: DEBUG oslo_concurrency.lockutils [req-8b5791a1-8429-443c-99f8-8e8538fc7c2f req-78f93423-5ed3-4d85-b1e0-24b05804f3c8 service nova] Acquired lock "refresh_cache-263f0866-49af-4ab6-8132-d993642ebedb" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 896.138826] env[61923]: DEBUG nova.network.neutron [req-8b5791a1-8429-443c-99f8-8e8538fc7c2f req-78f93423-5ed3-4d85-b1e0-24b05804f3c8 service nova] [instance: 263f0866-49af-4ab6-8132-d993642ebedb] Refreshing network info cache for port a16defc4-1374-44de-a7c1-e2d75f27e954 {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 896.203113] env[61923]: DEBUG nova.network.neutron [None req-6cea46fb-73b5-4047-8297-b9cdc5a261d0 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] [instance: bac71328-3796-4a65-aa41-ccda93bc7926] Successfully updated port: 923d9288-8294-4410-b6e7-529fefc2c6a7 {{(pid=61923) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 896.226057] env[61923]: DEBUG nova.compute.manager [req-6e31c777-2bd0-470d-bc17-2f6134951858 req-fb38b12e-daa2-454e-9d5f-ef0848641294 service nova] [instance: bac71328-3796-4a65-aa41-ccda93bc7926] Received event network-changed-923d9288-8294-4410-b6e7-529fefc2c6a7 {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 896.226284] env[61923]: DEBUG nova.compute.manager [req-6e31c777-2bd0-470d-bc17-2f6134951858 req-fb38b12e-daa2-454e-9d5f-ef0848641294 service nova] [instance: bac71328-3796-4a65-aa41-ccda93bc7926] Refreshing instance network info cache due to event network-changed-923d9288-8294-4410-b6e7-529fefc2c6a7. {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 896.226509] env[61923]: DEBUG oslo_concurrency.lockutils [req-6e31c777-2bd0-470d-bc17-2f6134951858 req-fb38b12e-daa2-454e-9d5f-ef0848641294 service nova] Acquiring lock "refresh_cache-bac71328-3796-4a65-aa41-ccda93bc7926" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 896.226655] env[61923]: DEBUG oslo_concurrency.lockutils [req-6e31c777-2bd0-470d-bc17-2f6134951858 req-fb38b12e-daa2-454e-9d5f-ef0848641294 service nova] Acquired lock "refresh_cache-bac71328-3796-4a65-aa41-ccda93bc7926" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 896.226818] env[61923]: DEBUG nova.network.neutron [req-6e31c777-2bd0-470d-bc17-2f6134951858 req-fb38b12e-daa2-454e-9d5f-ef0848641294 service nova] [instance: bac71328-3796-4a65-aa41-ccda93bc7926] Refreshing network info cache for port 923d9288-8294-4410-b6e7-529fefc2c6a7 {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 896.281776] env[61923]: DEBUG oslo_vmware.api [None req-3192dd2d-7513-4a93-8a8c-c00e6e67889c tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52de6185-8398-24a1-dd8a-d99b76a9d57c, 'name': SearchDatastore_Task, 'duration_secs': 0.011131} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 896.282099] env[61923]: DEBUG oslo_concurrency.lockutils [None req-3192dd2d-7513-4a93-8a8c-c00e6e67889c tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 896.282338] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-3192dd2d-7513-4a93-8a8c-c00e6e67889c tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] [instance: 263f0866-49af-4ab6-8132-d993642ebedb] Processing image 0f153f63-ae0a-45b1-b7f5-7f9b673c947f {{(pid=61923) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 896.282577] env[61923]: DEBUG oslo_concurrency.lockutils [None req-3192dd2d-7513-4a93-8a8c-c00e6e67889c tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 896.282726] env[61923]: DEBUG oslo_concurrency.lockutils [None req-3192dd2d-7513-4a93-8a8c-c00e6e67889c tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 896.282905] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-3192dd2d-7513-4a93-8a8c-c00e6e67889c tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61923) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 896.283201] env[61923]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-2f3feacc-dce2-4ee6-8974-03a8124abc6c {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.292416] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-3192dd2d-7513-4a93-8a8c-c00e6e67889c tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61923) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 896.292594] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-3192dd2d-7513-4a93-8a8c-c00e6e67889c tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61923) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 896.293333] env[61923]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ae8860e2-4825-4500-89a2-ed8c1040f833 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.298918] env[61923]: DEBUG oslo_vmware.api [None req-3192dd2d-7513-4a93-8a8c-c00e6e67889c tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] Waiting for the task: (returnval){ [ 896.298918] env[61923]: value = "session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]527eb2de-af4e-a987-ac27-c109dc1b9e6b" [ 896.298918] env[61923]: _type = "Task" [ 896.298918] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 896.312272] env[61923]: DEBUG oslo_vmware.api [None req-214ded71-bc45-4cf6-9e81-e6c214aa77f2 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Task: {'id': task-1377709, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.315751] env[61923]: DEBUG oslo_vmware.api [None req-3192dd2d-7513-4a93-8a8c-c00e6e67889c tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]527eb2de-af4e-a987-ac27-c109dc1b9e6b, 'name': SearchDatastore_Task, 'duration_secs': 0.009291} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 896.316440] env[61923]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f240f427-0be7-43ae-a9a0-ccdcf52e2fa9 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.321875] env[61923]: DEBUG oslo_vmware.api [None req-3192dd2d-7513-4a93-8a8c-c00e6e67889c tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] Waiting for the task: (returnval){ [ 896.321875] env[61923]: value = "session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]525b768b-79e2-6af7-43f4-577f55b5964e" [ 896.321875] env[61923]: _type = "Task" [ 896.321875] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 897.042386] env[61923]: DEBUG nova.compute.manager [None req-b7a2d28f-39dc-4a35-bb16-94846b77bc41 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] [instance: c98c96a5-d18e-4519-996a-949ded16225f] Start building block device mappings for instance. {{(pid=61923) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 897.054990] env[61923]: DEBUG oslo_concurrency.lockutils [None req-6cea46fb-73b5-4047-8297-b9cdc5a261d0 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Acquiring lock "refresh_cache-bac71328-3796-4a65-aa41-ccda93bc7926" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 897.057412] env[61923]: DEBUG oslo_vmware.api [None req-3192dd2d-7513-4a93-8a8c-c00e6e67889c tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]525b768b-79e2-6af7-43f4-577f55b5964e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.057613] env[61923]: WARNING oslo_vmware.common.loopingcall [None req-3192dd2d-7513-4a93-8a8c-c00e6e67889c tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] task run outlasted interval by 0.23551299999999997 sec [ 897.060373] env[61923]: DEBUG nova.virt.vmwareapi.volumeops [None req-11f7570e-7eca-4d2f-9bb4-0fd5b38bff3d tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] [instance: af3c317a-4007-4cea-a060-1e7dde5ce49e] Volume attach. Driver type: vmdk {{(pid=61923) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 897.060593] env[61923]: DEBUG nova.virt.vmwareapi.volumeops [None req-11f7570e-7eca-4d2f-9bb4-0fd5b38bff3d tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] [instance: af3c317a-4007-4cea-a060-1e7dde5ce49e] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-292706', 'volume_id': 'cddb3090-62b4-4639-850d-4eee7f2a2900', 'name': 'volume-cddb3090-62b4-4639-850d-4eee7f2a2900', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'af3c317a-4007-4cea-a060-1e7dde5ce49e', 'attached_at': '', 'detached_at': '', 'volume_id': 'cddb3090-62b4-4639-850d-4eee7f2a2900', 'serial': 'cddb3090-62b4-4639-850d-4eee7f2a2900'} {{(pid=61923) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 897.067337] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e73729fd-a2bc-44e6-8eba-af948260f0e4 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.070604] env[61923]: INFO nova.compute.manager [None req-55b58374-2c35-46b9-a508-d4833264ecac tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] [instance: 4cb3c397-ef1a-4267-a300-26da48206f3f] Took 23.42 seconds to build instance. [ 897.094052] env[61923]: DEBUG oslo_vmware.api [None req-214ded71-bc45-4cf6-9e81-e6c214aa77f2 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Task: {'id': task-1377709, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.100671] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db453e74-ee9c-4afe-9654-472937c62a5a {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.103702] env[61923]: DEBUG oslo_vmware.api [None req-3192dd2d-7513-4a93-8a8c-c00e6e67889c tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]525b768b-79e2-6af7-43f4-577f55b5964e, 'name': SearchDatastore_Task, 'duration_secs': 0.00983} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 897.104062] env[61923]: DEBUG oslo_concurrency.lockutils [None req-3192dd2d-7513-4a93-8a8c-c00e6e67889c tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 897.104348] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-3192dd2d-7513-4a93-8a8c-c00e6e67889c tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk to [datastore1] 263f0866-49af-4ab6-8132-d993642ebedb/263f0866-49af-4ab6-8132-d993642ebedb.vmdk {{(pid=61923) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 897.104987] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-6a483ec3-7290-48bd-b1f6-23b62c93749b {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.128175] env[61923]: DEBUG nova.virt.vmwareapi.volumeops [None req-11f7570e-7eca-4d2f-9bb4-0fd5b38bff3d tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] [instance: af3c317a-4007-4cea-a060-1e7dde5ce49e] Reconfiguring VM instance instance-00000047 to attach disk [datastore1] volume-cddb3090-62b4-4639-850d-4eee7f2a2900/volume-cddb3090-62b4-4639-850d-4eee7f2a2900.vmdk or device None with type thin {{(pid=61923) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 897.133766] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4946acf2-8d4e-4089-9341-462789497238 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.147772] env[61923]: DEBUG oslo_vmware.api [None req-3192dd2d-7513-4a93-8a8c-c00e6e67889c tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] Waiting for the task: (returnval){ [ 897.147772] env[61923]: value = "task-1377715" [ 897.147772] env[61923]: _type = "Task" [ 897.147772] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 897.148887] env[61923]: DEBUG nova.network.neutron [req-6e31c777-2bd0-470d-bc17-2f6134951858 req-fb38b12e-daa2-454e-9d5f-ef0848641294 service nova] [instance: bac71328-3796-4a65-aa41-ccda93bc7926] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 897.161386] env[61923]: DEBUG oslo_vmware.api [None req-11f7570e-7eca-4d2f-9bb4-0fd5b38bff3d tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] Waiting for the task: (returnval){ [ 897.161386] env[61923]: value = "task-1377716" [ 897.161386] env[61923]: _type = "Task" [ 897.161386] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 897.181531] env[61923]: DEBUG oslo_vmware.api [None req-3192dd2d-7513-4a93-8a8c-c00e6e67889c tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] Task: {'id': task-1377715, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.203369] env[61923]: DEBUG oslo_vmware.api [None req-11f7570e-7eca-4d2f-9bb4-0fd5b38bff3d tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] Task: {'id': task-1377716, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.356200] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11b6d5eb-251a-49f8-8170-85af41f44599 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.365143] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8f3829b-0d7c-481d-a711-02bac73f0859 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.410991] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19458bb6-5231-48eb-8ff0-ab9b705b56f7 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.420523] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f380d7c3-8f74-4f58-9eb2-d14aedb64334 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.440042] env[61923]: DEBUG nova.compute.provider_tree [None req-e1b75445-6f25-4dcb-9846-d40076ac5b42 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 897.572198] env[61923]: DEBUG oslo_concurrency.lockutils [None req-7225546b-1247-4af1-9195-074d155e3aeb tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Acquiring lock "4cb3c397-ef1a-4267-a300-26da48206f3f" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 897.578333] env[61923]: DEBUG oslo_concurrency.lockutils [None req-55b58374-2c35-46b9-a508-d4833264ecac tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Lock "4cb3c397-ef1a-4267-a300-26da48206f3f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 24.943s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 897.578333] env[61923]: DEBUG oslo_vmware.api [None req-214ded71-bc45-4cf6-9e81-e6c214aa77f2 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Task: {'id': task-1377709, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.578333] env[61923]: DEBUG oslo_concurrency.lockutils [None req-7225546b-1247-4af1-9195-074d155e3aeb tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Lock "4cb3c397-ef1a-4267-a300-26da48206f3f" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.006s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 897.578333] env[61923]: DEBUG oslo_concurrency.lockutils [None req-7225546b-1247-4af1-9195-074d155e3aeb tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Acquiring lock "4cb3c397-ef1a-4267-a300-26da48206f3f-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 897.578642] env[61923]: DEBUG oslo_concurrency.lockutils [None req-7225546b-1247-4af1-9195-074d155e3aeb tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Lock "4cb3c397-ef1a-4267-a300-26da48206f3f-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 897.578871] env[61923]: DEBUG oslo_concurrency.lockutils [None req-7225546b-1247-4af1-9195-074d155e3aeb tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Lock "4cb3c397-ef1a-4267-a300-26da48206f3f-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 897.581330] env[61923]: INFO nova.compute.manager [None req-7225546b-1247-4af1-9195-074d155e3aeb tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] [instance: 4cb3c397-ef1a-4267-a300-26da48206f3f] Terminating instance [ 897.583696] env[61923]: DEBUG nova.compute.manager [None req-7225546b-1247-4af1-9195-074d155e3aeb tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] [instance: 4cb3c397-ef1a-4267-a300-26da48206f3f] Start destroying the instance on the hypervisor. {{(pid=61923) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 897.583931] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-7225546b-1247-4af1-9195-074d155e3aeb tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] [instance: 4cb3c397-ef1a-4267-a300-26da48206f3f] Destroying instance {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 897.584896] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50249dce-1372-4a9e-9a1c-97bc9a2aa7a4 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.595281] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-7225546b-1247-4af1-9195-074d155e3aeb tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] [instance: 4cb3c397-ef1a-4267-a300-26da48206f3f] Powering off the VM {{(pid=61923) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 897.595585] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e4f9ff71-2965-42e9-860a-ec47cabd0516 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.603363] env[61923]: DEBUG nova.network.neutron [req-6e31c777-2bd0-470d-bc17-2f6134951858 req-fb38b12e-daa2-454e-9d5f-ef0848641294 service nova] [instance: bac71328-3796-4a65-aa41-ccda93bc7926] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 897.605012] env[61923]: DEBUG oslo_vmware.api [None req-7225546b-1247-4af1-9195-074d155e3aeb tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Waiting for the task: (returnval){ [ 897.605012] env[61923]: value = "task-1377717" [ 897.605012] env[61923]: _type = "Task" [ 897.605012] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 897.617107] env[61923]: DEBUG oslo_vmware.api [None req-7225546b-1247-4af1-9195-074d155e3aeb tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Task: {'id': task-1377717, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.663989] env[61923]: DEBUG oslo_vmware.api [None req-3192dd2d-7513-4a93-8a8c-c00e6e67889c tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] Task: {'id': task-1377715, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.676882] env[61923]: DEBUG oslo_vmware.api [None req-11f7570e-7eca-4d2f-9bb4-0fd5b38bff3d tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] Task: {'id': task-1377716, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.832244] env[61923]: DEBUG nova.network.neutron [req-8b5791a1-8429-443c-99f8-8e8538fc7c2f req-78f93423-5ed3-4d85-b1e0-24b05804f3c8 service nova] [instance: 263f0866-49af-4ab6-8132-d993642ebedb] Updated VIF entry in instance network info cache for port a16defc4-1374-44de-a7c1-e2d75f27e954. {{(pid=61923) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 897.833013] env[61923]: DEBUG nova.network.neutron [req-8b5791a1-8429-443c-99f8-8e8538fc7c2f req-78f93423-5ed3-4d85-b1e0-24b05804f3c8 service nova] [instance: 263f0866-49af-4ab6-8132-d993642ebedb] Updating instance_info_cache with network_info: [{"id": "99825a7a-e38f-449b-8633-318da7f11ab5", "address": "fa:16:3e:28:17:ff", "network": {"id": "83abfb75-61d1-4c48-9d06-077d2ae7c85b", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1481658242", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.143", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c304b5b5d4df485fbda55a570494a62f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fb68953b-dee5-4d9d-b47b-277336ba76dc", "external-id": "nsx-vlan-transportzone-168", "segmentation_id": 168, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap99825a7a-e3", "ovs_interfaceid": "99825a7a-e38f-449b-8633-318da7f11ab5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "a16defc4-1374-44de-a7c1-e2d75f27e954", "address": "fa:16:3e:a2:ea:09", "network": {"id": "6e4efac0-4973-49c2-bc09-71eb0339e4a3", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-558573136", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "c304b5b5d4df485fbda55a570494a62f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "713e54d5-283f-493d-b003-f13182deaf7b", "external-id": "cl2-zone-703", "segmentation_id": 703, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa16defc4-13", "ovs_interfaceid": "a16defc4-1374-44de-a7c1-e2d75f27e954", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 897.943952] env[61923]: DEBUG nova.scheduler.client.report [None req-e1b75445-6f25-4dcb-9846-d40076ac5b42 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 177, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 898.071205] env[61923]: DEBUG nova.compute.manager [None req-b7a2d28f-39dc-4a35-bb16-94846b77bc41 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] [instance: c98c96a5-d18e-4519-996a-949ded16225f] Start spawning the instance on the hypervisor. {{(pid=61923) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 898.073158] env[61923]: DEBUG oslo_vmware.api [None req-214ded71-bc45-4cf6-9e81-e6c214aa77f2 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Task: {'id': task-1377709, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.099256] env[61923]: DEBUG nova.virt.hardware [None req-b7a2d28f-39dc-4a35-bb16-94846b77bc41 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-29T20:07:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-29T20:07:35Z,direct_url=,disk_format='vmdk',id=0f153f63-ae0a-45b1-b7f5-7f9b673c947f,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='4e7b5e3b3c3443c8bd5c70f8a15739f5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-29T20:07:36Z,virtual_size=,visibility=), allow threads: False {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 898.099614] env[61923]: DEBUG nova.virt.hardware [None req-b7a2d28f-39dc-4a35-bb16-94846b77bc41 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Flavor limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 898.099819] env[61923]: DEBUG nova.virt.hardware [None req-b7a2d28f-39dc-4a35-bb16-94846b77bc41 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Image limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 898.100029] env[61923]: DEBUG nova.virt.hardware [None req-b7a2d28f-39dc-4a35-bb16-94846b77bc41 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Flavor pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 898.100424] env[61923]: DEBUG nova.virt.hardware [None req-b7a2d28f-39dc-4a35-bb16-94846b77bc41 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Image pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 898.100582] env[61923]: DEBUG nova.virt.hardware [None req-b7a2d28f-39dc-4a35-bb16-94846b77bc41 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 898.100798] env[61923]: DEBUG nova.virt.hardware [None req-b7a2d28f-39dc-4a35-bb16-94846b77bc41 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 898.100959] env[61923]: DEBUG nova.virt.hardware [None req-b7a2d28f-39dc-4a35-bb16-94846b77bc41 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 898.101142] env[61923]: DEBUG nova.virt.hardware [None req-b7a2d28f-39dc-4a35-bb16-94846b77bc41 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Got 1 possible topologies {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 898.101307] env[61923]: DEBUG nova.virt.hardware [None req-b7a2d28f-39dc-4a35-bb16-94846b77bc41 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 898.101479] env[61923]: DEBUG nova.virt.hardware [None req-b7a2d28f-39dc-4a35-bb16-94846b77bc41 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 898.102338] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13eb8a57-17b7-44b1-b873-6f48e213cbc8 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.106496] env[61923]: DEBUG oslo_concurrency.lockutils [req-6e31c777-2bd0-470d-bc17-2f6134951858 req-fb38b12e-daa2-454e-9d5f-ef0848641294 service nova] Releasing lock "refresh_cache-bac71328-3796-4a65-aa41-ccda93bc7926" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 898.106884] env[61923]: DEBUG oslo_concurrency.lockutils [None req-6cea46fb-73b5-4047-8297-b9cdc5a261d0 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Acquired lock "refresh_cache-bac71328-3796-4a65-aa41-ccda93bc7926" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 898.107079] env[61923]: DEBUG nova.network.neutron [None req-6cea46fb-73b5-4047-8297-b9cdc5a261d0 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] [instance: bac71328-3796-4a65-aa41-ccda93bc7926] Building network info cache for instance {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 898.114665] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1379b99-9a24-4aa9-9432-3d6ec57ccdbe {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.121867] env[61923]: DEBUG oslo_vmware.api [None req-7225546b-1247-4af1-9195-074d155e3aeb tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Task: {'id': task-1377717, 'name': PowerOffVM_Task, 'duration_secs': 0.202387} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 898.122449] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-7225546b-1247-4af1-9195-074d155e3aeb tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] [instance: 4cb3c397-ef1a-4267-a300-26da48206f3f] Powered off the VM {{(pid=61923) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 898.122664] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-7225546b-1247-4af1-9195-074d155e3aeb tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] [instance: 4cb3c397-ef1a-4267-a300-26da48206f3f] Unregistering the VM {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 898.123180] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-ffad5299-3e9f-4b50-b310-0fd52723eb1a {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.132507] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-b7a2d28f-39dc-4a35-bb16-94846b77bc41 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] [instance: c98c96a5-d18e-4519-996a-949ded16225f] Instance VIF info [] {{(pid=61923) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 898.137967] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-b7a2d28f-39dc-4a35-bb16-94846b77bc41 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Creating folder: Project (08fb8d0735c34ecab7b4811a1e3970f5). Parent ref: group-v292629. {{(pid=61923) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 898.138731] env[61923]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-11f69ca0-90ca-436f-b10b-f14f1d989312 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.150626] env[61923]: INFO nova.virt.vmwareapi.vm_util [None req-b7a2d28f-39dc-4a35-bb16-94846b77bc41 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Created folder: Project (08fb8d0735c34ecab7b4811a1e3970f5) in parent group-v292629. [ 898.150839] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-b7a2d28f-39dc-4a35-bb16-94846b77bc41 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Creating folder: Instances. Parent ref: group-v292708. {{(pid=61923) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 898.151093] env[61923]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-1e463cdf-1afa-4bb5-9c61-d570fa60b83f {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.161430] env[61923]: DEBUG oslo_vmware.api [None req-3192dd2d-7513-4a93-8a8c-c00e6e67889c tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] Task: {'id': task-1377715, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.68394} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 898.161692] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-3192dd2d-7513-4a93-8a8c-c00e6e67889c tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk to [datastore1] 263f0866-49af-4ab6-8132-d993642ebedb/263f0866-49af-4ab6-8132-d993642ebedb.vmdk {{(pid=61923) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 898.161907] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-3192dd2d-7513-4a93-8a8c-c00e6e67889c tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] [instance: 263f0866-49af-4ab6-8132-d993642ebedb] Extending root virtual disk to 1048576 {{(pid=61923) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 898.162165] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-3ff6fdd2-d39d-45c0-bfc8-227e7bbde246 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.165196] env[61923]: INFO nova.virt.vmwareapi.vm_util [None req-b7a2d28f-39dc-4a35-bb16-94846b77bc41 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Created folder: Instances in parent group-v292708. [ 898.165425] env[61923]: DEBUG oslo.service.loopingcall [None req-b7a2d28f-39dc-4a35-bb16-94846b77bc41 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61923) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 898.166227] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c98c96a5-d18e-4519-996a-949ded16225f] Creating VM on the ESX host {{(pid=61923) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 898.169237] env[61923]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-498df660-4176-454a-beff-d3f0c70b2a26 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.183321] env[61923]: DEBUG oslo_vmware.api [None req-3192dd2d-7513-4a93-8a8c-c00e6e67889c tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] Waiting for the task: (returnval){ [ 898.183321] env[61923]: value = "task-1377721" [ 898.183321] env[61923]: _type = "Task" [ 898.183321] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 898.192279] env[61923]: DEBUG oslo_vmware.api [None req-11f7570e-7eca-4d2f-9bb4-0fd5b38bff3d tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] Task: {'id': task-1377716, 'name': ReconfigVM_Task, 'duration_secs': 0.844181} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 898.192926] env[61923]: DEBUG nova.virt.vmwareapi.volumeops [None req-11f7570e-7eca-4d2f-9bb4-0fd5b38bff3d tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] [instance: af3c317a-4007-4cea-a060-1e7dde5ce49e] Reconfigured VM instance instance-00000047 to attach disk [datastore1] volume-cddb3090-62b4-4639-850d-4eee7f2a2900/volume-cddb3090-62b4-4639-850d-4eee7f2a2900.vmdk or device None with type thin {{(pid=61923) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 898.201372] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c2cb1de5-8961-4ccb-a83d-2c81a2a3b1cb {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.212989] env[61923]: DEBUG oslo_vmware.api [None req-3192dd2d-7513-4a93-8a8c-c00e6e67889c tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] Task: {'id': task-1377721, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.215798] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-7225546b-1247-4af1-9195-074d155e3aeb tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] [instance: 4cb3c397-ef1a-4267-a300-26da48206f3f] Unregistered the VM {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 898.216093] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-7225546b-1247-4af1-9195-074d155e3aeb tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] [instance: 4cb3c397-ef1a-4267-a300-26da48206f3f] Deleting contents of the VM from datastore datastore2 {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 898.216384] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-7225546b-1247-4af1-9195-074d155e3aeb tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Deleting the datastore file [datastore2] 4cb3c397-ef1a-4267-a300-26da48206f3f {{(pid=61923) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 898.216646] env[61923]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 898.216646] env[61923]: value = "task-1377722" [ 898.216646] env[61923]: _type = "Task" [ 898.216646] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 898.217243] env[61923]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-08c78513-15db-4d45-b109-c36c7be0e031 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.224098] env[61923]: DEBUG oslo_vmware.api [None req-11f7570e-7eca-4d2f-9bb4-0fd5b38bff3d tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] Waiting for the task: (returnval){ [ 898.224098] env[61923]: value = "task-1377723" [ 898.224098] env[61923]: _type = "Task" [ 898.224098] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 898.234068] env[61923]: DEBUG oslo_vmware.api [-] Task: {'id': task-1377722, 'name': CreateVM_Task} progress is 6%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.238063] env[61923]: DEBUG oslo_vmware.api [None req-7225546b-1247-4af1-9195-074d155e3aeb tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Waiting for the task: (returnval){ [ 898.238063] env[61923]: value = "task-1377724" [ 898.238063] env[61923]: _type = "Task" [ 898.238063] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 898.238378] env[61923]: DEBUG oslo_vmware.api [None req-11f7570e-7eca-4d2f-9bb4-0fd5b38bff3d tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] Task: {'id': task-1377723, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.248233] env[61923]: DEBUG oslo_vmware.api [None req-7225546b-1247-4af1-9195-074d155e3aeb tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Task: {'id': task-1377724, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.336724] env[61923]: DEBUG oslo_concurrency.lockutils [req-8b5791a1-8429-443c-99f8-8e8538fc7c2f req-78f93423-5ed3-4d85-b1e0-24b05804f3c8 service nova] Releasing lock "refresh_cache-263f0866-49af-4ab6-8132-d993642ebedb" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 898.450043] env[61923]: DEBUG oslo_concurrency.lockutils [None req-e1b75445-6f25-4dcb-9846-d40076ac5b42 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.057s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 898.450679] env[61923]: DEBUG nova.compute.manager [None req-e1b75445-6f25-4dcb-9846-d40076ac5b42 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] [instance: 624e9207-7f24-4f45-8f52-0fc7a60c1e84] Start building networks asynchronously for instance. {{(pid=61923) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 898.453599] env[61923]: DEBUG oslo_concurrency.lockutils [None req-cb709b57-999a-41a8-b165-7029c0613fc5 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.327s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 898.455175] env[61923]: INFO nova.compute.claims [None req-cb709b57-999a-41a8-b165-7029c0613fc5 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: 03c1e8ea-a322-46fa-9b07-e283080a7871] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 898.569763] env[61923]: DEBUG oslo_vmware.api [None req-214ded71-bc45-4cf6-9e81-e6c214aa77f2 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Task: {'id': task-1377709, 'name': ReconfigVM_Task} progress is 18%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.658055] env[61923]: DEBUG nova.network.neutron [None req-6cea46fb-73b5-4047-8297-b9cdc5a261d0 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] [instance: bac71328-3796-4a65-aa41-ccda93bc7926] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 898.694480] env[61923]: DEBUG oslo_vmware.api [None req-3192dd2d-7513-4a93-8a8c-c00e6e67889c tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] Task: {'id': task-1377721, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.067864} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 898.695222] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-3192dd2d-7513-4a93-8a8c-c00e6e67889c tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] [instance: 263f0866-49af-4ab6-8132-d993642ebedb] Extended root virtual disk {{(pid=61923) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 898.695654] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06549cf3-d08d-4838-bb6f-b23b7a291c30 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.721535] env[61923]: DEBUG nova.virt.vmwareapi.volumeops [None req-3192dd2d-7513-4a93-8a8c-c00e6e67889c tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] [instance: 263f0866-49af-4ab6-8132-d993642ebedb] Reconfiguring VM instance instance-00000052 to attach disk [datastore1] 263f0866-49af-4ab6-8132-d993642ebedb/263f0866-49af-4ab6-8132-d993642ebedb.vmdk or device None with type sparse {{(pid=61923) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 898.721879] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-021154bb-5c80-40ac-8b83-7c1dff86fc48 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.751190] env[61923]: DEBUG oslo_vmware.api [-] Task: {'id': task-1377722, 'name': CreateVM_Task, 'duration_secs': 0.386156} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 898.756966] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c98c96a5-d18e-4519-996a-949ded16225f] Created VM on the ESX host {{(pid=61923) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 898.757508] env[61923]: DEBUG oslo_vmware.api [None req-3192dd2d-7513-4a93-8a8c-c00e6e67889c tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] Waiting for the task: (returnval){ [ 898.757508] env[61923]: value = "task-1377725" [ 898.757508] env[61923]: _type = "Task" [ 898.757508] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 898.757669] env[61923]: DEBUG oslo_vmware.api [None req-7225546b-1247-4af1-9195-074d155e3aeb tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Task: {'id': task-1377724, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.259028} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 898.758133] env[61923]: DEBUG oslo_vmware.api [None req-11f7570e-7eca-4d2f-9bb4-0fd5b38bff3d tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] Task: {'id': task-1377723, 'name': ReconfigVM_Task, 'duration_secs': 0.189113} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 898.758526] env[61923]: DEBUG oslo_concurrency.lockutils [None req-b7a2d28f-39dc-4a35-bb16-94846b77bc41 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 898.758653] env[61923]: DEBUG oslo_concurrency.lockutils [None req-b7a2d28f-39dc-4a35-bb16-94846b77bc41 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 898.758978] env[61923]: DEBUG oslo_concurrency.lockutils [None req-b7a2d28f-39dc-4a35-bb16-94846b77bc41 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 898.759314] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-7225546b-1247-4af1-9195-074d155e3aeb tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Deleted the datastore file {{(pid=61923) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 898.759628] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-7225546b-1247-4af1-9195-074d155e3aeb tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] [instance: 4cb3c397-ef1a-4267-a300-26da48206f3f] Deleted contents of the VM from datastore datastore2 {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 898.759699] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-7225546b-1247-4af1-9195-074d155e3aeb tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] [instance: 4cb3c397-ef1a-4267-a300-26da48206f3f] Instance destroyed {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 898.759829] env[61923]: INFO nova.compute.manager [None req-7225546b-1247-4af1-9195-074d155e3aeb tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] [instance: 4cb3c397-ef1a-4267-a300-26da48206f3f] Took 1.18 seconds to destroy the instance on the hypervisor. [ 898.760060] env[61923]: DEBUG oslo.service.loopingcall [None req-7225546b-1247-4af1-9195-074d155e3aeb tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61923) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 898.760301] env[61923]: DEBUG nova.virt.vmwareapi.volumeops [None req-11f7570e-7eca-4d2f-9bb4-0fd5b38bff3d tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] [instance: af3c317a-4007-4cea-a060-1e7dde5ce49e] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-292706', 'volume_id': 'cddb3090-62b4-4639-850d-4eee7f2a2900', 'name': 'volume-cddb3090-62b4-4639-850d-4eee7f2a2900', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'af3c317a-4007-4cea-a060-1e7dde5ce49e', 'attached_at': '', 'detached_at': '', 'volume_id': 'cddb3090-62b4-4639-850d-4eee7f2a2900', 'serial': 'cddb3090-62b4-4639-850d-4eee7f2a2900'} {{(pid=61923) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 898.761963] env[61923]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a5c89241-4131-4f4d-9875-4cddcb92e8ab {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.769503] env[61923]: DEBUG nova.compute.manager [-] [instance: 4cb3c397-ef1a-4267-a300-26da48206f3f] Deallocating network for instance {{(pid=61923) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 898.769503] env[61923]: DEBUG nova.network.neutron [-] [instance: 4cb3c397-ef1a-4267-a300-26da48206f3f] deallocate_for_instance() {{(pid=61923) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 898.778797] env[61923]: DEBUG oslo_vmware.api [None req-b7a2d28f-39dc-4a35-bb16-94846b77bc41 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Waiting for the task: (returnval){ [ 898.778797] env[61923]: value = "session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52f3b59b-82cb-7759-2526-791f158c16b2" [ 898.778797] env[61923]: _type = "Task" [ 898.778797] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 898.779211] env[61923]: DEBUG oslo_vmware.api [None req-3192dd2d-7513-4a93-8a8c-c00e6e67889c tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] Task: {'id': task-1377725, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.788722] env[61923]: DEBUG oslo_vmware.api [None req-b7a2d28f-39dc-4a35-bb16-94846b77bc41 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52f3b59b-82cb-7759-2526-791f158c16b2, 'name': SearchDatastore_Task, 'duration_secs': 0.013058} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 898.789026] env[61923]: DEBUG oslo_concurrency.lockutils [None req-b7a2d28f-39dc-4a35-bb16-94846b77bc41 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 898.790461] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-b7a2d28f-39dc-4a35-bb16-94846b77bc41 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] [instance: c98c96a5-d18e-4519-996a-949ded16225f] Processing image 0f153f63-ae0a-45b1-b7f5-7f9b673c947f {{(pid=61923) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 898.790461] env[61923]: DEBUG oslo_concurrency.lockutils [None req-b7a2d28f-39dc-4a35-bb16-94846b77bc41 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 898.790461] env[61923]: DEBUG oslo_concurrency.lockutils [None req-b7a2d28f-39dc-4a35-bb16-94846b77bc41 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 898.790461] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-b7a2d28f-39dc-4a35-bb16-94846b77bc41 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61923) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 898.790461] env[61923]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f797f7aa-6bc6-46cf-adbf-3037a6b84383 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.803022] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-b7a2d28f-39dc-4a35-bb16-94846b77bc41 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61923) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 898.803022] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-b7a2d28f-39dc-4a35-bb16-94846b77bc41 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61923) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 898.803022] env[61923]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2f5b05ed-4089-4c62-bac0-c2fb9e4df855 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.806107] env[61923]: DEBUG oslo_vmware.api [None req-b7a2d28f-39dc-4a35-bb16-94846b77bc41 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Waiting for the task: (returnval){ [ 898.806107] env[61923]: value = "session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]524f1b78-d696-b278-f5bd-539efb41f6b2" [ 898.806107] env[61923]: _type = "Task" [ 898.806107] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 898.814785] env[61923]: DEBUG oslo_vmware.api [None req-b7a2d28f-39dc-4a35-bb16-94846b77bc41 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]524f1b78-d696-b278-f5bd-539efb41f6b2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.959799] env[61923]: DEBUG nova.compute.utils [None req-e1b75445-6f25-4dcb-9846-d40076ac5b42 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Using /dev/sd instead of None {{(pid=61923) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 898.964206] env[61923]: DEBUG nova.compute.manager [None req-e1b75445-6f25-4dcb-9846-d40076ac5b42 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] [instance: 624e9207-7f24-4f45-8f52-0fc7a60c1e84] Not allocating networking since 'none' was specified. {{(pid=61923) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 898.994476] env[61923]: DEBUG nova.network.neutron [None req-6cea46fb-73b5-4047-8297-b9cdc5a261d0 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] [instance: bac71328-3796-4a65-aa41-ccda93bc7926] Updating instance_info_cache with network_info: [{"id": "923d9288-8294-4410-b6e7-529fefc2c6a7", "address": "fa:16:3e:1a:ec:9f", "network": {"id": "07fc148e-c3ba-4e53-bec2-36500ec38add", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1374533437-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0d24f4a47d1f4b79a7d999a4dacb99ba", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b2907cce-d529-4809-af05-d29397bed211", "external-id": "nsx-vlan-transportzone-427", "segmentation_id": 427, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap923d9288-82", "ovs_interfaceid": "923d9288-8294-4410-b6e7-529fefc2c6a7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 899.072426] env[61923]: DEBUG oslo_vmware.api [None req-214ded71-bc45-4cf6-9e81-e6c214aa77f2 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Task: {'id': task-1377709, 'name': ReconfigVM_Task, 'duration_secs': 5.758204} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 899.072738] env[61923]: DEBUG oslo_concurrency.lockutils [None req-214ded71-bc45-4cf6-9e81-e6c214aa77f2 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Releasing lock "422d06d6-6932-46c3-bb25-841e6f65c028" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 899.072931] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-214ded71-bc45-4cf6-9e81-e6c214aa77f2 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] [instance: 422d06d6-6932-46c3-bb25-841e6f65c028] Reconfigured VM to detach interface {{(pid=61923) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1978}} [ 899.195398] env[61923]: DEBUG oslo_vmware.rw_handles [None req-4482c5e6-0d59-468c-9d9f-cee661c7eaed tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52c8888b-9c22-85ef-0fcf-b5c96df0cb46/disk-0.vmdk. {{(pid=61923) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 899.196572] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-670e2578-c291-4ebb-948a-d253665fb36c {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.203827] env[61923]: DEBUG oslo_vmware.rw_handles [None req-4482c5e6-0d59-468c-9d9f-cee661c7eaed tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52c8888b-9c22-85ef-0fcf-b5c96df0cb46/disk-0.vmdk is in state: ready. {{(pid=61923) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 899.204064] env[61923]: ERROR oslo_vmware.rw_handles [None req-4482c5e6-0d59-468c-9d9f-cee661c7eaed tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52c8888b-9c22-85ef-0fcf-b5c96df0cb46/disk-0.vmdk due to incomplete transfer. [ 899.204344] env[61923]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-184babcc-29c3-45d1-bfc8-33203073ede4 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.212303] env[61923]: DEBUG oslo_vmware.rw_handles [None req-4482c5e6-0d59-468c-9d9f-cee661c7eaed tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52c8888b-9c22-85ef-0fcf-b5c96df0cb46/disk-0.vmdk. {{(pid=61923) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 899.212540] env[61923]: DEBUG nova.virt.vmwareapi.images [None req-4482c5e6-0d59-468c-9d9f-cee661c7eaed tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] [instance: 2719569b-8572-4199-8158-7bb367d17dc5] Uploaded image b1a1622a-32c0-45e0-b073-5adb9b6d5da6 to the Glance image server {{(pid=61923) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 899.215837] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-4482c5e6-0d59-468c-9d9f-cee661c7eaed tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] [instance: 2719569b-8572-4199-8158-7bb367d17dc5] Destroying the VM {{(pid=61923) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 899.216168] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-2aef1253-4c7d-46d6-ac78-ed5c0664b3de {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.227037] env[61923]: DEBUG oslo_vmware.api [None req-4482c5e6-0d59-468c-9d9f-cee661c7eaed tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Waiting for the task: (returnval){ [ 899.227037] env[61923]: value = "task-1377726" [ 899.227037] env[61923]: _type = "Task" [ 899.227037] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 899.235538] env[61923]: DEBUG oslo_vmware.api [None req-4482c5e6-0d59-468c-9d9f-cee661c7eaed tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Task: {'id': task-1377726, 'name': Destroy_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.273763] env[61923]: DEBUG oslo_vmware.api [None req-3192dd2d-7513-4a93-8a8c-c00e6e67889c tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] Task: {'id': task-1377725, 'name': ReconfigVM_Task, 'duration_secs': 0.342038} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 899.274873] env[61923]: DEBUG nova.virt.vmwareapi.volumeops [None req-3192dd2d-7513-4a93-8a8c-c00e6e67889c tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] [instance: 263f0866-49af-4ab6-8132-d993642ebedb] Reconfigured VM instance instance-00000052 to attach disk [datastore1] 263f0866-49af-4ab6-8132-d993642ebedb/263f0866-49af-4ab6-8132-d993642ebedb.vmdk or device None with type sparse {{(pid=61923) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 899.276121] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-8fb80265-6038-4d07-9d2e-9ab7b30c3213 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.284466] env[61923]: DEBUG oslo_vmware.api [None req-3192dd2d-7513-4a93-8a8c-c00e6e67889c tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] Waiting for the task: (returnval){ [ 899.284466] env[61923]: value = "task-1377727" [ 899.284466] env[61923]: _type = "Task" [ 899.284466] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 899.294487] env[61923]: DEBUG nova.compute.manager [req-9ef35019-a38b-4df4-a8d4-39bffdbc89f0 req-13847a4b-4d99-499d-8912-079542e5ca40 service nova] [instance: 4cb3c397-ef1a-4267-a300-26da48206f3f] Received event network-vif-deleted-83e705c4-d432-4fdf-8486-d31b46e1e6a5 {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 899.294694] env[61923]: INFO nova.compute.manager [req-9ef35019-a38b-4df4-a8d4-39bffdbc89f0 req-13847a4b-4d99-499d-8912-079542e5ca40 service nova] [instance: 4cb3c397-ef1a-4267-a300-26da48206f3f] Neutron deleted interface 83e705c4-d432-4fdf-8486-d31b46e1e6a5; detaching it from the instance and deleting it from the info cache [ 899.294868] env[61923]: DEBUG nova.network.neutron [req-9ef35019-a38b-4df4-a8d4-39bffdbc89f0 req-13847a4b-4d99-499d-8912-079542e5ca40 service nova] [instance: 4cb3c397-ef1a-4267-a300-26da48206f3f] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 899.303395] env[61923]: DEBUG oslo_vmware.api [None req-3192dd2d-7513-4a93-8a8c-c00e6e67889c tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] Task: {'id': task-1377727, 'name': Rename_Task} progress is 10%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.319120] env[61923]: DEBUG oslo_vmware.api [None req-b7a2d28f-39dc-4a35-bb16-94846b77bc41 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]524f1b78-d696-b278-f5bd-539efb41f6b2, 'name': SearchDatastore_Task, 'duration_secs': 0.00977} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 899.322127] env[61923]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9712e29a-6d29-4956-b34a-51c1397a0704 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.325984] env[61923]: DEBUG oslo_vmware.api [None req-b7a2d28f-39dc-4a35-bb16-94846b77bc41 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Waiting for the task: (returnval){ [ 899.325984] env[61923]: value = "session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52e2a9e8-0e2a-75e1-319c-a9c5e204aeb3" [ 899.325984] env[61923]: _type = "Task" [ 899.325984] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 899.335054] env[61923]: DEBUG oslo_vmware.api [None req-b7a2d28f-39dc-4a35-bb16-94846b77bc41 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52e2a9e8-0e2a-75e1-319c-a9c5e204aeb3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.468567] env[61923]: DEBUG nova.compute.manager [None req-e1b75445-6f25-4dcb-9846-d40076ac5b42 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] [instance: 624e9207-7f24-4f45-8f52-0fc7a60c1e84] Start building block device mappings for instance. {{(pid=61923) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 899.498635] env[61923]: DEBUG oslo_concurrency.lockutils [None req-6cea46fb-73b5-4047-8297-b9cdc5a261d0 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Releasing lock "refresh_cache-bac71328-3796-4a65-aa41-ccda93bc7926" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 899.498959] env[61923]: DEBUG nova.compute.manager [None req-6cea46fb-73b5-4047-8297-b9cdc5a261d0 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] [instance: bac71328-3796-4a65-aa41-ccda93bc7926] Instance network_info: |[{"id": "923d9288-8294-4410-b6e7-529fefc2c6a7", "address": "fa:16:3e:1a:ec:9f", "network": {"id": "07fc148e-c3ba-4e53-bec2-36500ec38add", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1374533437-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0d24f4a47d1f4b79a7d999a4dacb99ba", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b2907cce-d529-4809-af05-d29397bed211", "external-id": "nsx-vlan-transportzone-427", "segmentation_id": 427, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap923d9288-82", "ovs_interfaceid": "923d9288-8294-4410-b6e7-529fefc2c6a7", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61923) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 899.499726] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-6cea46fb-73b5-4047-8297-b9cdc5a261d0 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] [instance: bac71328-3796-4a65-aa41-ccda93bc7926] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:1a:ec:9f', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'b2907cce-d529-4809-af05-d29397bed211', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '923d9288-8294-4410-b6e7-529fefc2c6a7', 'vif_model': 'vmxnet3'}] {{(pid=61923) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 899.508247] env[61923]: DEBUG oslo.service.loopingcall [None req-6cea46fb-73b5-4047-8297-b9cdc5a261d0 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61923) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 899.508516] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: bac71328-3796-4a65-aa41-ccda93bc7926] Creating VM on the ESX host {{(pid=61923) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 899.508765] env[61923]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-59ab43d1-5ef6-4270-a107-b0f2aee99442 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.536392] env[61923]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 899.536392] env[61923]: value = "task-1377728" [ 899.536392] env[61923]: _type = "Task" [ 899.536392] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 899.546963] env[61923]: DEBUG oslo_vmware.api [-] Task: {'id': task-1377728, 'name': CreateVM_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.731789] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-292ac863-9bec-4cdd-aaed-f934d2440d3e {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.742540] env[61923]: DEBUG oslo_vmware.api [None req-4482c5e6-0d59-468c-9d9f-cee661c7eaed tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Task: {'id': task-1377726, 'name': Destroy_Task} progress is 33%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.743200] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d798182e-00f2-48bc-95d8-e5917389839f {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.777969] env[61923]: DEBUG nova.network.neutron [-] [instance: 4cb3c397-ef1a-4267-a300-26da48206f3f] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 899.782058] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ceb44909-0834-4533-8c13-2ea7fe0a470f {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.794182] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0eebf609-0886-4ec4-a609-f2f1d1e8ff24 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.802755] env[61923]: DEBUG oslo_vmware.api [None req-3192dd2d-7513-4a93-8a8c-c00e6e67889c tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] Task: {'id': task-1377727, 'name': Rename_Task, 'duration_secs': 0.229109} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 899.802755] env[61923]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-04f95656-29ea-4caf-819e-06610d77957c {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.805267] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-3192dd2d-7513-4a93-8a8c-c00e6e67889c tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] [instance: 263f0866-49af-4ab6-8132-d993642ebedb] Powering on the VM {{(pid=61923) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 899.805848] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-9ae64615-589d-4317-8a3f-abd790ded0f2 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.818197] env[61923]: DEBUG nova.objects.instance [None req-11f7570e-7eca-4d2f-9bb4-0fd5b38bff3d tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] Lazy-loading 'flavor' on Instance uuid af3c317a-4007-4cea-a060-1e7dde5ce49e {{(pid=61923) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 899.822423] env[61923]: DEBUG nova.compute.provider_tree [None req-cb709b57-999a-41a8-b165-7029c0613fc5 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 899.828281] env[61923]: DEBUG oslo_vmware.api [None req-3192dd2d-7513-4a93-8a8c-c00e6e67889c tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] Waiting for the task: (returnval){ [ 899.828281] env[61923]: value = "task-1377729" [ 899.828281] env[61923]: _type = "Task" [ 899.828281] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 899.834725] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dfea8997-6540-41fe-aa90-83bdd3781173 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.859995] env[61923]: DEBUG oslo_vmware.api [None req-3192dd2d-7513-4a93-8a8c-c00e6e67889c tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] Task: {'id': task-1377729, 'name': PowerOnVM_Task} progress is 33%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.863771] env[61923]: DEBUG oslo_vmware.api [None req-b7a2d28f-39dc-4a35-bb16-94846b77bc41 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52e2a9e8-0e2a-75e1-319c-a9c5e204aeb3, 'name': SearchDatastore_Task, 'duration_secs': 0.010805} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 899.864410] env[61923]: DEBUG oslo_concurrency.lockutils [None req-b7a2d28f-39dc-4a35-bb16-94846b77bc41 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 899.864795] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-b7a2d28f-39dc-4a35-bb16-94846b77bc41 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk to [datastore2] c98c96a5-d18e-4519-996a-949ded16225f/c98c96a5-d18e-4519-996a-949ded16225f.vmdk {{(pid=61923) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 899.864960] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-0cb3fa93-0194-436c-a813-eda68e5d7902 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.878121] env[61923]: DEBUG nova.compute.manager [req-9ef35019-a38b-4df4-a8d4-39bffdbc89f0 req-13847a4b-4d99-499d-8912-079542e5ca40 service nova] [instance: 4cb3c397-ef1a-4267-a300-26da48206f3f] Detach interface failed, port_id=83e705c4-d432-4fdf-8486-d31b46e1e6a5, reason: Instance 4cb3c397-ef1a-4267-a300-26da48206f3f could not be found. {{(pid=61923) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 899.885166] env[61923]: DEBUG oslo_vmware.api [None req-b7a2d28f-39dc-4a35-bb16-94846b77bc41 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Waiting for the task: (returnval){ [ 899.885166] env[61923]: value = "task-1377730" [ 899.885166] env[61923]: _type = "Task" [ 899.885166] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 899.895589] env[61923]: DEBUG oslo_vmware.api [None req-b7a2d28f-39dc-4a35-bb16-94846b77bc41 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Task: {'id': task-1377730, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.048615] env[61923]: DEBUG oslo_vmware.api [-] Task: {'id': task-1377728, 'name': CreateVM_Task} progress is 99%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.236738] env[61923]: DEBUG oslo_vmware.api [None req-4482c5e6-0d59-468c-9d9f-cee661c7eaed tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Task: {'id': task-1377726, 'name': Destroy_Task, 'duration_secs': 0.562117} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 900.237051] env[61923]: INFO nova.virt.vmwareapi.vm_util [None req-4482c5e6-0d59-468c-9d9f-cee661c7eaed tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] [instance: 2719569b-8572-4199-8158-7bb367d17dc5] Destroyed the VM [ 900.237431] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-4482c5e6-0d59-468c-9d9f-cee661c7eaed tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] [instance: 2719569b-8572-4199-8158-7bb367d17dc5] Deleting Snapshot of the VM instance {{(pid=61923) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 900.237758] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-ece48a3e-6037-41d9-a682-1c57380444b8 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.246364] env[61923]: DEBUG oslo_vmware.api [None req-4482c5e6-0d59-468c-9d9f-cee661c7eaed tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Waiting for the task: (returnval){ [ 900.246364] env[61923]: value = "task-1377731" [ 900.246364] env[61923]: _type = "Task" [ 900.246364] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 900.255519] env[61923]: DEBUG oslo_vmware.api [None req-4482c5e6-0d59-468c-9d9f-cee661c7eaed tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Task: {'id': task-1377731, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.284605] env[61923]: INFO nova.compute.manager [-] [instance: 4cb3c397-ef1a-4267-a300-26da48206f3f] Took 1.52 seconds to deallocate network for instance. [ 900.329787] env[61923]: DEBUG nova.scheduler.client.report [None req-cb709b57-999a-41a8-b165-7029c0613fc5 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 177, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 900.333262] env[61923]: DEBUG oslo_concurrency.lockutils [None req-11f7570e-7eca-4d2f-9bb4-0fd5b38bff3d tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] Lock "af3c317a-4007-4cea-a060-1e7dde5ce49e" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 8.463s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 900.344537] env[61923]: DEBUG oslo_vmware.api [None req-3192dd2d-7513-4a93-8a8c-c00e6e67889c tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] Task: {'id': task-1377729, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.395254] env[61923]: DEBUG oslo_vmware.api [None req-b7a2d28f-39dc-4a35-bb16-94846b77bc41 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Task: {'id': task-1377730, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.479258] env[61923]: DEBUG nova.compute.manager [None req-e1b75445-6f25-4dcb-9846-d40076ac5b42 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] [instance: 624e9207-7f24-4f45-8f52-0fc7a60c1e84] Start spawning the instance on the hypervisor. {{(pid=61923) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 900.492549] env[61923]: DEBUG oslo_concurrency.lockutils [None req-f4c3a587-afc9-4326-a453-4024294c25b5 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] Acquiring lock "af3c317a-4007-4cea-a060-1e7dde5ce49e" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 900.492746] env[61923]: DEBUG oslo_concurrency.lockutils [None req-f4c3a587-afc9-4326-a453-4024294c25b5 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] Lock "af3c317a-4007-4cea-a060-1e7dde5ce49e" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.001s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 900.507606] env[61923]: DEBUG nova.virt.hardware [None req-e1b75445-6f25-4dcb-9846-d40076ac5b42 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-29T20:07:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-29T20:07:35Z,direct_url=,disk_format='vmdk',id=0f153f63-ae0a-45b1-b7f5-7f9b673c947f,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='4e7b5e3b3c3443c8bd5c70f8a15739f5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-29T20:07:36Z,virtual_size=,visibility=), allow threads: False {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 900.507809] env[61923]: DEBUG nova.virt.hardware [None req-e1b75445-6f25-4dcb-9846-d40076ac5b42 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Flavor limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 900.507934] env[61923]: DEBUG nova.virt.hardware [None req-e1b75445-6f25-4dcb-9846-d40076ac5b42 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Image limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 900.508139] env[61923]: DEBUG nova.virt.hardware [None req-e1b75445-6f25-4dcb-9846-d40076ac5b42 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Flavor pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 900.508291] env[61923]: DEBUG nova.virt.hardware [None req-e1b75445-6f25-4dcb-9846-d40076ac5b42 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Image pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 900.508441] env[61923]: DEBUG nova.virt.hardware [None req-e1b75445-6f25-4dcb-9846-d40076ac5b42 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 900.508659] env[61923]: DEBUG nova.virt.hardware [None req-e1b75445-6f25-4dcb-9846-d40076ac5b42 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 900.508818] env[61923]: DEBUG nova.virt.hardware [None req-e1b75445-6f25-4dcb-9846-d40076ac5b42 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 900.509122] env[61923]: DEBUG nova.virt.hardware [None req-e1b75445-6f25-4dcb-9846-d40076ac5b42 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Got 1 possible topologies {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 900.509385] env[61923]: DEBUG nova.virt.hardware [None req-e1b75445-6f25-4dcb-9846-d40076ac5b42 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 900.509608] env[61923]: DEBUG nova.virt.hardware [None req-e1b75445-6f25-4dcb-9846-d40076ac5b42 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 900.510921] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-713b4e6a-3abc-44f2-b54c-ca5a50aefc7a {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.523020] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19ccaba3-50d2-4f3a-aed3-04917759a4a0 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.538765] env[61923]: DEBUG nova.compute.manager [req-bfc4d372-54f0-4569-8793-5a5c05278256 req-f57d85d8-2133-49ba-b9c2-262a73e962b5 service nova] [instance: 422d06d6-6932-46c3-bb25-841e6f65c028] Received event network-changed-97caa1b8-427c-46b0-be7a-d59b932900c3 {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 900.539156] env[61923]: DEBUG nova.compute.manager [req-bfc4d372-54f0-4569-8793-5a5c05278256 req-f57d85d8-2133-49ba-b9c2-262a73e962b5 service nova] [instance: 422d06d6-6932-46c3-bb25-841e6f65c028] Refreshing instance network info cache due to event network-changed-97caa1b8-427c-46b0-be7a-d59b932900c3. {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 900.539475] env[61923]: DEBUG oslo_concurrency.lockutils [req-bfc4d372-54f0-4569-8793-5a5c05278256 req-f57d85d8-2133-49ba-b9c2-262a73e962b5 service nova] Acquiring lock "refresh_cache-422d06d6-6932-46c3-bb25-841e6f65c028" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 900.539722] env[61923]: DEBUG oslo_concurrency.lockutils [req-bfc4d372-54f0-4569-8793-5a5c05278256 req-f57d85d8-2133-49ba-b9c2-262a73e962b5 service nova] Acquired lock "refresh_cache-422d06d6-6932-46c3-bb25-841e6f65c028" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 900.540010] env[61923]: DEBUG nova.network.neutron [req-bfc4d372-54f0-4569-8793-5a5c05278256 req-f57d85d8-2133-49ba-b9c2-262a73e962b5 service nova] [instance: 422d06d6-6932-46c3-bb25-841e6f65c028] Refreshing network info cache for port 97caa1b8-427c-46b0-be7a-d59b932900c3 {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 900.563252] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-e1b75445-6f25-4dcb-9846-d40076ac5b42 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] [instance: 624e9207-7f24-4f45-8f52-0fc7a60c1e84] Instance VIF info [] {{(pid=61923) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 900.573334] env[61923]: DEBUG oslo.service.loopingcall [None req-e1b75445-6f25-4dcb-9846-d40076ac5b42 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61923) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 900.575104] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 624e9207-7f24-4f45-8f52-0fc7a60c1e84] Creating VM on the ESX host {{(pid=61923) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 900.575489] env[61923]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-afd79462-1557-488d-a0ce-d262381c97c7 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.603123] env[61923]: DEBUG oslo_vmware.api [-] Task: {'id': task-1377728, 'name': CreateVM_Task} progress is 99%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.612668] env[61923]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 900.612668] env[61923]: value = "task-1377732" [ 900.612668] env[61923]: _type = "Task" [ 900.612668] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 900.613518] env[61923]: DEBUG oslo_concurrency.lockutils [None req-214ded71-bc45-4cf6-9e81-e6c214aa77f2 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Acquiring lock "refresh_cache-422d06d6-6932-46c3-bb25-841e6f65c028" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 900.628074] env[61923]: DEBUG oslo_vmware.api [-] Task: {'id': task-1377732, 'name': CreateVM_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.761503] env[61923]: DEBUG oslo_vmware.api [None req-4482c5e6-0d59-468c-9d9f-cee661c7eaed tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Task: {'id': task-1377731, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.791389] env[61923]: DEBUG oslo_concurrency.lockutils [None req-7225546b-1247-4af1-9195-074d155e3aeb tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 900.838944] env[61923]: DEBUG oslo_concurrency.lockutils [None req-cb709b57-999a-41a8-b165-7029c0613fc5 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.385s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 900.839571] env[61923]: DEBUG nova.compute.manager [None req-cb709b57-999a-41a8-b165-7029c0613fc5 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: 03c1e8ea-a322-46fa-9b07-e283080a7871] Start building networks asynchronously for instance. {{(pid=61923) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 900.846427] env[61923]: DEBUG oslo_concurrency.lockutils [None req-7225546b-1247-4af1-9195-074d155e3aeb tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.055s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 900.846723] env[61923]: DEBUG nova.objects.instance [None req-7225546b-1247-4af1-9195-074d155e3aeb tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Lazy-loading 'resources' on Instance uuid 4cb3c397-ef1a-4267-a300-26da48206f3f {{(pid=61923) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 900.847989] env[61923]: DEBUG oslo_vmware.api [None req-3192dd2d-7513-4a93-8a8c-c00e6e67889c tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] Task: {'id': task-1377729, 'name': PowerOnVM_Task} progress is 94%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.901683] env[61923]: DEBUG oslo_vmware.api [None req-b7a2d28f-39dc-4a35-bb16-94846b77bc41 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Task: {'id': task-1377730, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.996800] env[61923]: INFO nova.compute.manager [None req-f4c3a587-afc9-4326-a453-4024294c25b5 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] [instance: af3c317a-4007-4cea-a060-1e7dde5ce49e] Detaching volume cddb3090-62b4-4639-850d-4eee7f2a2900 [ 901.043540] env[61923]: INFO nova.virt.block_device [None req-f4c3a587-afc9-4326-a453-4024294c25b5 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] [instance: af3c317a-4007-4cea-a060-1e7dde5ce49e] Attempting to driver detach volume cddb3090-62b4-4639-850d-4eee7f2a2900 from mountpoint /dev/sdb [ 901.044027] env[61923]: DEBUG nova.virt.vmwareapi.volumeops [None req-f4c3a587-afc9-4326-a453-4024294c25b5 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] [instance: af3c317a-4007-4cea-a060-1e7dde5ce49e] Volume detach. Driver type: vmdk {{(pid=61923) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 901.044027] env[61923]: DEBUG nova.virt.vmwareapi.volumeops [None req-f4c3a587-afc9-4326-a453-4024294c25b5 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] [instance: af3c317a-4007-4cea-a060-1e7dde5ce49e] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-292706', 'volume_id': 'cddb3090-62b4-4639-850d-4eee7f2a2900', 'name': 'volume-cddb3090-62b4-4639-850d-4eee7f2a2900', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'af3c317a-4007-4cea-a060-1e7dde5ce49e', 'attached_at': '', 'detached_at': '', 'volume_id': 'cddb3090-62b4-4639-850d-4eee7f2a2900', 'serial': 'cddb3090-62b4-4639-850d-4eee7f2a2900'} {{(pid=61923) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 901.045246] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-725a4eea-b20a-4569-af9c-919cad416de9 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.059586] env[61923]: DEBUG oslo_vmware.api [-] Task: {'id': task-1377728, 'name': CreateVM_Task, 'duration_secs': 1.261755} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 901.086244] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: bac71328-3796-4a65-aa41-ccda93bc7926] Created VM on the ESX host {{(pid=61923) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 901.090596] env[61923]: DEBUG oslo_concurrency.lockutils [None req-6cea46fb-73b5-4047-8297-b9cdc5a261d0 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 901.090596] env[61923]: DEBUG oslo_concurrency.lockutils [None req-6cea46fb-73b5-4047-8297-b9cdc5a261d0 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 901.090596] env[61923]: DEBUG oslo_concurrency.lockutils [None req-6cea46fb-73b5-4047-8297-b9cdc5a261d0 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 901.090596] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5fc89500-1fad-4ed2-8876-f8ba538fe479 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.093432] env[61923]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7e80e53f-d3cf-4cf7-a512-03bba06c31d5 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.101274] env[61923]: DEBUG oslo_vmware.api [None req-6cea46fb-73b5-4047-8297-b9cdc5a261d0 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Waiting for the task: (returnval){ [ 901.101274] env[61923]: value = "session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]5208eb64-1d04-71dd-f94d-1cd6c2f19905" [ 901.101274] env[61923]: _type = "Task" [ 901.101274] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 901.108186] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9680704-d544-49c3-b20c-4dd7e4a20b85 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.138680] env[61923]: DEBUG oslo_vmware.api [None req-6cea46fb-73b5-4047-8297-b9cdc5a261d0 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]5208eb64-1d04-71dd-f94d-1cd6c2f19905, 'name': SearchDatastore_Task, 'duration_secs': 0.014892} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 901.142087] env[61923]: DEBUG oslo_concurrency.lockutils [None req-6cea46fb-73b5-4047-8297-b9cdc5a261d0 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 901.142311] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-6cea46fb-73b5-4047-8297-b9cdc5a261d0 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] [instance: bac71328-3796-4a65-aa41-ccda93bc7926] Processing image 0f153f63-ae0a-45b1-b7f5-7f9b673c947f {{(pid=61923) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 901.142611] env[61923]: DEBUG oslo_concurrency.lockutils [None req-6cea46fb-73b5-4047-8297-b9cdc5a261d0 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 901.142722] env[61923]: DEBUG oslo_concurrency.lockutils [None req-6cea46fb-73b5-4047-8297-b9cdc5a261d0 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 901.142891] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-6cea46fb-73b5-4047-8297-b9cdc5a261d0 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61923) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 901.143835] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d2c45ec-8c36-4f73-b769-b25eca0523ef {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.149736] env[61923]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-5cd8f8a1-0216-411b-bdfe-2a27f7e156bf {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.152355] env[61923]: DEBUG oslo_vmware.api [-] Task: {'id': task-1377732, 'name': CreateVM_Task} progress is 99%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.177544] env[61923]: DEBUG nova.virt.vmwareapi.volumeops [None req-f4c3a587-afc9-4326-a453-4024294c25b5 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] The volume has not been displaced from its original location: [datastore1] volume-cddb3090-62b4-4639-850d-4eee7f2a2900/volume-cddb3090-62b4-4639-850d-4eee7f2a2900.vmdk. No consolidation needed. {{(pid=61923) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 901.182396] env[61923]: DEBUG nova.virt.vmwareapi.volumeops [None req-f4c3a587-afc9-4326-a453-4024294c25b5 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] [instance: af3c317a-4007-4cea-a060-1e7dde5ce49e] Reconfiguring VM instance instance-00000047 to detach disk 2001 {{(pid=61923) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 901.183974] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-746f6833-a686-4333-8183-5aaf07bc0d4e {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.197400] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-6cea46fb-73b5-4047-8297-b9cdc5a261d0 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61923) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 901.197823] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-6cea46fb-73b5-4047-8297-b9cdc5a261d0 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61923) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 901.198873] env[61923]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-aacceab0-bd7a-4229-a014-56433c22a5a7 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.207208] env[61923]: DEBUG oslo_vmware.api [None req-6cea46fb-73b5-4047-8297-b9cdc5a261d0 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Waiting for the task: (returnval){ [ 901.207208] env[61923]: value = "session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]5236346e-e3fe-72a5-6dd1-b8582543ab31" [ 901.207208] env[61923]: _type = "Task" [ 901.207208] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 901.212695] env[61923]: DEBUG oslo_vmware.api [None req-f4c3a587-afc9-4326-a453-4024294c25b5 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] Waiting for the task: (returnval){ [ 901.212695] env[61923]: value = "task-1377733" [ 901.212695] env[61923]: _type = "Task" [ 901.212695] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 901.220569] env[61923]: DEBUG oslo_vmware.api [None req-6cea46fb-73b5-4047-8297-b9cdc5a261d0 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]5236346e-e3fe-72a5-6dd1-b8582543ab31, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.226203] env[61923]: DEBUG oslo_vmware.api [None req-f4c3a587-afc9-4326-a453-4024294c25b5 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] Task: {'id': task-1377733, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.259294] env[61923]: DEBUG oslo_vmware.api [None req-4482c5e6-0d59-468c-9d9f-cee661c7eaed tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Task: {'id': task-1377731, 'name': RemoveSnapshot_Task} progress is 15%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.267988] env[61923]: DEBUG oslo_concurrency.lockutils [None req-74d3449c-2b90-473c-a8b0-7be80c06daed tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Acquiring lock "interface-6012d3b5-739c-4762-9bb4-09c51171dcd7-b395b183-1c30-42de-929f-690ea10d9d75" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 901.268323] env[61923]: DEBUG oslo_concurrency.lockutils [None req-74d3449c-2b90-473c-a8b0-7be80c06daed tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Lock "interface-6012d3b5-739c-4762-9bb4-09c51171dcd7-b395b183-1c30-42de-929f-690ea10d9d75" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 901.268719] env[61923]: DEBUG nova.objects.instance [None req-74d3449c-2b90-473c-a8b0-7be80c06daed tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Lazy-loading 'flavor' on Instance uuid 6012d3b5-739c-4762-9bb4-09c51171dcd7 {{(pid=61923) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 901.345597] env[61923]: DEBUG oslo_vmware.api [None req-3192dd2d-7513-4a93-8a8c-c00e6e67889c tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] Task: {'id': task-1377729, 'name': PowerOnVM_Task, 'duration_secs': 1.250921} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 901.345924] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-3192dd2d-7513-4a93-8a8c-c00e6e67889c tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] [instance: 263f0866-49af-4ab6-8132-d993642ebedb] Powered on the VM {{(pid=61923) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 901.346114] env[61923]: INFO nova.compute.manager [None req-3192dd2d-7513-4a93-8a8c-c00e6e67889c tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] [instance: 263f0866-49af-4ab6-8132-d993642ebedb] Took 11.66 seconds to spawn the instance on the hypervisor. [ 901.346333] env[61923]: DEBUG nova.compute.manager [None req-3192dd2d-7513-4a93-8a8c-c00e6e67889c tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] [instance: 263f0866-49af-4ab6-8132-d993642ebedb] Checking state {{(pid=61923) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 901.347131] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83122102-814a-416c-bceb-461794430fe7 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.351108] env[61923]: DEBUG nova.compute.utils [None req-cb709b57-999a-41a8-b165-7029c0613fc5 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Using /dev/sd instead of None {{(pid=61923) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 901.356426] env[61923]: DEBUG nova.compute.manager [None req-cb709b57-999a-41a8-b165-7029c0613fc5 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: 03c1e8ea-a322-46fa-9b07-e283080a7871] Allocating IP information in the background. {{(pid=61923) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 901.356469] env[61923]: DEBUG nova.network.neutron [None req-cb709b57-999a-41a8-b165-7029c0613fc5 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: 03c1e8ea-a322-46fa-9b07-e283080a7871] allocate_for_instance() {{(pid=61923) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 901.402590] env[61923]: DEBUG oslo_vmware.api [None req-b7a2d28f-39dc-4a35-bb16-94846b77bc41 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Task: {'id': task-1377730, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.12449} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 901.403080] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-b7a2d28f-39dc-4a35-bb16-94846b77bc41 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk to [datastore2] c98c96a5-d18e-4519-996a-949ded16225f/c98c96a5-d18e-4519-996a-949ded16225f.vmdk {{(pid=61923) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 901.403439] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-b7a2d28f-39dc-4a35-bb16-94846b77bc41 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] [instance: c98c96a5-d18e-4519-996a-949ded16225f] Extending root virtual disk to 1048576 {{(pid=61923) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 901.404410] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-2892d4f1-87ba-4352-9818-0b24cf98aa10 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.414347] env[61923]: DEBUG oslo_vmware.api [None req-b7a2d28f-39dc-4a35-bb16-94846b77bc41 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Waiting for the task: (returnval){ [ 901.414347] env[61923]: value = "task-1377734" [ 901.414347] env[61923]: _type = "Task" [ 901.414347] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 901.428683] env[61923]: DEBUG oslo_vmware.api [None req-b7a2d28f-39dc-4a35-bb16-94846b77bc41 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Task: {'id': task-1377734, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.465060] env[61923]: DEBUG nova.policy [None req-cb709b57-999a-41a8-b165-7029c0613fc5 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5b8a639e99914e75857b4571f0d58a3c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '69cc941eb6dd4780ac12aa29656c37f7', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61923) authorize /opt/stack/nova/nova/policy.py:201}} [ 901.488088] env[61923]: DEBUG nova.network.neutron [req-bfc4d372-54f0-4569-8793-5a5c05278256 req-f57d85d8-2133-49ba-b9c2-262a73e962b5 service nova] [instance: 422d06d6-6932-46c3-bb25-841e6f65c028] Updated VIF entry in instance network info cache for port 97caa1b8-427c-46b0-be7a-d59b932900c3. {{(pid=61923) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 901.488783] env[61923]: DEBUG nova.network.neutron [req-bfc4d372-54f0-4569-8793-5a5c05278256 req-f57d85d8-2133-49ba-b9c2-262a73e962b5 service nova] [instance: 422d06d6-6932-46c3-bb25-841e6f65c028] Updating instance_info_cache with network_info: [{"id": "97caa1b8-427c-46b0-be7a-d59b932900c3", "address": "fa:16:3e:fa:f8:d1", "network": {"id": "a8b7e1b4-1ec6-4226-bdd6-535dbce450d2", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1928758125-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "912300a09b80452e85e3cd13c4a644cc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7806fe18-2b89-4386-87b1-f22876f82af2", "external-id": "nsx-vlan-transportzone-727", "segmentation_id": 727, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap97caa1b8-42", "ovs_interfaceid": "97caa1b8-427c-46b0-be7a-d59b932900c3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "b395b183-1c30-42de-929f-690ea10d9d75", "address": "fa:16:3e:7a:d5:2a", "network": {"id": "a8b7e1b4-1ec6-4226-bdd6-535dbce450d2", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1928758125-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "912300a09b80452e85e3cd13c4a644cc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7806fe18-2b89-4386-87b1-f22876f82af2", "external-id": "nsx-vlan-transportzone-727", "segmentation_id": 727, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb395b183-1c", "ovs_interfaceid": "b395b183-1c30-42de-929f-690ea10d9d75", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 901.599543] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47e8edd5-fc20-4d10-b605-3a7af760e2a5 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.607801] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06bbdd9c-9258-485e-a7ea-e4fcccc90877 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.643735] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9870b6ac-6e33-4e5b-a9b3-989f3cfd8255 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.649324] env[61923]: DEBUG oslo_vmware.api [-] Task: {'id': task-1377732, 'name': CreateVM_Task, 'duration_secs': 0.557476} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 901.649849] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 624e9207-7f24-4f45-8f52-0fc7a60c1e84] Created VM on the ESX host {{(pid=61923) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 901.650336] env[61923]: DEBUG oslo_concurrency.lockutils [None req-e1b75445-6f25-4dcb-9846-d40076ac5b42 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 901.650503] env[61923]: DEBUG oslo_concurrency.lockutils [None req-e1b75445-6f25-4dcb-9846-d40076ac5b42 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 901.650825] env[61923]: DEBUG oslo_concurrency.lockutils [None req-e1b75445-6f25-4dcb-9846-d40076ac5b42 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 901.651103] env[61923]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5cc0ce66-82ec-4b70-85ab-5ce9da4220b5 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.656188] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26f8ffd4-250c-462d-8bcc-c31766ef8899 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.661150] env[61923]: DEBUG oslo_vmware.api [None req-e1b75445-6f25-4dcb-9846-d40076ac5b42 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Waiting for the task: (returnval){ [ 901.661150] env[61923]: value = "session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]525b514d-144f-301d-c5eb-ee535ee680f9" [ 901.661150] env[61923]: _type = "Task" [ 901.661150] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 901.673077] env[61923]: DEBUG nova.compute.provider_tree [None req-7225546b-1247-4af1-9195-074d155e3aeb tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 901.681027] env[61923]: DEBUG oslo_vmware.api [None req-e1b75445-6f25-4dcb-9846-d40076ac5b42 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]525b514d-144f-301d-c5eb-ee535ee680f9, 'name': SearchDatastore_Task, 'duration_secs': 0.018888} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 901.681355] env[61923]: DEBUG oslo_concurrency.lockutils [None req-e1b75445-6f25-4dcb-9846-d40076ac5b42 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 901.681603] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-e1b75445-6f25-4dcb-9846-d40076ac5b42 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] [instance: 624e9207-7f24-4f45-8f52-0fc7a60c1e84] Processing image 0f153f63-ae0a-45b1-b7f5-7f9b673c947f {{(pid=61923) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 901.681818] env[61923]: DEBUG oslo_concurrency.lockutils [None req-e1b75445-6f25-4dcb-9846-d40076ac5b42 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 901.721374] env[61923]: DEBUG oslo_vmware.api [None req-6cea46fb-73b5-4047-8297-b9cdc5a261d0 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]5236346e-e3fe-72a5-6dd1-b8582543ab31, 'name': SearchDatastore_Task, 'duration_secs': 0.036456} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 901.722589] env[61923]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-63ecda16-aa81-4542-afff-bae0a572fdf3 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.728101] env[61923]: DEBUG oslo_vmware.api [None req-f4c3a587-afc9-4326-a453-4024294c25b5 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] Task: {'id': task-1377733, 'name': ReconfigVM_Task, 'duration_secs': 0.350629} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 901.728773] env[61923]: DEBUG nova.virt.vmwareapi.volumeops [None req-f4c3a587-afc9-4326-a453-4024294c25b5 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] [instance: af3c317a-4007-4cea-a060-1e7dde5ce49e] Reconfigured VM instance instance-00000047 to detach disk 2001 {{(pid=61923) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 901.734658] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1a6522da-af04-4a16-bb06-700d92b0a1d8 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.745880] env[61923]: DEBUG oslo_vmware.api [None req-6cea46fb-73b5-4047-8297-b9cdc5a261d0 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Waiting for the task: (returnval){ [ 901.745880] env[61923]: value = "session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52070077-d4f8-521e-25cd-f94e956a9a30" [ 901.745880] env[61923]: _type = "Task" [ 901.745880] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 901.757653] env[61923]: DEBUG oslo_vmware.api [None req-f4c3a587-afc9-4326-a453-4024294c25b5 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] Waiting for the task: (returnval){ [ 901.757653] env[61923]: value = "task-1377735" [ 901.757653] env[61923]: _type = "Task" [ 901.757653] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 901.770658] env[61923]: DEBUG oslo_vmware.api [None req-4482c5e6-0d59-468c-9d9f-cee661c7eaed tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Task: {'id': task-1377731, 'name': RemoveSnapshot_Task} progress is 100%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.770968] env[61923]: DEBUG oslo_vmware.api [None req-6cea46fb-73b5-4047-8297-b9cdc5a261d0 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52070077-d4f8-521e-25cd-f94e956a9a30, 'name': SearchDatastore_Task, 'duration_secs': 0.01395} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 901.775050] env[61923]: DEBUG oslo_concurrency.lockutils [None req-6cea46fb-73b5-4047-8297-b9cdc5a261d0 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 901.775405] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-6cea46fb-73b5-4047-8297-b9cdc5a261d0 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk to [datastore1] bac71328-3796-4a65-aa41-ccda93bc7926/bac71328-3796-4a65-aa41-ccda93bc7926.vmdk {{(pid=61923) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 901.779636] env[61923]: DEBUG oslo_concurrency.lockutils [None req-e1b75445-6f25-4dcb-9846-d40076ac5b42 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 901.779933] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-e1b75445-6f25-4dcb-9846-d40076ac5b42 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61923) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 901.781087] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-6208864e-f803-4459-88dd-65196fe2ae91 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.784363] env[61923]: DEBUG oslo_vmware.api [None req-f4c3a587-afc9-4326-a453-4024294c25b5 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] Task: {'id': task-1377735, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.786827] env[61923]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-2928164b-3ac5-415b-968d-7e650e8b7ca7 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.795377] env[61923]: DEBUG oslo_vmware.api [None req-6cea46fb-73b5-4047-8297-b9cdc5a261d0 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Waiting for the task: (returnval){ [ 901.795377] env[61923]: value = "task-1377736" [ 901.795377] env[61923]: _type = "Task" [ 901.795377] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 901.801702] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-e1b75445-6f25-4dcb-9846-d40076ac5b42 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61923) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 901.802013] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-e1b75445-6f25-4dcb-9846-d40076ac5b42 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61923) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 901.803361] env[61923]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4734934b-5b2a-4a44-84ff-e9d3d5165c98 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.809837] env[61923]: DEBUG oslo_vmware.api [None req-6cea46fb-73b5-4047-8297-b9cdc5a261d0 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Task: {'id': task-1377736, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.815233] env[61923]: DEBUG oslo_vmware.api [None req-e1b75445-6f25-4dcb-9846-d40076ac5b42 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Waiting for the task: (returnval){ [ 901.815233] env[61923]: value = "session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52d2c5d0-9d87-7c61-2252-562e6ad083aa" [ 901.815233] env[61923]: _type = "Task" [ 901.815233] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 901.825810] env[61923]: DEBUG oslo_vmware.api [None req-e1b75445-6f25-4dcb-9846-d40076ac5b42 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52d2c5d0-9d87-7c61-2252-562e6ad083aa, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.826425] env[61923]: DEBUG nova.network.neutron [None req-cb709b57-999a-41a8-b165-7029c0613fc5 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: 03c1e8ea-a322-46fa-9b07-e283080a7871] Successfully created port: 37dd225f-bb7f-471b-8484-5387c0e22183 {{(pid=61923) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 901.856806] env[61923]: DEBUG nova.compute.manager [None req-cb709b57-999a-41a8-b165-7029c0613fc5 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: 03c1e8ea-a322-46fa-9b07-e283080a7871] Start building block device mappings for instance. {{(pid=61923) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 901.880220] env[61923]: INFO nova.compute.manager [None req-3192dd2d-7513-4a93-8a8c-c00e6e67889c tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] [instance: 263f0866-49af-4ab6-8132-d993642ebedb] Took 25.60 seconds to build instance. [ 901.926823] env[61923]: DEBUG oslo_vmware.api [None req-b7a2d28f-39dc-4a35-bb16-94846b77bc41 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Task: {'id': task-1377734, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.142007} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 901.927207] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-b7a2d28f-39dc-4a35-bb16-94846b77bc41 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] [instance: c98c96a5-d18e-4519-996a-949ded16225f] Extended root virtual disk {{(pid=61923) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 901.928173] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4aba4dcb-c79e-44be-bfe7-459da45f88cf {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.948977] env[61923]: DEBUG nova.virt.vmwareapi.volumeops [None req-b7a2d28f-39dc-4a35-bb16-94846b77bc41 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] [instance: c98c96a5-d18e-4519-996a-949ded16225f] Reconfiguring VM instance instance-00000054 to attach disk [datastore2] c98c96a5-d18e-4519-996a-949ded16225f/c98c96a5-d18e-4519-996a-949ded16225f.vmdk or device None with type sparse {{(pid=61923) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 901.949287] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e1c67676-bc0d-448a-bdab-36482ccd9b14 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.970346] env[61923]: DEBUG oslo_vmware.api [None req-b7a2d28f-39dc-4a35-bb16-94846b77bc41 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Waiting for the task: (returnval){ [ 901.970346] env[61923]: value = "task-1377737" [ 901.970346] env[61923]: _type = "Task" [ 901.970346] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 901.981862] env[61923]: DEBUG oslo_vmware.api [None req-b7a2d28f-39dc-4a35-bb16-94846b77bc41 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Task: {'id': task-1377737, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.994932] env[61923]: DEBUG oslo_concurrency.lockutils [req-bfc4d372-54f0-4569-8793-5a5c05278256 req-f57d85d8-2133-49ba-b9c2-262a73e962b5 service nova] Releasing lock "refresh_cache-422d06d6-6932-46c3-bb25-841e6f65c028" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 901.995624] env[61923]: DEBUG oslo_concurrency.lockutils [None req-214ded71-bc45-4cf6-9e81-e6c214aa77f2 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Acquired lock "refresh_cache-422d06d6-6932-46c3-bb25-841e6f65c028" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 901.996657] env[61923]: DEBUG nova.network.neutron [None req-214ded71-bc45-4cf6-9e81-e6c214aa77f2 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] [instance: 422d06d6-6932-46c3-bb25-841e6f65c028] Building network info cache for instance {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 902.054610] env[61923]: DEBUG nova.objects.instance [None req-74d3449c-2b90-473c-a8b0-7be80c06daed tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Lazy-loading 'pci_requests' on Instance uuid 6012d3b5-739c-4762-9bb4-09c51171dcd7 {{(pid=61923) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 902.177238] env[61923]: DEBUG nova.scheduler.client.report [None req-7225546b-1247-4af1-9195-074d155e3aeb tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 177, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 902.265229] env[61923]: DEBUG oslo_vmware.api [None req-4482c5e6-0d59-468c-9d9f-cee661c7eaed tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Task: {'id': task-1377731, 'name': RemoveSnapshot_Task, 'duration_secs': 1.564311} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 902.265874] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-4482c5e6-0d59-468c-9d9f-cee661c7eaed tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] [instance: 2719569b-8572-4199-8158-7bb367d17dc5] Deleted Snapshot of the VM instance {{(pid=61923) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 902.266167] env[61923]: INFO nova.compute.manager [None req-4482c5e6-0d59-468c-9d9f-cee661c7eaed tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] [instance: 2719569b-8572-4199-8158-7bb367d17dc5] Took 19.03 seconds to snapshot the instance on the hypervisor. [ 902.272807] env[61923]: DEBUG oslo_vmware.api [None req-f4c3a587-afc9-4326-a453-4024294c25b5 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] Task: {'id': task-1377735, 'name': ReconfigVM_Task, 'duration_secs': 0.225308} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 902.275209] env[61923]: DEBUG nova.virt.vmwareapi.volumeops [None req-f4c3a587-afc9-4326-a453-4024294c25b5 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] [instance: af3c317a-4007-4cea-a060-1e7dde5ce49e] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-292706', 'volume_id': 'cddb3090-62b4-4639-850d-4eee7f2a2900', 'name': 'volume-cddb3090-62b4-4639-850d-4eee7f2a2900', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'af3c317a-4007-4cea-a060-1e7dde5ce49e', 'attached_at': '', 'detached_at': '', 'volume_id': 'cddb3090-62b4-4639-850d-4eee7f2a2900', 'serial': 'cddb3090-62b4-4639-850d-4eee7f2a2900'} {{(pid=61923) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 902.306552] env[61923]: DEBUG oslo_vmware.api [None req-6cea46fb-73b5-4047-8297-b9cdc5a261d0 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Task: {'id': task-1377736, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.325988] env[61923]: DEBUG oslo_vmware.api [None req-e1b75445-6f25-4dcb-9846-d40076ac5b42 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52d2c5d0-9d87-7c61-2252-562e6ad083aa, 'name': SearchDatastore_Task, 'duration_secs': 0.039606} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 902.326821] env[61923]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ee0e615e-b897-4b00-9e68-86794ed47dec {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.332427] env[61923]: DEBUG oslo_vmware.api [None req-e1b75445-6f25-4dcb-9846-d40076ac5b42 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Waiting for the task: (returnval){ [ 902.332427] env[61923]: value = "session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52879d2b-dd75-2186-0eaf-b633f87be7b5" [ 902.332427] env[61923]: _type = "Task" [ 902.332427] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 902.340673] env[61923]: DEBUG oslo_vmware.api [None req-e1b75445-6f25-4dcb-9846-d40076ac5b42 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52879d2b-dd75-2186-0eaf-b633f87be7b5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.381157] env[61923]: DEBUG oslo_concurrency.lockutils [None req-3192dd2d-7513-4a93-8a8c-c00e6e67889c tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] Lock "263f0866-49af-4ab6-8132-d993642ebedb" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 27.108s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 902.480599] env[61923]: DEBUG oslo_vmware.api [None req-b7a2d28f-39dc-4a35-bb16-94846b77bc41 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Task: {'id': task-1377737, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.560728] env[61923]: DEBUG nova.objects.base [None req-74d3449c-2b90-473c-a8b0-7be80c06daed tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Object Instance<6012d3b5-739c-4762-9bb4-09c51171dcd7> lazy-loaded attributes: flavor,pci_requests {{(pid=61923) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 902.560956] env[61923]: DEBUG nova.network.neutron [None req-74d3449c-2b90-473c-a8b0-7be80c06daed tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] [instance: 6012d3b5-739c-4762-9bb4-09c51171dcd7] allocate_for_instance() {{(pid=61923) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 902.633292] env[61923]: DEBUG nova.policy [None req-74d3449c-2b90-473c-a8b0-7be80c06daed tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'cd5039bf8f4344e490df828e643d2779', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '912300a09b80452e85e3cd13c4a644cc', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61923) authorize /opt/stack/nova/nova/policy.py:201}} [ 902.683835] env[61923]: DEBUG oslo_concurrency.lockutils [None req-7225546b-1247-4af1-9195-074d155e3aeb tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.837s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 902.719539] env[61923]: INFO nova.scheduler.client.report [None req-7225546b-1247-4af1-9195-074d155e3aeb tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Deleted allocations for instance 4cb3c397-ef1a-4267-a300-26da48206f3f [ 902.808935] env[61923]: DEBUG oslo_vmware.api [None req-6cea46fb-73b5-4047-8297-b9cdc5a261d0 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Task: {'id': task-1377736, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.830435] env[61923]: DEBUG nova.objects.instance [None req-f4c3a587-afc9-4326-a453-4024294c25b5 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] Lazy-loading 'flavor' on Instance uuid af3c317a-4007-4cea-a060-1e7dde5ce49e {{(pid=61923) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 902.842765] env[61923]: DEBUG oslo_vmware.api [None req-e1b75445-6f25-4dcb-9846-d40076ac5b42 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52879d2b-dd75-2186-0eaf-b633f87be7b5, 'name': SearchDatastore_Task, 'duration_secs': 0.04349} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 902.843066] env[61923]: DEBUG oslo_concurrency.lockutils [None req-e1b75445-6f25-4dcb-9846-d40076ac5b42 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 902.843339] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-e1b75445-6f25-4dcb-9846-d40076ac5b42 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk to [datastore1] 624e9207-7f24-4f45-8f52-0fc7a60c1e84/624e9207-7f24-4f45-8f52-0fc7a60c1e84.vmdk {{(pid=61923) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 902.843631] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-781da19c-9ca9-49f6-b4ba-f7c02d23f145 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.851985] env[61923]: DEBUG oslo_vmware.api [None req-e1b75445-6f25-4dcb-9846-d40076ac5b42 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Waiting for the task: (returnval){ [ 902.851985] env[61923]: value = "task-1377738" [ 902.851985] env[61923]: _type = "Task" [ 902.851985] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 902.863290] env[61923]: DEBUG oslo_vmware.api [None req-e1b75445-6f25-4dcb-9846-d40076ac5b42 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Task: {'id': task-1377738, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.867766] env[61923]: DEBUG nova.compute.manager [None req-cb709b57-999a-41a8-b165-7029c0613fc5 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: 03c1e8ea-a322-46fa-9b07-e283080a7871] Start spawning the instance on the hypervisor. {{(pid=61923) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 902.872590] env[61923]: INFO nova.network.neutron [None req-214ded71-bc45-4cf6-9e81-e6c214aa77f2 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] [instance: 422d06d6-6932-46c3-bb25-841e6f65c028] Port b395b183-1c30-42de-929f-690ea10d9d75 from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 902.873753] env[61923]: DEBUG nova.network.neutron [None req-214ded71-bc45-4cf6-9e81-e6c214aa77f2 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] [instance: 422d06d6-6932-46c3-bb25-841e6f65c028] Updating instance_info_cache with network_info: [{"id": "97caa1b8-427c-46b0-be7a-d59b932900c3", "address": "fa:16:3e:fa:f8:d1", "network": {"id": "a8b7e1b4-1ec6-4226-bdd6-535dbce450d2", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1928758125-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "912300a09b80452e85e3cd13c4a644cc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7806fe18-2b89-4386-87b1-f22876f82af2", "external-id": "nsx-vlan-transportzone-727", "segmentation_id": 727, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap97caa1b8-42", "ovs_interfaceid": "97caa1b8-427c-46b0-be7a-d59b932900c3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 902.906319] env[61923]: DEBUG nova.virt.hardware [None req-cb709b57-999a-41a8-b165-7029c0613fc5 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-29T20:07:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-29T20:07:35Z,direct_url=,disk_format='vmdk',id=0f153f63-ae0a-45b1-b7f5-7f9b673c947f,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='4e7b5e3b3c3443c8bd5c70f8a15739f5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-29T20:07:36Z,virtual_size=,visibility=), allow threads: False {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 902.906691] env[61923]: DEBUG nova.virt.hardware [None req-cb709b57-999a-41a8-b165-7029c0613fc5 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Flavor limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 902.906776] env[61923]: DEBUG nova.virt.hardware [None req-cb709b57-999a-41a8-b165-7029c0613fc5 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Image limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 902.906913] env[61923]: DEBUG nova.virt.hardware [None req-cb709b57-999a-41a8-b165-7029c0613fc5 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Flavor pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 902.907183] env[61923]: DEBUG nova.virt.hardware [None req-cb709b57-999a-41a8-b165-7029c0613fc5 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Image pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 902.907423] env[61923]: DEBUG nova.virt.hardware [None req-cb709b57-999a-41a8-b165-7029c0613fc5 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 902.907758] env[61923]: DEBUG nova.virt.hardware [None req-cb709b57-999a-41a8-b165-7029c0613fc5 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 902.908026] env[61923]: DEBUG nova.virt.hardware [None req-cb709b57-999a-41a8-b165-7029c0613fc5 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 902.908424] env[61923]: DEBUG nova.virt.hardware [None req-cb709b57-999a-41a8-b165-7029c0613fc5 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Got 1 possible topologies {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 902.908712] env[61923]: DEBUG nova.virt.hardware [None req-cb709b57-999a-41a8-b165-7029c0613fc5 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 902.908862] env[61923]: DEBUG nova.virt.hardware [None req-cb709b57-999a-41a8-b165-7029c0613fc5 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 902.910103] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ac51201-0173-4310-b19d-ffded1c7b58b {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.915251] env[61923]: DEBUG nova.compute.manager [req-b00d9c40-6e79-4a30-8efc-3607eeb53d20 req-76a27ade-d77b-45c2-869d-af1e9d13a6e1 service nova] [instance: 6012d3b5-739c-4762-9bb4-09c51171dcd7] Received event network-changed-269a41b0-75e7-46af-9497-8eae3102df11 {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 902.915479] env[61923]: DEBUG nova.compute.manager [req-b00d9c40-6e79-4a30-8efc-3607eeb53d20 req-76a27ade-d77b-45c2-869d-af1e9d13a6e1 service nova] [instance: 6012d3b5-739c-4762-9bb4-09c51171dcd7] Refreshing instance network info cache due to event network-changed-269a41b0-75e7-46af-9497-8eae3102df11. {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 902.915723] env[61923]: DEBUG oslo_concurrency.lockutils [req-b00d9c40-6e79-4a30-8efc-3607eeb53d20 req-76a27ade-d77b-45c2-869d-af1e9d13a6e1 service nova] Acquiring lock "refresh_cache-6012d3b5-739c-4762-9bb4-09c51171dcd7" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 902.915888] env[61923]: DEBUG oslo_concurrency.lockutils [req-b00d9c40-6e79-4a30-8efc-3607eeb53d20 req-76a27ade-d77b-45c2-869d-af1e9d13a6e1 service nova] Acquired lock "refresh_cache-6012d3b5-739c-4762-9bb4-09c51171dcd7" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 902.916070] env[61923]: DEBUG nova.network.neutron [req-b00d9c40-6e79-4a30-8efc-3607eeb53d20 req-76a27ade-d77b-45c2-869d-af1e9d13a6e1 service nova] [instance: 6012d3b5-739c-4762-9bb4-09c51171dcd7] Refreshing network info cache for port 269a41b0-75e7-46af-9497-8eae3102df11 {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 902.923283] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-126287c4-e564-4000-ba21-0c903c93ad88 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.980229] env[61923]: DEBUG oslo_vmware.api [None req-b7a2d28f-39dc-4a35-bb16-94846b77bc41 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Task: {'id': task-1377737, 'name': ReconfigVM_Task, 'duration_secs': 0.936687} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 902.980462] env[61923]: DEBUG nova.virt.vmwareapi.volumeops [None req-b7a2d28f-39dc-4a35-bb16-94846b77bc41 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] [instance: c98c96a5-d18e-4519-996a-949ded16225f] Reconfigured VM instance instance-00000054 to attach disk [datastore2] c98c96a5-d18e-4519-996a-949ded16225f/c98c96a5-d18e-4519-996a-949ded16225f.vmdk or device None with type sparse {{(pid=61923) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 902.981079] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f34c5d07-1d93-4f9a-9bc4-d0c8d6020d64 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.987009] env[61923]: DEBUG oslo_vmware.api [None req-b7a2d28f-39dc-4a35-bb16-94846b77bc41 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Waiting for the task: (returnval){ [ 902.987009] env[61923]: value = "task-1377739" [ 902.987009] env[61923]: _type = "Task" [ 902.987009] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 902.994850] env[61923]: DEBUG oslo_vmware.api [None req-b7a2d28f-39dc-4a35-bb16-94846b77bc41 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Task: {'id': task-1377739, 'name': Rename_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.018816] env[61923]: DEBUG oslo_concurrency.lockutils [None req-4a664699-f279-4048-aa87-1277ae6e408d tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] Acquiring lock "263f0866-49af-4ab6-8132-d993642ebedb" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 903.019096] env[61923]: DEBUG oslo_concurrency.lockutils [None req-4a664699-f279-4048-aa87-1277ae6e408d tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] Lock "263f0866-49af-4ab6-8132-d993642ebedb" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 903.019377] env[61923]: DEBUG oslo_concurrency.lockutils [None req-4a664699-f279-4048-aa87-1277ae6e408d tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] Acquiring lock "263f0866-49af-4ab6-8132-d993642ebedb-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 903.019620] env[61923]: DEBUG oslo_concurrency.lockutils [None req-4a664699-f279-4048-aa87-1277ae6e408d tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] Lock "263f0866-49af-4ab6-8132-d993642ebedb-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 903.019812] env[61923]: DEBUG oslo_concurrency.lockutils [None req-4a664699-f279-4048-aa87-1277ae6e408d tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] Lock "263f0866-49af-4ab6-8132-d993642ebedb-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 903.022043] env[61923]: INFO nova.compute.manager [None req-4a664699-f279-4048-aa87-1277ae6e408d tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] [instance: 263f0866-49af-4ab6-8132-d993642ebedb] Terminating instance [ 903.023673] env[61923]: DEBUG nova.compute.manager [None req-4a664699-f279-4048-aa87-1277ae6e408d tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] [instance: 263f0866-49af-4ab6-8132-d993642ebedb] Start destroying the instance on the hypervisor. {{(pid=61923) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 903.023877] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-4a664699-f279-4048-aa87-1277ae6e408d tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] [instance: 263f0866-49af-4ab6-8132-d993642ebedb] Destroying instance {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 903.024878] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92b877b6-4ff3-4516-893f-efe9e1118410 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.032941] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-4a664699-f279-4048-aa87-1277ae6e408d tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] [instance: 263f0866-49af-4ab6-8132-d993642ebedb] Powering off the VM {{(pid=61923) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 903.033188] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6f768ec1-fad9-4d98-a364-f0e93f966c43 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.040442] env[61923]: DEBUG oslo_vmware.api [None req-4a664699-f279-4048-aa87-1277ae6e408d tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] Waiting for the task: (returnval){ [ 903.040442] env[61923]: value = "task-1377740" [ 903.040442] env[61923]: _type = "Task" [ 903.040442] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 903.049063] env[61923]: DEBUG oslo_vmware.api [None req-4a664699-f279-4048-aa87-1277ae6e408d tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] Task: {'id': task-1377740, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.231111] env[61923]: DEBUG oslo_concurrency.lockutils [None req-7225546b-1247-4af1-9195-074d155e3aeb tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Lock "4cb3c397-ef1a-4267-a300-26da48206f3f" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.653s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 903.306735] env[61923]: DEBUG oslo_vmware.api [None req-6cea46fb-73b5-4047-8297-b9cdc5a261d0 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Task: {'id': task-1377736, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.144134} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 903.307078] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-6cea46fb-73b5-4047-8297-b9cdc5a261d0 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk to [datastore1] bac71328-3796-4a65-aa41-ccda93bc7926/bac71328-3796-4a65-aa41-ccda93bc7926.vmdk {{(pid=61923) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 903.307308] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-6cea46fb-73b5-4047-8297-b9cdc5a261d0 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] [instance: bac71328-3796-4a65-aa41-ccda93bc7926] Extending root virtual disk to 1048576 {{(pid=61923) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 903.307611] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c15afcf3-8e66-4d6a-9602-7c66c14cd25a {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.314974] env[61923]: DEBUG oslo_vmware.api [None req-6cea46fb-73b5-4047-8297-b9cdc5a261d0 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Waiting for the task: (returnval){ [ 903.314974] env[61923]: value = "task-1377741" [ 903.314974] env[61923]: _type = "Task" [ 903.314974] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 903.324083] env[61923]: DEBUG oslo_vmware.api [None req-6cea46fb-73b5-4047-8297-b9cdc5a261d0 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Task: {'id': task-1377741, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.363235] env[61923]: DEBUG oslo_vmware.api [None req-e1b75445-6f25-4dcb-9846-d40076ac5b42 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Task: {'id': task-1377738, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.376359] env[61923]: DEBUG oslo_concurrency.lockutils [None req-214ded71-bc45-4cf6-9e81-e6c214aa77f2 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Releasing lock "refresh_cache-422d06d6-6932-46c3-bb25-841e6f65c028" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 903.497921] env[61923]: DEBUG oslo_vmware.api [None req-b7a2d28f-39dc-4a35-bb16-94846b77bc41 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Task: {'id': task-1377739, 'name': Rename_Task, 'duration_secs': 0.235761} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 903.498262] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-b7a2d28f-39dc-4a35-bb16-94846b77bc41 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] [instance: c98c96a5-d18e-4519-996a-949ded16225f] Powering on the VM {{(pid=61923) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 903.498535] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-33751f7a-798f-4833-b54d-aef8b51138e8 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.505835] env[61923]: DEBUG oslo_vmware.api [None req-b7a2d28f-39dc-4a35-bb16-94846b77bc41 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Waiting for the task: (returnval){ [ 903.505835] env[61923]: value = "task-1377742" [ 903.505835] env[61923]: _type = "Task" [ 903.505835] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 903.515128] env[61923]: DEBUG oslo_vmware.api [None req-b7a2d28f-39dc-4a35-bb16-94846b77bc41 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Task: {'id': task-1377742, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.550691] env[61923]: DEBUG oslo_vmware.api [None req-4a664699-f279-4048-aa87-1277ae6e408d tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] Task: {'id': task-1377740, 'name': PowerOffVM_Task, 'duration_secs': 0.207312} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 903.551008] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-4a664699-f279-4048-aa87-1277ae6e408d tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] [instance: 263f0866-49af-4ab6-8132-d993642ebedb] Powered off the VM {{(pid=61923) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 903.551192] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-4a664699-f279-4048-aa87-1277ae6e408d tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] [instance: 263f0866-49af-4ab6-8132-d993642ebedb] Unregistering the VM {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 903.551465] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-0323824d-cd1f-4925-b164-27c7eedab81f {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.643391] env[61923]: DEBUG nova.network.neutron [None req-cb709b57-999a-41a8-b165-7029c0613fc5 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: 03c1e8ea-a322-46fa-9b07-e283080a7871] Successfully updated port: 37dd225f-bb7f-471b-8484-5387c0e22183 {{(pid=61923) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 903.650886] env[61923]: DEBUG nova.compute.manager [req-1730042d-9a0d-4083-83c5-c4fb3bf1f08c req-87062a52-1433-4ca9-8987-2102e2eb78b4 service nova] [instance: 03c1e8ea-a322-46fa-9b07-e283080a7871] Received event network-vif-plugged-37dd225f-bb7f-471b-8484-5387c0e22183 {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 903.651150] env[61923]: DEBUG oslo_concurrency.lockutils [req-1730042d-9a0d-4083-83c5-c4fb3bf1f08c req-87062a52-1433-4ca9-8987-2102e2eb78b4 service nova] Acquiring lock "03c1e8ea-a322-46fa-9b07-e283080a7871-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 903.651778] env[61923]: DEBUG oslo_concurrency.lockutils [req-1730042d-9a0d-4083-83c5-c4fb3bf1f08c req-87062a52-1433-4ca9-8987-2102e2eb78b4 service nova] Lock "03c1e8ea-a322-46fa-9b07-e283080a7871-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 903.651997] env[61923]: DEBUG oslo_concurrency.lockutils [req-1730042d-9a0d-4083-83c5-c4fb3bf1f08c req-87062a52-1433-4ca9-8987-2102e2eb78b4 service nova] Lock "03c1e8ea-a322-46fa-9b07-e283080a7871-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.001s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 903.652218] env[61923]: DEBUG nova.compute.manager [req-1730042d-9a0d-4083-83c5-c4fb3bf1f08c req-87062a52-1433-4ca9-8987-2102e2eb78b4 service nova] [instance: 03c1e8ea-a322-46fa-9b07-e283080a7871] No waiting events found dispatching network-vif-plugged-37dd225f-bb7f-471b-8484-5387c0e22183 {{(pid=61923) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 903.652401] env[61923]: WARNING nova.compute.manager [req-1730042d-9a0d-4083-83c5-c4fb3bf1f08c req-87062a52-1433-4ca9-8987-2102e2eb78b4 service nova] [instance: 03c1e8ea-a322-46fa-9b07-e283080a7871] Received unexpected event network-vif-plugged-37dd225f-bb7f-471b-8484-5387c0e22183 for instance with vm_state building and task_state spawning. [ 903.703860] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-4a664699-f279-4048-aa87-1277ae6e408d tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] [instance: 263f0866-49af-4ab6-8132-d993642ebedb] Unregistered the VM {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 903.704135] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-4a664699-f279-4048-aa87-1277ae6e408d tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] [instance: 263f0866-49af-4ab6-8132-d993642ebedb] Deleting contents of the VM from datastore datastore1 {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 903.704309] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-4a664699-f279-4048-aa87-1277ae6e408d tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] Deleting the datastore file [datastore1] 263f0866-49af-4ab6-8132-d993642ebedb {{(pid=61923) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 903.704576] env[61923]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-35222c4a-fab0-438f-b7fe-1e9cd731899f {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.711258] env[61923]: DEBUG oslo_vmware.api [None req-4a664699-f279-4048-aa87-1277ae6e408d tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] Waiting for the task: (returnval){ [ 903.711258] env[61923]: value = "task-1377744" [ 903.711258] env[61923]: _type = "Task" [ 903.711258] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 903.723869] env[61923]: DEBUG oslo_vmware.api [None req-4a664699-f279-4048-aa87-1277ae6e408d tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] Task: {'id': task-1377744, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.780575] env[61923]: DEBUG nova.network.neutron [req-b00d9c40-6e79-4a30-8efc-3607eeb53d20 req-76a27ade-d77b-45c2-869d-af1e9d13a6e1 service nova] [instance: 6012d3b5-739c-4762-9bb4-09c51171dcd7] Updated VIF entry in instance network info cache for port 269a41b0-75e7-46af-9497-8eae3102df11. {{(pid=61923) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 903.781339] env[61923]: DEBUG nova.network.neutron [req-b00d9c40-6e79-4a30-8efc-3607eeb53d20 req-76a27ade-d77b-45c2-869d-af1e9d13a6e1 service nova] [instance: 6012d3b5-739c-4762-9bb4-09c51171dcd7] Updating instance_info_cache with network_info: [{"id": "269a41b0-75e7-46af-9497-8eae3102df11", "address": "fa:16:3e:e1:02:b8", "network": {"id": "a8b7e1b4-1ec6-4226-bdd6-535dbce450d2", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1928758125-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.146", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "912300a09b80452e85e3cd13c4a644cc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7806fe18-2b89-4386-87b1-f22876f82af2", "external-id": "nsx-vlan-transportzone-727", "segmentation_id": 727, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap269a41b0-75", "ovs_interfaceid": "269a41b0-75e7-46af-9497-8eae3102df11", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 903.825010] env[61923]: DEBUG oslo_vmware.api [None req-6cea46fb-73b5-4047-8297-b9cdc5a261d0 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Task: {'id': task-1377741, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.293878} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 903.825372] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-6cea46fb-73b5-4047-8297-b9cdc5a261d0 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] [instance: bac71328-3796-4a65-aa41-ccda93bc7926] Extended root virtual disk {{(pid=61923) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 903.826161] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-136d9651-0ae6-4654-a87b-4930d3eafe50 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.840824] env[61923]: DEBUG oslo_concurrency.lockutils [None req-f4c3a587-afc9-4326-a453-4024294c25b5 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] Lock "af3c317a-4007-4cea-a060-1e7dde5ce49e" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.348s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 903.850180] env[61923]: DEBUG nova.virt.vmwareapi.volumeops [None req-6cea46fb-73b5-4047-8297-b9cdc5a261d0 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] [instance: bac71328-3796-4a65-aa41-ccda93bc7926] Reconfiguring VM instance instance-00000053 to attach disk [datastore1] bac71328-3796-4a65-aa41-ccda93bc7926/bac71328-3796-4a65-aa41-ccda93bc7926.vmdk or device None with type sparse {{(pid=61923) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 903.850721] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4b91f5c4-7e58-43b3-b1d9-0947e44380c1 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.875935] env[61923]: DEBUG oslo_vmware.api [None req-e1b75445-6f25-4dcb-9846-d40076ac5b42 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Task: {'id': task-1377738, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.747255} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 903.877138] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-e1b75445-6f25-4dcb-9846-d40076ac5b42 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk to [datastore1] 624e9207-7f24-4f45-8f52-0fc7a60c1e84/624e9207-7f24-4f45-8f52-0fc7a60c1e84.vmdk {{(pid=61923) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 903.877365] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-e1b75445-6f25-4dcb-9846-d40076ac5b42 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] [instance: 624e9207-7f24-4f45-8f52-0fc7a60c1e84] Extending root virtual disk to 1048576 {{(pid=61923) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 903.877898] env[61923]: DEBUG oslo_vmware.api [None req-6cea46fb-73b5-4047-8297-b9cdc5a261d0 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Waiting for the task: (returnval){ [ 903.877898] env[61923]: value = "task-1377745" [ 903.877898] env[61923]: _type = "Task" [ 903.877898] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 903.878108] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-e65cf685-718f-433d-be1f-5db9ba977b2e {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.883383] env[61923]: DEBUG oslo_concurrency.lockutils [None req-214ded71-bc45-4cf6-9e81-e6c214aa77f2 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Lock "interface-422d06d6-6932-46c3-bb25-841e6f65c028-b395b183-1c30-42de-929f-690ea10d9d75" "released" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: held 11.161s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 903.889831] env[61923]: DEBUG oslo_vmware.api [None req-e1b75445-6f25-4dcb-9846-d40076ac5b42 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Waiting for the task: (returnval){ [ 903.889831] env[61923]: value = "task-1377746" [ 903.889831] env[61923]: _type = "Task" [ 903.889831] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 903.890140] env[61923]: DEBUG oslo_vmware.api [None req-6cea46fb-73b5-4047-8297-b9cdc5a261d0 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Task: {'id': task-1377745, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.900203] env[61923]: DEBUG oslo_vmware.api [None req-e1b75445-6f25-4dcb-9846-d40076ac5b42 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Task: {'id': task-1377746, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.015891] env[61923]: DEBUG oslo_vmware.api [None req-b7a2d28f-39dc-4a35-bb16-94846b77bc41 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Task: {'id': task-1377742, 'name': PowerOnVM_Task, 'duration_secs': 0.494614} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 904.016159] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-b7a2d28f-39dc-4a35-bb16-94846b77bc41 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] [instance: c98c96a5-d18e-4519-996a-949ded16225f] Powered on the VM {{(pid=61923) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 904.016396] env[61923]: INFO nova.compute.manager [None req-b7a2d28f-39dc-4a35-bb16-94846b77bc41 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] [instance: c98c96a5-d18e-4519-996a-949ded16225f] Took 5.95 seconds to spawn the instance on the hypervisor. [ 904.016591] env[61923]: DEBUG nova.compute.manager [None req-b7a2d28f-39dc-4a35-bb16-94846b77bc41 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] [instance: c98c96a5-d18e-4519-996a-949ded16225f] Checking state {{(pid=61923) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 904.017390] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-146f8e74-2836-42bb-a35f-4e29c98678fb {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.146296] env[61923]: DEBUG oslo_concurrency.lockutils [None req-cb709b57-999a-41a8-b165-7029c0613fc5 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Acquiring lock "refresh_cache-03c1e8ea-a322-46fa-9b07-e283080a7871" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 904.146455] env[61923]: DEBUG oslo_concurrency.lockutils [None req-cb709b57-999a-41a8-b165-7029c0613fc5 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Acquired lock "refresh_cache-03c1e8ea-a322-46fa-9b07-e283080a7871" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 904.146673] env[61923]: DEBUG nova.network.neutron [None req-cb709b57-999a-41a8-b165-7029c0613fc5 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: 03c1e8ea-a322-46fa-9b07-e283080a7871] Building network info cache for instance {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 904.218311] env[61923]: DEBUG nova.network.neutron [None req-74d3449c-2b90-473c-a8b0-7be80c06daed tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] [instance: 6012d3b5-739c-4762-9bb4-09c51171dcd7] Successfully updated port: b395b183-1c30-42de-929f-690ea10d9d75 {{(pid=61923) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 904.225529] env[61923]: DEBUG oslo_vmware.api [None req-4a664699-f279-4048-aa87-1277ae6e408d tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] Task: {'id': task-1377744, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.237439} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 904.225742] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-4a664699-f279-4048-aa87-1277ae6e408d tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] Deleted the datastore file {{(pid=61923) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 904.225951] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-4a664699-f279-4048-aa87-1277ae6e408d tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] [instance: 263f0866-49af-4ab6-8132-d993642ebedb] Deleted contents of the VM from datastore datastore1 {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 904.226151] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-4a664699-f279-4048-aa87-1277ae6e408d tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] [instance: 263f0866-49af-4ab6-8132-d993642ebedb] Instance destroyed {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 904.226337] env[61923]: INFO nova.compute.manager [None req-4a664699-f279-4048-aa87-1277ae6e408d tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] [instance: 263f0866-49af-4ab6-8132-d993642ebedb] Took 1.20 seconds to destroy the instance on the hypervisor. [ 904.226580] env[61923]: DEBUG oslo.service.loopingcall [None req-4a664699-f279-4048-aa87-1277ae6e408d tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61923) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 904.226772] env[61923]: DEBUG nova.compute.manager [-] [instance: 263f0866-49af-4ab6-8132-d993642ebedb] Deallocating network for instance {{(pid=61923) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 904.226869] env[61923]: DEBUG nova.network.neutron [-] [instance: 263f0866-49af-4ab6-8132-d993642ebedb] deallocate_for_instance() {{(pid=61923) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 904.283992] env[61923]: DEBUG oslo_concurrency.lockutils [req-b00d9c40-6e79-4a30-8efc-3607eeb53d20 req-76a27ade-d77b-45c2-869d-af1e9d13a6e1 service nova] Releasing lock "refresh_cache-6012d3b5-739c-4762-9bb4-09c51171dcd7" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 904.389467] env[61923]: DEBUG oslo_vmware.api [None req-6cea46fb-73b5-4047-8297-b9cdc5a261d0 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Task: {'id': task-1377745, 'name': ReconfigVM_Task, 'duration_secs': 0.295155} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 904.389814] env[61923]: DEBUG nova.virt.vmwareapi.volumeops [None req-6cea46fb-73b5-4047-8297-b9cdc5a261d0 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] [instance: bac71328-3796-4a65-aa41-ccda93bc7926] Reconfigured VM instance instance-00000053 to attach disk [datastore1] bac71328-3796-4a65-aa41-ccda93bc7926/bac71328-3796-4a65-aa41-ccda93bc7926.vmdk or device None with type sparse {{(pid=61923) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 904.390592] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-bff4ce0e-b80a-448c-b11a-afc663bc4cbe {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.401370] env[61923]: DEBUG oslo_vmware.api [None req-e1b75445-6f25-4dcb-9846-d40076ac5b42 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Task: {'id': task-1377746, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.069997} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 904.402868] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-e1b75445-6f25-4dcb-9846-d40076ac5b42 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] [instance: 624e9207-7f24-4f45-8f52-0fc7a60c1e84] Extended root virtual disk {{(pid=61923) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 904.403266] env[61923]: DEBUG oslo_vmware.api [None req-6cea46fb-73b5-4047-8297-b9cdc5a261d0 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Waiting for the task: (returnval){ [ 904.403266] env[61923]: value = "task-1377747" [ 904.403266] env[61923]: _type = "Task" [ 904.403266] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 904.404065] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50f69fe1-7f56-4baa-9f5d-40eae2634afc {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.426165] env[61923]: DEBUG nova.virt.vmwareapi.volumeops [None req-e1b75445-6f25-4dcb-9846-d40076ac5b42 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] [instance: 624e9207-7f24-4f45-8f52-0fc7a60c1e84] Reconfiguring VM instance instance-00000055 to attach disk [datastore1] 624e9207-7f24-4f45-8f52-0fc7a60c1e84/624e9207-7f24-4f45-8f52-0fc7a60c1e84.vmdk or device None with type sparse {{(pid=61923) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 904.429785] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b90839da-2cb1-4b32-9d8d-021d7a75ae40 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.444446] env[61923]: DEBUG oslo_vmware.api [None req-6cea46fb-73b5-4047-8297-b9cdc5a261d0 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Task: {'id': task-1377747, 'name': Rename_Task} progress is 6%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.449875] env[61923]: DEBUG oslo_vmware.api [None req-e1b75445-6f25-4dcb-9846-d40076ac5b42 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Waiting for the task: (returnval){ [ 904.449875] env[61923]: value = "task-1377748" [ 904.449875] env[61923]: _type = "Task" [ 904.449875] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 904.459053] env[61923]: DEBUG oslo_vmware.api [None req-e1b75445-6f25-4dcb-9846-d40076ac5b42 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Task: {'id': task-1377748, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.488912] env[61923]: DEBUG oslo_concurrency.lockutils [None req-69c7da51-15c3-4307-92c9-fa8aea180f17 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] Acquiring lock "af3c317a-4007-4cea-a060-1e7dde5ce49e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 904.489361] env[61923]: DEBUG oslo_concurrency.lockutils [None req-69c7da51-15c3-4307-92c9-fa8aea180f17 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] Lock "af3c317a-4007-4cea-a060-1e7dde5ce49e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 904.489651] env[61923]: DEBUG oslo_concurrency.lockutils [None req-69c7da51-15c3-4307-92c9-fa8aea180f17 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] Acquiring lock "af3c317a-4007-4cea-a060-1e7dde5ce49e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 904.491070] env[61923]: DEBUG oslo_concurrency.lockutils [None req-69c7da51-15c3-4307-92c9-fa8aea180f17 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] Lock "af3c317a-4007-4cea-a060-1e7dde5ce49e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 904.491070] env[61923]: DEBUG oslo_concurrency.lockutils [None req-69c7da51-15c3-4307-92c9-fa8aea180f17 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] Lock "af3c317a-4007-4cea-a060-1e7dde5ce49e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 904.492952] env[61923]: INFO nova.compute.manager [None req-69c7da51-15c3-4307-92c9-fa8aea180f17 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] [instance: af3c317a-4007-4cea-a060-1e7dde5ce49e] Terminating instance [ 904.494869] env[61923]: DEBUG nova.compute.manager [None req-69c7da51-15c3-4307-92c9-fa8aea180f17 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] [instance: af3c317a-4007-4cea-a060-1e7dde5ce49e] Start destroying the instance on the hypervisor. {{(pid=61923) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 904.495175] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-69c7da51-15c3-4307-92c9-fa8aea180f17 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] [instance: af3c317a-4007-4cea-a060-1e7dde5ce49e] Destroying instance {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 904.496025] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa902f65-a341-4d6f-bb6c-5f89d9e09a34 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.507845] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-69c7da51-15c3-4307-92c9-fa8aea180f17 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] [instance: af3c317a-4007-4cea-a060-1e7dde5ce49e] Powering off the VM {{(pid=61923) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 904.507845] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-93e8f1db-53ab-40ca-9152-b98d436ef09a {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.513593] env[61923]: DEBUG oslo_vmware.api [None req-69c7da51-15c3-4307-92c9-fa8aea180f17 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] Waiting for the task: (returnval){ [ 904.513593] env[61923]: value = "task-1377749" [ 904.513593] env[61923]: _type = "Task" [ 904.513593] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 904.523796] env[61923]: DEBUG oslo_vmware.api [None req-69c7da51-15c3-4307-92c9-fa8aea180f17 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] Task: {'id': task-1377749, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.531972] env[61923]: INFO nova.compute.manager [None req-b7a2d28f-39dc-4a35-bb16-94846b77bc41 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] [instance: c98c96a5-d18e-4519-996a-949ded16225f] Took 14.52 seconds to build instance. [ 904.587467] env[61923]: DEBUG nova.compute.manager [None req-81989905-2c22-4589-a2f0-03b8d81c0d2f tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] [instance: 87ae37dc-d0d0-4b76-afdd-0ba3e8f6ce0b] Checking state {{(pid=61923) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 904.588393] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e5831f4-30b0-4a9b-8b5a-293a4f3d67f0 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.690262] env[61923]: DEBUG nova.network.neutron [None req-cb709b57-999a-41a8-b165-7029c0613fc5 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: 03c1e8ea-a322-46fa-9b07-e283080a7871] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 904.721695] env[61923]: DEBUG oslo_concurrency.lockutils [None req-74d3449c-2b90-473c-a8b0-7be80c06daed tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Acquiring lock "refresh_cache-6012d3b5-739c-4762-9bb4-09c51171dcd7" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 904.721894] env[61923]: DEBUG oslo_concurrency.lockutils [None req-74d3449c-2b90-473c-a8b0-7be80c06daed tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Acquired lock "refresh_cache-6012d3b5-739c-4762-9bb4-09c51171dcd7" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 904.722095] env[61923]: DEBUG nova.network.neutron [None req-74d3449c-2b90-473c-a8b0-7be80c06daed tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] [instance: 6012d3b5-739c-4762-9bb4-09c51171dcd7] Building network info cache for instance {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 904.744047] env[61923]: DEBUG oslo_concurrency.lockutils [None req-697de1ff-d6e2-4a1e-86f8-0419a0e8777a tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Acquiring lock "59198f4d-4dde-4eaf-9f6c-a962cbe53c6e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 904.744311] env[61923]: DEBUG oslo_concurrency.lockutils [None req-697de1ff-d6e2-4a1e-86f8-0419a0e8777a tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Lock "59198f4d-4dde-4eaf-9f6c-a962cbe53c6e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 904.744522] env[61923]: DEBUG oslo_concurrency.lockutils [None req-697de1ff-d6e2-4a1e-86f8-0419a0e8777a tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Acquiring lock "59198f4d-4dde-4eaf-9f6c-a962cbe53c6e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 904.744759] env[61923]: DEBUG oslo_concurrency.lockutils [None req-697de1ff-d6e2-4a1e-86f8-0419a0e8777a tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Lock "59198f4d-4dde-4eaf-9f6c-a962cbe53c6e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 904.744968] env[61923]: DEBUG oslo_concurrency.lockutils [None req-697de1ff-d6e2-4a1e-86f8-0419a0e8777a tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Lock "59198f4d-4dde-4eaf-9f6c-a962cbe53c6e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 904.747385] env[61923]: INFO nova.compute.manager [None req-697de1ff-d6e2-4a1e-86f8-0419a0e8777a tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] [instance: 59198f4d-4dde-4eaf-9f6c-a962cbe53c6e] Terminating instance [ 904.749402] env[61923]: DEBUG nova.compute.manager [None req-697de1ff-d6e2-4a1e-86f8-0419a0e8777a tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] [instance: 59198f4d-4dde-4eaf-9f6c-a962cbe53c6e] Start destroying the instance on the hypervisor. {{(pid=61923) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 904.749601] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-697de1ff-d6e2-4a1e-86f8-0419a0e8777a tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] [instance: 59198f4d-4dde-4eaf-9f6c-a962cbe53c6e] Destroying instance {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 904.750891] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-668e9a6a-18f0-4493-8956-3d9a2f2be0f1 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.760367] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-697de1ff-d6e2-4a1e-86f8-0419a0e8777a tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] [instance: 59198f4d-4dde-4eaf-9f6c-a962cbe53c6e] Powering off the VM {{(pid=61923) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 904.760367] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-65119362-c2a9-42e7-8d7b-3b2d0941b72e {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.766790] env[61923]: DEBUG oslo_vmware.api [None req-697de1ff-d6e2-4a1e-86f8-0419a0e8777a tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Waiting for the task: (returnval){ [ 904.766790] env[61923]: value = "task-1377750" [ 904.766790] env[61923]: _type = "Task" [ 904.766790] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 904.778720] env[61923]: DEBUG oslo_vmware.api [None req-697de1ff-d6e2-4a1e-86f8-0419a0e8777a tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Task: {'id': task-1377750, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.869266] env[61923]: DEBUG nova.network.neutron [None req-cb709b57-999a-41a8-b165-7029c0613fc5 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: 03c1e8ea-a322-46fa-9b07-e283080a7871] Updating instance_info_cache with network_info: [{"id": "37dd225f-bb7f-471b-8484-5387c0e22183", "address": "fa:16:3e:d5:a8:cf", "network": {"id": "d1c38272-0461-43e5-93ae-3964346f77d0", "bridge": "br-int", "label": "tempest-ServersTestJSON-151839165-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "69cc941eb6dd4780ac12aa29656c37f7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8ba3bd22-c936-470e-89bd-b3a5587e87a0", "external-id": "nsx-vlan-transportzone-605", "segmentation_id": 605, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap37dd225f-bb", "ovs_interfaceid": "37dd225f-bb7f-471b-8484-5387c0e22183", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 904.917995] env[61923]: DEBUG oslo_vmware.api [None req-6cea46fb-73b5-4047-8297-b9cdc5a261d0 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Task: {'id': task-1377747, 'name': Rename_Task, 'duration_secs': 0.181391} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 904.918327] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-6cea46fb-73b5-4047-8297-b9cdc5a261d0 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] [instance: bac71328-3796-4a65-aa41-ccda93bc7926] Powering on the VM {{(pid=61923) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 904.918602] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-29e74329-230b-4b7e-a4ec-b3b3eeaeeb2f {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.925526] env[61923]: DEBUG oslo_vmware.api [None req-6cea46fb-73b5-4047-8297-b9cdc5a261d0 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Waiting for the task: (returnval){ [ 904.925526] env[61923]: value = "task-1377751" [ 904.925526] env[61923]: _type = "Task" [ 904.925526] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 904.934684] env[61923]: DEBUG oslo_vmware.api [None req-6cea46fb-73b5-4047-8297-b9cdc5a261d0 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Task: {'id': task-1377751, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.963057] env[61923]: DEBUG oslo_vmware.api [None req-e1b75445-6f25-4dcb-9846-d40076ac5b42 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Task: {'id': task-1377748, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.025312] env[61923]: DEBUG oslo_vmware.api [None req-69c7da51-15c3-4307-92c9-fa8aea180f17 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] Task: {'id': task-1377749, 'name': PowerOffVM_Task, 'duration_secs': 0.365484} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 905.025639] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-69c7da51-15c3-4307-92c9-fa8aea180f17 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] [instance: af3c317a-4007-4cea-a060-1e7dde5ce49e] Powered off the VM {{(pid=61923) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 905.026034] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-69c7da51-15c3-4307-92c9-fa8aea180f17 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] [instance: af3c317a-4007-4cea-a060-1e7dde5ce49e] Unregistering the VM {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 905.026307] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-312d1c35-b73a-4df8-b294-982d5000ad0a {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.034509] env[61923]: DEBUG oslo_concurrency.lockutils [None req-b7a2d28f-39dc-4a35-bb16-94846b77bc41 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Lock "c98c96a5-d18e-4519-996a-949ded16225f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 16.030s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 905.045280] env[61923]: DEBUG nova.compute.manager [req-921b1676-2d6b-43e1-b976-ca5c3c9dc5ae req-602c27ca-8a3d-4849-b1bd-0cac7b9b9266 service nova] [instance: 6012d3b5-739c-4762-9bb4-09c51171dcd7] Received event network-vif-plugged-b395b183-1c30-42de-929f-690ea10d9d75 {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 905.045522] env[61923]: DEBUG oslo_concurrency.lockutils [req-921b1676-2d6b-43e1-b976-ca5c3c9dc5ae req-602c27ca-8a3d-4849-b1bd-0cac7b9b9266 service nova] Acquiring lock "6012d3b5-739c-4762-9bb4-09c51171dcd7-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 905.045703] env[61923]: DEBUG oslo_concurrency.lockutils [req-921b1676-2d6b-43e1-b976-ca5c3c9dc5ae req-602c27ca-8a3d-4849-b1bd-0cac7b9b9266 service nova] Lock "6012d3b5-739c-4762-9bb4-09c51171dcd7-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 905.045877] env[61923]: DEBUG oslo_concurrency.lockutils [req-921b1676-2d6b-43e1-b976-ca5c3c9dc5ae req-602c27ca-8a3d-4849-b1bd-0cac7b9b9266 service nova] Lock "6012d3b5-739c-4762-9bb4-09c51171dcd7-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 905.046198] env[61923]: DEBUG nova.compute.manager [req-921b1676-2d6b-43e1-b976-ca5c3c9dc5ae req-602c27ca-8a3d-4849-b1bd-0cac7b9b9266 service nova] [instance: 6012d3b5-739c-4762-9bb4-09c51171dcd7] No waiting events found dispatching network-vif-plugged-b395b183-1c30-42de-929f-690ea10d9d75 {{(pid=61923) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 905.046418] env[61923]: WARNING nova.compute.manager [req-921b1676-2d6b-43e1-b976-ca5c3c9dc5ae req-602c27ca-8a3d-4849-b1bd-0cac7b9b9266 service nova] [instance: 6012d3b5-739c-4762-9bb4-09c51171dcd7] Received unexpected event network-vif-plugged-b395b183-1c30-42de-929f-690ea10d9d75 for instance with vm_state active and task_state None. [ 905.046594] env[61923]: DEBUG nova.compute.manager [req-921b1676-2d6b-43e1-b976-ca5c3c9dc5ae req-602c27ca-8a3d-4849-b1bd-0cac7b9b9266 service nova] [instance: 6012d3b5-739c-4762-9bb4-09c51171dcd7] Received event network-changed-b395b183-1c30-42de-929f-690ea10d9d75 {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 905.046758] env[61923]: DEBUG nova.compute.manager [req-921b1676-2d6b-43e1-b976-ca5c3c9dc5ae req-602c27ca-8a3d-4849-b1bd-0cac7b9b9266 service nova] [instance: 6012d3b5-739c-4762-9bb4-09c51171dcd7] Refreshing instance network info cache due to event network-changed-b395b183-1c30-42de-929f-690ea10d9d75. {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 905.046931] env[61923]: DEBUG oslo_concurrency.lockutils [req-921b1676-2d6b-43e1-b976-ca5c3c9dc5ae req-602c27ca-8a3d-4849-b1bd-0cac7b9b9266 service nova] Acquiring lock "refresh_cache-6012d3b5-739c-4762-9bb4-09c51171dcd7" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 905.099856] env[61923]: INFO nova.compute.manager [None req-81989905-2c22-4589-a2f0-03b8d81c0d2f tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] [instance: 87ae37dc-d0d0-4b76-afdd-0ba3e8f6ce0b] instance snapshotting [ 905.103598] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1c5399d-f23a-4987-9754-fbb3b89b5758 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.124304] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f38cca17-3eae-4f03-97b3-f6deca9ded46 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.166039] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-69c7da51-15c3-4307-92c9-fa8aea180f17 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] [instance: af3c317a-4007-4cea-a060-1e7dde5ce49e] Unregistered the VM {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 905.166039] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-69c7da51-15c3-4307-92c9-fa8aea180f17 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] [instance: af3c317a-4007-4cea-a060-1e7dde5ce49e] Deleting contents of the VM from datastore datastore1 {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 905.166039] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-69c7da51-15c3-4307-92c9-fa8aea180f17 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] Deleting the datastore file [datastore1] af3c317a-4007-4cea-a060-1e7dde5ce49e {{(pid=61923) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 905.166314] env[61923]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f961b461-93de-43be-8b25-2e2b072ca613 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.174084] env[61923]: DEBUG oslo_vmware.api [None req-69c7da51-15c3-4307-92c9-fa8aea180f17 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] Waiting for the task: (returnval){ [ 905.174084] env[61923]: value = "task-1377753" [ 905.174084] env[61923]: _type = "Task" [ 905.174084] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 905.183983] env[61923]: DEBUG oslo_vmware.api [None req-69c7da51-15c3-4307-92c9-fa8aea180f17 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] Task: {'id': task-1377753, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.203723] env[61923]: DEBUG nova.network.neutron [-] [instance: 263f0866-49af-4ab6-8132-d993642ebedb] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 905.267116] env[61923]: WARNING nova.network.neutron [None req-74d3449c-2b90-473c-a8b0-7be80c06daed tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] [instance: 6012d3b5-739c-4762-9bb4-09c51171dcd7] a8b7e1b4-1ec6-4226-bdd6-535dbce450d2 already exists in list: networks containing: ['a8b7e1b4-1ec6-4226-bdd6-535dbce450d2']. ignoring it [ 905.279668] env[61923]: DEBUG oslo_vmware.api [None req-697de1ff-d6e2-4a1e-86f8-0419a0e8777a tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Task: {'id': task-1377750, 'name': PowerOffVM_Task, 'duration_secs': 0.228371} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 905.280323] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-697de1ff-d6e2-4a1e-86f8-0419a0e8777a tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] [instance: 59198f4d-4dde-4eaf-9f6c-a962cbe53c6e] Powered off the VM {{(pid=61923) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 905.280538] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-697de1ff-d6e2-4a1e-86f8-0419a0e8777a tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] [instance: 59198f4d-4dde-4eaf-9f6c-a962cbe53c6e] Unregistering the VM {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 905.280825] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-5d860489-6a29-47ae-a64e-8b5d318d7bbd {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.343880] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-697de1ff-d6e2-4a1e-86f8-0419a0e8777a tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] [instance: 59198f4d-4dde-4eaf-9f6c-a962cbe53c6e] Unregistered the VM {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 905.344164] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-697de1ff-d6e2-4a1e-86f8-0419a0e8777a tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] [instance: 59198f4d-4dde-4eaf-9f6c-a962cbe53c6e] Deleting contents of the VM from datastore datastore2 {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 905.344387] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-697de1ff-d6e2-4a1e-86f8-0419a0e8777a tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Deleting the datastore file [datastore2] 59198f4d-4dde-4eaf-9f6c-a962cbe53c6e {{(pid=61923) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 905.345053] env[61923]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-509ce01a-7c0f-4e52-9734-b1358c51dc2d {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.352615] env[61923]: DEBUG oslo_vmware.api [None req-697de1ff-d6e2-4a1e-86f8-0419a0e8777a tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Waiting for the task: (returnval){ [ 905.352615] env[61923]: value = "task-1377755" [ 905.352615] env[61923]: _type = "Task" [ 905.352615] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 905.362562] env[61923]: DEBUG oslo_vmware.api [None req-697de1ff-d6e2-4a1e-86f8-0419a0e8777a tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Task: {'id': task-1377755, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.372406] env[61923]: DEBUG oslo_concurrency.lockutils [None req-cb709b57-999a-41a8-b165-7029c0613fc5 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Releasing lock "refresh_cache-03c1e8ea-a322-46fa-9b07-e283080a7871" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 905.372761] env[61923]: DEBUG nova.compute.manager [None req-cb709b57-999a-41a8-b165-7029c0613fc5 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: 03c1e8ea-a322-46fa-9b07-e283080a7871] Instance network_info: |[{"id": "37dd225f-bb7f-471b-8484-5387c0e22183", "address": "fa:16:3e:d5:a8:cf", "network": {"id": "d1c38272-0461-43e5-93ae-3964346f77d0", "bridge": "br-int", "label": "tempest-ServersTestJSON-151839165-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "69cc941eb6dd4780ac12aa29656c37f7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8ba3bd22-c936-470e-89bd-b3a5587e87a0", "external-id": "nsx-vlan-transportzone-605", "segmentation_id": 605, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap37dd225f-bb", "ovs_interfaceid": "37dd225f-bb7f-471b-8484-5387c0e22183", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61923) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 905.373217] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-cb709b57-999a-41a8-b165-7029c0613fc5 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: 03c1e8ea-a322-46fa-9b07-e283080a7871] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:d5:a8:cf', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '8ba3bd22-c936-470e-89bd-b3a5587e87a0', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '37dd225f-bb7f-471b-8484-5387c0e22183', 'vif_model': 'vmxnet3'}] {{(pid=61923) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 905.382582] env[61923]: DEBUG oslo.service.loopingcall [None req-cb709b57-999a-41a8-b165-7029c0613fc5 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61923) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 905.382871] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 03c1e8ea-a322-46fa-9b07-e283080a7871] Creating VM on the ESX host {{(pid=61923) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 905.383216] env[61923]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-44395c51-cd70-4696-93c6-662feb3dfb88 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.402732] env[61923]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 905.402732] env[61923]: value = "task-1377756" [ 905.402732] env[61923]: _type = "Task" [ 905.402732] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 905.414672] env[61923]: DEBUG oslo_vmware.api [-] Task: {'id': task-1377756, 'name': CreateVM_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.428687] env[61923]: DEBUG oslo_concurrency.lockutils [None req-9a89d0be-1b30-454c-a15f-f7048c8714c5 tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Acquiring lock "62c75dfd-6046-4de1-b1e3-13a307af1394" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 905.429019] env[61923]: DEBUG oslo_concurrency.lockutils [None req-9a89d0be-1b30-454c-a15f-f7048c8714c5 tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Lock "62c75dfd-6046-4de1-b1e3-13a307af1394" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 905.447819] env[61923]: DEBUG oslo_vmware.api [None req-6cea46fb-73b5-4047-8297-b9cdc5a261d0 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Task: {'id': task-1377751, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.461300] env[61923]: DEBUG oslo_vmware.api [None req-e1b75445-6f25-4dcb-9846-d40076ac5b42 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Task: {'id': task-1377748, 'name': ReconfigVM_Task, 'duration_secs': 0.585967} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 905.461845] env[61923]: DEBUG nova.virt.vmwareapi.volumeops [None req-e1b75445-6f25-4dcb-9846-d40076ac5b42 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] [instance: 624e9207-7f24-4f45-8f52-0fc7a60c1e84] Reconfigured VM instance instance-00000055 to attach disk [datastore1] 624e9207-7f24-4f45-8f52-0fc7a60c1e84/624e9207-7f24-4f45-8f52-0fc7a60c1e84.vmdk or device None with type sparse {{(pid=61923) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 905.462596] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b80910de-534b-44db-b8f8-d2299c5f4339 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.469876] env[61923]: DEBUG oslo_vmware.api [None req-e1b75445-6f25-4dcb-9846-d40076ac5b42 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Waiting for the task: (returnval){ [ 905.469876] env[61923]: value = "task-1377757" [ 905.469876] env[61923]: _type = "Task" [ 905.469876] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 905.484906] env[61923]: DEBUG oslo_vmware.api [None req-e1b75445-6f25-4dcb-9846-d40076ac5b42 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Task: {'id': task-1377757, 'name': Rename_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.637705] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-81989905-2c22-4589-a2f0-03b8d81c0d2f tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] [instance: 87ae37dc-d0d0-4b76-afdd-0ba3e8f6ce0b] Creating Snapshot of the VM instance {{(pid=61923) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 905.638070] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-70ba3953-d16d-42ba-b480-567a39c5e08c {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.641360] env[61923]: DEBUG nova.network.neutron [None req-74d3449c-2b90-473c-a8b0-7be80c06daed tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] [instance: 6012d3b5-739c-4762-9bb4-09c51171dcd7] Updating instance_info_cache with network_info: [{"id": "269a41b0-75e7-46af-9497-8eae3102df11", "address": "fa:16:3e:e1:02:b8", "network": {"id": "a8b7e1b4-1ec6-4226-bdd6-535dbce450d2", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1928758125-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.146", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "912300a09b80452e85e3cd13c4a644cc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7806fe18-2b89-4386-87b1-f22876f82af2", "external-id": "nsx-vlan-transportzone-727", "segmentation_id": 727, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap269a41b0-75", "ovs_interfaceid": "269a41b0-75e7-46af-9497-8eae3102df11", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "b395b183-1c30-42de-929f-690ea10d9d75", "address": "fa:16:3e:7a:d5:2a", "network": {"id": "a8b7e1b4-1ec6-4226-bdd6-535dbce450d2", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1928758125-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "912300a09b80452e85e3cd13c4a644cc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7806fe18-2b89-4386-87b1-f22876f82af2", "external-id": "nsx-vlan-transportzone-727", "segmentation_id": 727, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb395b183-1c", "ovs_interfaceid": "b395b183-1c30-42de-929f-690ea10d9d75", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 905.653985] env[61923]: DEBUG oslo_vmware.api [None req-81989905-2c22-4589-a2f0-03b8d81c0d2f tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Waiting for the task: (returnval){ [ 905.653985] env[61923]: value = "task-1377758" [ 905.653985] env[61923]: _type = "Task" [ 905.653985] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 905.666063] env[61923]: DEBUG oslo_vmware.api [None req-81989905-2c22-4589-a2f0-03b8d81c0d2f tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Task: {'id': task-1377758, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.682242] env[61923]: DEBUG nova.compute.manager [req-dac6f300-f8e7-4444-8300-a12a7a7e2a0c req-f2ced70c-4e7e-44af-8fbc-410dd0c878b2 service nova] [instance: 03c1e8ea-a322-46fa-9b07-e283080a7871] Received event network-changed-37dd225f-bb7f-471b-8484-5387c0e22183 {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 905.682511] env[61923]: DEBUG nova.compute.manager [req-dac6f300-f8e7-4444-8300-a12a7a7e2a0c req-f2ced70c-4e7e-44af-8fbc-410dd0c878b2 service nova] [instance: 03c1e8ea-a322-46fa-9b07-e283080a7871] Refreshing instance network info cache due to event network-changed-37dd225f-bb7f-471b-8484-5387c0e22183. {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 905.682779] env[61923]: DEBUG oslo_concurrency.lockutils [req-dac6f300-f8e7-4444-8300-a12a7a7e2a0c req-f2ced70c-4e7e-44af-8fbc-410dd0c878b2 service nova] Acquiring lock "refresh_cache-03c1e8ea-a322-46fa-9b07-e283080a7871" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 905.682952] env[61923]: DEBUG oslo_concurrency.lockutils [req-dac6f300-f8e7-4444-8300-a12a7a7e2a0c req-f2ced70c-4e7e-44af-8fbc-410dd0c878b2 service nova] Acquired lock "refresh_cache-03c1e8ea-a322-46fa-9b07-e283080a7871" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 905.683138] env[61923]: DEBUG nova.network.neutron [req-dac6f300-f8e7-4444-8300-a12a7a7e2a0c req-f2ced70c-4e7e-44af-8fbc-410dd0c878b2 service nova] [instance: 03c1e8ea-a322-46fa-9b07-e283080a7871] Refreshing network info cache for port 37dd225f-bb7f-471b-8484-5387c0e22183 {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 905.689516] env[61923]: DEBUG oslo_vmware.api [None req-69c7da51-15c3-4307-92c9-fa8aea180f17 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] Task: {'id': task-1377753, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.254877} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 905.689753] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-69c7da51-15c3-4307-92c9-fa8aea180f17 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] Deleted the datastore file {{(pid=61923) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 905.689936] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-69c7da51-15c3-4307-92c9-fa8aea180f17 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] [instance: af3c317a-4007-4cea-a060-1e7dde5ce49e] Deleted contents of the VM from datastore datastore1 {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 905.690128] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-69c7da51-15c3-4307-92c9-fa8aea180f17 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] [instance: af3c317a-4007-4cea-a060-1e7dde5ce49e] Instance destroyed {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 905.690310] env[61923]: INFO nova.compute.manager [None req-69c7da51-15c3-4307-92c9-fa8aea180f17 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] [instance: af3c317a-4007-4cea-a060-1e7dde5ce49e] Took 1.20 seconds to destroy the instance on the hypervisor. [ 905.690565] env[61923]: DEBUG oslo.service.loopingcall [None req-69c7da51-15c3-4307-92c9-fa8aea180f17 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61923) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 905.690760] env[61923]: DEBUG nova.compute.manager [-] [instance: af3c317a-4007-4cea-a060-1e7dde5ce49e] Deallocating network for instance {{(pid=61923) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 905.690908] env[61923]: DEBUG nova.network.neutron [-] [instance: af3c317a-4007-4cea-a060-1e7dde5ce49e] deallocate_for_instance() {{(pid=61923) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 905.706507] env[61923]: INFO nova.compute.manager [-] [instance: 263f0866-49af-4ab6-8132-d993642ebedb] Took 1.48 seconds to deallocate network for instance. [ 905.863997] env[61923]: DEBUG oslo_vmware.api [None req-697de1ff-d6e2-4a1e-86f8-0419a0e8777a tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Task: {'id': task-1377755, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.218719} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 905.865772] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-697de1ff-d6e2-4a1e-86f8-0419a0e8777a tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Deleted the datastore file {{(pid=61923) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 905.865975] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-697de1ff-d6e2-4a1e-86f8-0419a0e8777a tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] [instance: 59198f4d-4dde-4eaf-9f6c-a962cbe53c6e] Deleted contents of the VM from datastore datastore2 {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 905.866182] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-697de1ff-d6e2-4a1e-86f8-0419a0e8777a tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] [instance: 59198f4d-4dde-4eaf-9f6c-a962cbe53c6e] Instance destroyed {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 905.866389] env[61923]: INFO nova.compute.manager [None req-697de1ff-d6e2-4a1e-86f8-0419a0e8777a tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] [instance: 59198f4d-4dde-4eaf-9f6c-a962cbe53c6e] Took 1.12 seconds to destroy the instance on the hypervisor. [ 905.866736] env[61923]: DEBUG oslo.service.loopingcall [None req-697de1ff-d6e2-4a1e-86f8-0419a0e8777a tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61923) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 905.866942] env[61923]: DEBUG nova.compute.manager [-] [instance: 59198f4d-4dde-4eaf-9f6c-a962cbe53c6e] Deallocating network for instance {{(pid=61923) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 905.867050] env[61923]: DEBUG nova.network.neutron [-] [instance: 59198f4d-4dde-4eaf-9f6c-a962cbe53c6e] deallocate_for_instance() {{(pid=61923) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 905.912498] env[61923]: DEBUG oslo_vmware.api [-] Task: {'id': task-1377756, 'name': CreateVM_Task, 'duration_secs': 0.387301} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 905.912689] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 03c1e8ea-a322-46fa-9b07-e283080a7871] Created VM on the ESX host {{(pid=61923) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 905.913385] env[61923]: DEBUG oslo_concurrency.lockutils [None req-cb709b57-999a-41a8-b165-7029c0613fc5 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 905.913557] env[61923]: DEBUG oslo_concurrency.lockutils [None req-cb709b57-999a-41a8-b165-7029c0613fc5 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 905.913944] env[61923]: DEBUG oslo_concurrency.lockutils [None req-cb709b57-999a-41a8-b165-7029c0613fc5 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 905.914150] env[61923]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-52665ec0-26f5-44f8-adb4-665212217c63 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.919578] env[61923]: DEBUG oslo_vmware.api [None req-cb709b57-999a-41a8-b165-7029c0613fc5 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Waiting for the task: (returnval){ [ 905.919578] env[61923]: value = "session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]525d6f29-0cd2-5759-16b8-a2a63ce6fc0c" [ 905.919578] env[61923]: _type = "Task" [ 905.919578] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 905.927331] env[61923]: DEBUG oslo_vmware.api [None req-cb709b57-999a-41a8-b165-7029c0613fc5 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]525d6f29-0cd2-5759-16b8-a2a63ce6fc0c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.936373] env[61923]: DEBUG nova.compute.manager [None req-9a89d0be-1b30-454c-a15f-f7048c8714c5 tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] [instance: 62c75dfd-6046-4de1-b1e3-13a307af1394] Starting instance... {{(pid=61923) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 905.939054] env[61923]: DEBUG oslo_vmware.api [None req-6cea46fb-73b5-4047-8297-b9cdc5a261d0 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Task: {'id': task-1377751, 'name': PowerOnVM_Task, 'duration_secs': 0.586781} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 905.941531] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-6cea46fb-73b5-4047-8297-b9cdc5a261d0 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] [instance: bac71328-3796-4a65-aa41-ccda93bc7926] Powered on the VM {{(pid=61923) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 905.941776] env[61923]: INFO nova.compute.manager [None req-6cea46fb-73b5-4047-8297-b9cdc5a261d0 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] [instance: bac71328-3796-4a65-aa41-ccda93bc7926] Took 10.93 seconds to spawn the instance on the hypervisor. [ 905.941973] env[61923]: DEBUG nova.compute.manager [None req-6cea46fb-73b5-4047-8297-b9cdc5a261d0 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] [instance: bac71328-3796-4a65-aa41-ccda93bc7926] Checking state {{(pid=61923) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 905.942772] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4fd68d53-d67e-4fe6-bd8e-8e5bc452bd15 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.980842] env[61923]: DEBUG oslo_vmware.api [None req-e1b75445-6f25-4dcb-9846-d40076ac5b42 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Task: {'id': task-1377757, 'name': Rename_Task, 'duration_secs': 0.212511} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 905.981263] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-e1b75445-6f25-4dcb-9846-d40076ac5b42 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] [instance: 624e9207-7f24-4f45-8f52-0fc7a60c1e84] Powering on the VM {{(pid=61923) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 905.981593] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-27dacb91-42b4-41aa-8a02-26f6921e60f2 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.988094] env[61923]: DEBUG oslo_vmware.api [None req-e1b75445-6f25-4dcb-9846-d40076ac5b42 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Waiting for the task: (returnval){ [ 905.988094] env[61923]: value = "task-1377759" [ 905.988094] env[61923]: _type = "Task" [ 905.988094] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 905.996897] env[61923]: DEBUG oslo_vmware.api [None req-e1b75445-6f25-4dcb-9846-d40076ac5b42 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Task: {'id': task-1377759, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.148043] env[61923]: DEBUG oslo_concurrency.lockutils [None req-74d3449c-2b90-473c-a8b0-7be80c06daed tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Releasing lock "refresh_cache-6012d3b5-739c-4762-9bb4-09c51171dcd7" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 906.148043] env[61923]: DEBUG oslo_concurrency.lockutils [None req-74d3449c-2b90-473c-a8b0-7be80c06daed tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Acquiring lock "6012d3b5-739c-4762-9bb4-09c51171dcd7" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 906.148043] env[61923]: DEBUG oslo_concurrency.lockutils [None req-74d3449c-2b90-473c-a8b0-7be80c06daed tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Acquired lock "6012d3b5-739c-4762-9bb4-09c51171dcd7" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 906.148043] env[61923]: DEBUG oslo_concurrency.lockutils [req-921b1676-2d6b-43e1-b976-ca5c3c9dc5ae req-602c27ca-8a3d-4849-b1bd-0cac7b9b9266 service nova] Acquired lock "refresh_cache-6012d3b5-739c-4762-9bb4-09c51171dcd7" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 906.148043] env[61923]: DEBUG nova.network.neutron [req-921b1676-2d6b-43e1-b976-ca5c3c9dc5ae req-602c27ca-8a3d-4849-b1bd-0cac7b9b9266 service nova] [instance: 6012d3b5-739c-4762-9bb4-09c51171dcd7] Refreshing network info cache for port b395b183-1c30-42de-929f-690ea10d9d75 {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 906.148043] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0eb9d309-3ebe-455e-acb0-25e1b8267e42 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.172270] env[61923]: DEBUG nova.virt.hardware [None req-74d3449c-2b90-473c-a8b0-7be80c06daed tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-29T20:07:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 906.172512] env[61923]: DEBUG nova.virt.hardware [None req-74d3449c-2b90-473c-a8b0-7be80c06daed tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Flavor limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 906.172670] env[61923]: DEBUG nova.virt.hardware [None req-74d3449c-2b90-473c-a8b0-7be80c06daed tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Image limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 906.172866] env[61923]: DEBUG nova.virt.hardware [None req-74d3449c-2b90-473c-a8b0-7be80c06daed tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Flavor pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 906.173043] env[61923]: DEBUG nova.virt.hardware [None req-74d3449c-2b90-473c-a8b0-7be80c06daed tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Image pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 906.173215] env[61923]: DEBUG nova.virt.hardware [None req-74d3449c-2b90-473c-a8b0-7be80c06daed tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 906.173428] env[61923]: DEBUG nova.virt.hardware [None req-74d3449c-2b90-473c-a8b0-7be80c06daed tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 906.173587] env[61923]: DEBUG nova.virt.hardware [None req-74d3449c-2b90-473c-a8b0-7be80c06daed tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 906.173764] env[61923]: DEBUG nova.virt.hardware [None req-74d3449c-2b90-473c-a8b0-7be80c06daed tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Got 1 possible topologies {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 906.173946] env[61923]: DEBUG nova.virt.hardware [None req-74d3449c-2b90-473c-a8b0-7be80c06daed tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 906.174147] env[61923]: DEBUG nova.virt.hardware [None req-74d3449c-2b90-473c-a8b0-7be80c06daed tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 906.181076] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-74d3449c-2b90-473c-a8b0-7be80c06daed tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] [instance: 6012d3b5-739c-4762-9bb4-09c51171dcd7] Reconfiguring VM to attach interface {{(pid=61923) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1923}} [ 906.181290] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-33f80e65-3a24-42e3-830a-a27398c8a64b {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.201639] env[61923]: DEBUG oslo_vmware.api [None req-81989905-2c22-4589-a2f0-03b8d81c0d2f tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Task: {'id': task-1377758, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.208368] env[61923]: DEBUG oslo_vmware.api [None req-74d3449c-2b90-473c-a8b0-7be80c06daed tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Waiting for the task: (returnval){ [ 906.208368] env[61923]: value = "task-1377760" [ 906.208368] env[61923]: _type = "Task" [ 906.208368] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 906.212651] env[61923]: DEBUG oslo_concurrency.lockutils [None req-4a664699-f279-4048-aa87-1277ae6e408d tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 906.212907] env[61923]: DEBUG oslo_concurrency.lockutils [None req-4a664699-f279-4048-aa87-1277ae6e408d tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 906.213136] env[61923]: DEBUG nova.objects.instance [None req-4a664699-f279-4048-aa87-1277ae6e408d tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] Lazy-loading 'resources' on Instance uuid 263f0866-49af-4ab6-8132-d993642ebedb {{(pid=61923) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 906.220414] env[61923]: DEBUG oslo_vmware.api [None req-74d3449c-2b90-473c-a8b0-7be80c06daed tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Task: {'id': task-1377760, 'name': ReconfigVM_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.432035] env[61923]: DEBUG oslo_vmware.api [None req-cb709b57-999a-41a8-b165-7029c0613fc5 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]525d6f29-0cd2-5759-16b8-a2a63ce6fc0c, 'name': SearchDatastore_Task, 'duration_secs': 0.016943} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 906.433052] env[61923]: DEBUG nova.network.neutron [req-dac6f300-f8e7-4444-8300-a12a7a7e2a0c req-f2ced70c-4e7e-44af-8fbc-410dd0c878b2 service nova] [instance: 03c1e8ea-a322-46fa-9b07-e283080a7871] Updated VIF entry in instance network info cache for port 37dd225f-bb7f-471b-8484-5387c0e22183. {{(pid=61923) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 906.433425] env[61923]: DEBUG nova.network.neutron [req-dac6f300-f8e7-4444-8300-a12a7a7e2a0c req-f2ced70c-4e7e-44af-8fbc-410dd0c878b2 service nova] [instance: 03c1e8ea-a322-46fa-9b07-e283080a7871] Updating instance_info_cache with network_info: [{"id": "37dd225f-bb7f-471b-8484-5387c0e22183", "address": "fa:16:3e:d5:a8:cf", "network": {"id": "d1c38272-0461-43e5-93ae-3964346f77d0", "bridge": "br-int", "label": "tempest-ServersTestJSON-151839165-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "69cc941eb6dd4780ac12aa29656c37f7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8ba3bd22-c936-470e-89bd-b3a5587e87a0", "external-id": "nsx-vlan-transportzone-605", "segmentation_id": 605, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap37dd225f-bb", "ovs_interfaceid": "37dd225f-bb7f-471b-8484-5387c0e22183", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 906.434816] env[61923]: DEBUG oslo_concurrency.lockutils [None req-cb709b57-999a-41a8-b165-7029c0613fc5 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 906.435087] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-cb709b57-999a-41a8-b165-7029c0613fc5 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: 03c1e8ea-a322-46fa-9b07-e283080a7871] Processing image 0f153f63-ae0a-45b1-b7f5-7f9b673c947f {{(pid=61923) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 906.435349] env[61923]: DEBUG oslo_concurrency.lockutils [None req-cb709b57-999a-41a8-b165-7029c0613fc5 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 906.435516] env[61923]: DEBUG oslo_concurrency.lockutils [None req-cb709b57-999a-41a8-b165-7029c0613fc5 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 906.435709] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-cb709b57-999a-41a8-b165-7029c0613fc5 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61923) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 906.435999] env[61923]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-8dda3fbf-4b51-45f6-9e5f-5bd9571ed982 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.448320] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-cb709b57-999a-41a8-b165-7029c0613fc5 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61923) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 906.448542] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-cb709b57-999a-41a8-b165-7029c0613fc5 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61923) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 906.449447] env[61923]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-01d5efd7-9bd5-40b2-8767-fdaf8fc86473 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.458893] env[61923]: DEBUG oslo_vmware.api [None req-cb709b57-999a-41a8-b165-7029c0613fc5 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Waiting for the task: (returnval){ [ 906.458893] env[61923]: value = "session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52c779ca-f80d-6608-99c3-71343c1336f7" [ 906.458893] env[61923]: _type = "Task" [ 906.458893] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 906.461426] env[61923]: INFO nova.compute.manager [None req-6cea46fb-73b5-4047-8297-b9cdc5a261d0 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] [instance: bac71328-3796-4a65-aa41-ccda93bc7926] Took 17.29 seconds to build instance. [ 906.474969] env[61923]: DEBUG oslo_vmware.api [None req-cb709b57-999a-41a8-b165-7029c0613fc5 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52c779ca-f80d-6608-99c3-71343c1336f7, 'name': SearchDatastore_Task, 'duration_secs': 0.010769} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 906.476155] env[61923]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-79b83600-4b70-495b-9953-3a6919cd83a5 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.483022] env[61923]: DEBUG oslo_vmware.api [None req-cb709b57-999a-41a8-b165-7029c0613fc5 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Waiting for the task: (returnval){ [ 906.483022] env[61923]: value = "session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52bbff61-e00c-50df-e59c-e9cbab0e420b" [ 906.483022] env[61923]: _type = "Task" [ 906.483022] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 906.484220] env[61923]: DEBUG oslo_concurrency.lockutils [None req-9a89d0be-1b30-454c-a15f-f7048c8714c5 tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 906.495677] env[61923]: DEBUG oslo_vmware.api [None req-cb709b57-999a-41a8-b165-7029c0613fc5 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52bbff61-e00c-50df-e59c-e9cbab0e420b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.502227] env[61923]: DEBUG oslo_vmware.api [None req-e1b75445-6f25-4dcb-9846-d40076ac5b42 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Task: {'id': task-1377759, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.646822] env[61923]: DEBUG nova.network.neutron [-] [instance: 59198f4d-4dde-4eaf-9f6c-a962cbe53c6e] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 906.667858] env[61923]: DEBUG oslo_vmware.api [None req-81989905-2c22-4589-a2f0-03b8d81c0d2f tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Task: {'id': task-1377758, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.724380] env[61923]: DEBUG oslo_vmware.api [None req-74d3449c-2b90-473c-a8b0-7be80c06daed tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Task: {'id': task-1377760, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.942093] env[61923]: DEBUG oslo_concurrency.lockutils [req-dac6f300-f8e7-4444-8300-a12a7a7e2a0c req-f2ced70c-4e7e-44af-8fbc-410dd0c878b2 service nova] Releasing lock "refresh_cache-03c1e8ea-a322-46fa-9b07-e283080a7871" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 906.946786] env[61923]: DEBUG nova.network.neutron [req-921b1676-2d6b-43e1-b976-ca5c3c9dc5ae req-602c27ca-8a3d-4849-b1bd-0cac7b9b9266 service nova] [instance: 6012d3b5-739c-4762-9bb4-09c51171dcd7] Updated VIF entry in instance network info cache for port b395b183-1c30-42de-929f-690ea10d9d75. {{(pid=61923) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 906.947304] env[61923]: DEBUG nova.network.neutron [req-921b1676-2d6b-43e1-b976-ca5c3c9dc5ae req-602c27ca-8a3d-4849-b1bd-0cac7b9b9266 service nova] [instance: 6012d3b5-739c-4762-9bb4-09c51171dcd7] Updating instance_info_cache with network_info: [{"id": "269a41b0-75e7-46af-9497-8eae3102df11", "address": "fa:16:3e:e1:02:b8", "network": {"id": "a8b7e1b4-1ec6-4226-bdd6-535dbce450d2", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1928758125-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.146", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "912300a09b80452e85e3cd13c4a644cc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7806fe18-2b89-4386-87b1-f22876f82af2", "external-id": "nsx-vlan-transportzone-727", "segmentation_id": 727, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap269a41b0-75", "ovs_interfaceid": "269a41b0-75e7-46af-9497-8eae3102df11", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "b395b183-1c30-42de-929f-690ea10d9d75", "address": "fa:16:3e:7a:d5:2a", "network": {"id": "a8b7e1b4-1ec6-4226-bdd6-535dbce450d2", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1928758125-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "912300a09b80452e85e3cd13c4a644cc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7806fe18-2b89-4386-87b1-f22876f82af2", "external-id": "nsx-vlan-transportzone-727", "segmentation_id": 727, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb395b183-1c", "ovs_interfaceid": "b395b183-1c30-42de-929f-690ea10d9d75", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 906.953142] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82bd5aea-a4d2-41f4-8c09-6a0ad007765f {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.961956] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41c2d919-5687-4b5f-880b-29a82d9ec507 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.968473] env[61923]: DEBUG oslo_concurrency.lockutils [None req-6cea46fb-73b5-4047-8297-b9cdc5a261d0 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Lock "bac71328-3796-4a65-aa41-ccda93bc7926" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 18.818s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 907.005306] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5a3aeb5-4b16-4870-a330-5030ff41dd6c {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.017809] env[61923]: DEBUG oslo_vmware.api [None req-cb709b57-999a-41a8-b165-7029c0613fc5 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52bbff61-e00c-50df-e59c-e9cbab0e420b, 'name': SearchDatastore_Task, 'duration_secs': 0.011048} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 907.018338] env[61923]: DEBUG oslo_vmware.api [None req-e1b75445-6f25-4dcb-9846-d40076ac5b42 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Task: {'id': task-1377759, 'name': PowerOnVM_Task, 'duration_secs': 0.60126} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 907.020488] env[61923]: DEBUG oslo_concurrency.lockutils [None req-cb709b57-999a-41a8-b165-7029c0613fc5 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 907.021706] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-cb709b57-999a-41a8-b165-7029c0613fc5 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk to [datastore1] 03c1e8ea-a322-46fa-9b07-e283080a7871/03c1e8ea-a322-46fa-9b07-e283080a7871.vmdk {{(pid=61923) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 907.021706] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-e1b75445-6f25-4dcb-9846-d40076ac5b42 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] [instance: 624e9207-7f24-4f45-8f52-0fc7a60c1e84] Powered on the VM {{(pid=61923) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 907.021885] env[61923]: INFO nova.compute.manager [None req-e1b75445-6f25-4dcb-9846-d40076ac5b42 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] [instance: 624e9207-7f24-4f45-8f52-0fc7a60c1e84] Took 6.54 seconds to spawn the instance on the hypervisor. [ 907.022196] env[61923]: DEBUG nova.compute.manager [None req-e1b75445-6f25-4dcb-9846-d40076ac5b42 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] [instance: 624e9207-7f24-4f45-8f52-0fc7a60c1e84] Checking state {{(pid=61923) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 907.022570] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-dcf7e681-87d7-4a8b-8ef1-50176be295b1 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.026182] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-152034c1-d327-41e7-9ad5-c9c6c8f04c38 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.029162] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef315d3c-5112-45b0-a2f1-4e2642270185 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.034176] env[61923]: DEBUG nova.network.neutron [-] [instance: af3c317a-4007-4cea-a060-1e7dde5ce49e] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 907.050039] env[61923]: DEBUG nova.compute.provider_tree [None req-4a664699-f279-4048-aa87-1277ae6e408d tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 907.054992] env[61923]: DEBUG oslo_vmware.api [None req-cb709b57-999a-41a8-b165-7029c0613fc5 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Waiting for the task: (returnval){ [ 907.054992] env[61923]: value = "task-1377761" [ 907.054992] env[61923]: _type = "Task" [ 907.054992] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 907.065721] env[61923]: DEBUG oslo_vmware.api [None req-cb709b57-999a-41a8-b165-7029c0613fc5 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Task: {'id': task-1377761, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.075119] env[61923]: DEBUG nova.compute.manager [req-628b94f1-0df7-4c6e-94b7-37bc816c352e req-082aab72-4681-4b2f-b2f4-75ce61e0921b service nova] [instance: 59198f4d-4dde-4eaf-9f6c-a962cbe53c6e] Received event network-vif-deleted-cf99020c-d7cc-4912-a021-8861628b7540 {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 907.116398] env[61923]: DEBUG oslo_concurrency.lockutils [None req-15de031b-9972-4bde-a9c6-0e58042e23f7 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Acquiring lock "bac71328-3796-4a65-aa41-ccda93bc7926" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 907.116692] env[61923]: DEBUG oslo_concurrency.lockutils [None req-15de031b-9972-4bde-a9c6-0e58042e23f7 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Lock "bac71328-3796-4a65-aa41-ccda93bc7926" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 907.116965] env[61923]: DEBUG oslo_concurrency.lockutils [None req-15de031b-9972-4bde-a9c6-0e58042e23f7 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Acquiring lock "bac71328-3796-4a65-aa41-ccda93bc7926-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 907.117203] env[61923]: DEBUG oslo_concurrency.lockutils [None req-15de031b-9972-4bde-a9c6-0e58042e23f7 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Lock "bac71328-3796-4a65-aa41-ccda93bc7926-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 907.117503] env[61923]: DEBUG oslo_concurrency.lockutils [None req-15de031b-9972-4bde-a9c6-0e58042e23f7 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Lock "bac71328-3796-4a65-aa41-ccda93bc7926-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 907.119853] env[61923]: INFO nova.compute.manager [None req-15de031b-9972-4bde-a9c6-0e58042e23f7 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] [instance: bac71328-3796-4a65-aa41-ccda93bc7926] Terminating instance [ 907.122015] env[61923]: DEBUG nova.compute.manager [None req-15de031b-9972-4bde-a9c6-0e58042e23f7 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] [instance: bac71328-3796-4a65-aa41-ccda93bc7926] Start destroying the instance on the hypervisor. {{(pid=61923) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 907.122237] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-15de031b-9972-4bde-a9c6-0e58042e23f7 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] [instance: bac71328-3796-4a65-aa41-ccda93bc7926] Destroying instance {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 907.123163] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7fd556c-dedd-431d-a715-0b6cb41fd0c8 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.132153] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-15de031b-9972-4bde-a9c6-0e58042e23f7 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] [instance: bac71328-3796-4a65-aa41-ccda93bc7926] Powering off the VM {{(pid=61923) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 907.132438] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a63d0c9f-a638-41fb-80aa-a85156c41158 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.141568] env[61923]: DEBUG oslo_vmware.api [None req-15de031b-9972-4bde-a9c6-0e58042e23f7 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Waiting for the task: (returnval){ [ 907.141568] env[61923]: value = "task-1377762" [ 907.141568] env[61923]: _type = "Task" [ 907.141568] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 907.150732] env[61923]: INFO nova.compute.manager [-] [instance: 59198f4d-4dde-4eaf-9f6c-a962cbe53c6e] Took 1.28 seconds to deallocate network for instance. [ 907.151097] env[61923]: DEBUG oslo_vmware.api [None req-15de031b-9972-4bde-a9c6-0e58042e23f7 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Task: {'id': task-1377762, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.165758] env[61923]: DEBUG oslo_vmware.api [None req-81989905-2c22-4589-a2f0-03b8d81c0d2f tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Task: {'id': task-1377758, 'name': CreateSnapshot_Task, 'duration_secs': 1.026129} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 907.166150] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-81989905-2c22-4589-a2f0-03b8d81c0d2f tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] [instance: 87ae37dc-d0d0-4b76-afdd-0ba3e8f6ce0b] Created Snapshot of the VM instance {{(pid=61923) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 907.166996] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97e430c4-1b98-419e-8a80-949c4dfa0eeb {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.220485] env[61923]: DEBUG oslo_vmware.api [None req-74d3449c-2b90-473c-a8b0-7be80c06daed tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Task: {'id': task-1377760, 'name': ReconfigVM_Task, 'duration_secs': 0.915317} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 907.221066] env[61923]: DEBUG oslo_concurrency.lockutils [None req-74d3449c-2b90-473c-a8b0-7be80c06daed tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Releasing lock "6012d3b5-739c-4762-9bb4-09c51171dcd7" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 907.221313] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-74d3449c-2b90-473c-a8b0-7be80c06daed tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] [instance: 6012d3b5-739c-4762-9bb4-09c51171dcd7] Reconfigured VM to attach interface {{(pid=61923) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1937}} [ 907.451060] env[61923]: DEBUG oslo_concurrency.lockutils [req-921b1676-2d6b-43e1-b976-ca5c3c9dc5ae req-602c27ca-8a3d-4849-b1bd-0cac7b9b9266 service nova] Releasing lock "refresh_cache-6012d3b5-739c-4762-9bb4-09c51171dcd7" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 907.451327] env[61923]: DEBUG nova.compute.manager [req-921b1676-2d6b-43e1-b976-ca5c3c9dc5ae req-602c27ca-8a3d-4849-b1bd-0cac7b9b9266 service nova] [instance: 263f0866-49af-4ab6-8132-d993642ebedb] Received event network-vif-deleted-a16defc4-1374-44de-a7c1-e2d75f27e954 {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 907.451327] env[61923]: INFO nova.compute.manager [req-921b1676-2d6b-43e1-b976-ca5c3c9dc5ae req-602c27ca-8a3d-4849-b1bd-0cac7b9b9266 service nova] [instance: 263f0866-49af-4ab6-8132-d993642ebedb] Neutron deleted interface a16defc4-1374-44de-a7c1-e2d75f27e954; detaching it from the instance and deleting it from the info cache [ 907.451677] env[61923]: DEBUG nova.network.neutron [req-921b1676-2d6b-43e1-b976-ca5c3c9dc5ae req-602c27ca-8a3d-4849-b1bd-0cac7b9b9266 service nova] [instance: 263f0866-49af-4ab6-8132-d993642ebedb] Updating instance_info_cache with network_info: [{"id": "99825a7a-e38f-449b-8633-318da7f11ab5", "address": "fa:16:3e:28:17:ff", "network": {"id": "83abfb75-61d1-4c48-9d06-077d2ae7c85b", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1481658242", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.143", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c304b5b5d4df485fbda55a570494a62f", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fb68953b-dee5-4d9d-b47b-277336ba76dc", "external-id": "nsx-vlan-transportzone-168", "segmentation_id": 168, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap99825a7a-e3", "ovs_interfaceid": "99825a7a-e38f-449b-8633-318da7f11ab5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 907.537734] env[61923]: INFO nova.compute.manager [-] [instance: af3c317a-4007-4cea-a060-1e7dde5ce49e] Took 1.85 seconds to deallocate network for instance. [ 907.557574] env[61923]: DEBUG nova.scheduler.client.report [None req-4a664699-f279-4048-aa87-1277ae6e408d tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 177, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 907.566414] env[61923]: INFO nova.compute.manager [None req-e1b75445-6f25-4dcb-9846-d40076ac5b42 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] [instance: 624e9207-7f24-4f45-8f52-0fc7a60c1e84] Took 16.66 seconds to build instance. [ 907.574068] env[61923]: DEBUG oslo_vmware.api [None req-cb709b57-999a-41a8-b165-7029c0613fc5 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Task: {'id': task-1377761, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.652098] env[61923]: DEBUG oslo_vmware.api [None req-15de031b-9972-4bde-a9c6-0e58042e23f7 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Task: {'id': task-1377762, 'name': PowerOffVM_Task, 'duration_secs': 0.271094} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 907.652341] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-15de031b-9972-4bde-a9c6-0e58042e23f7 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] [instance: bac71328-3796-4a65-aa41-ccda93bc7926] Powered off the VM {{(pid=61923) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 907.652452] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-15de031b-9972-4bde-a9c6-0e58042e23f7 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] [instance: bac71328-3796-4a65-aa41-ccda93bc7926] Unregistering the VM {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 907.652729] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-0d6f0055-15fa-450c-8fc6-6fe92f6eceac {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.660780] env[61923]: DEBUG oslo_concurrency.lockutils [None req-697de1ff-d6e2-4a1e-86f8-0419a0e8777a tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 907.686124] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-81989905-2c22-4589-a2f0-03b8d81c0d2f tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] [instance: 87ae37dc-d0d0-4b76-afdd-0ba3e8f6ce0b] Creating linked-clone VM from snapshot {{(pid=61923) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 907.686482] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-25fdb90e-9cb3-433c-bcf8-436d3319546c {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.696813] env[61923]: DEBUG oslo_vmware.api [None req-81989905-2c22-4589-a2f0-03b8d81c0d2f tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Waiting for the task: (returnval){ [ 907.696813] env[61923]: value = "task-1377764" [ 907.696813] env[61923]: _type = "Task" [ 907.696813] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 907.706884] env[61923]: DEBUG oslo_vmware.api [None req-81989905-2c22-4589-a2f0-03b8d81c0d2f tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Task: {'id': task-1377764, 'name': CloneVM_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.710311] env[61923]: DEBUG nova.compute.manager [req-63155894-7c9a-47c8-b426-51a4ebbf7c44 req-2a6dd94c-b7d6-48d7-a9f4-937f2b6b3c28 service nova] [instance: af3c317a-4007-4cea-a060-1e7dde5ce49e] Received event network-vif-deleted-e5c62d99-3250-4fca-9798-25780ae50b00 {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 907.726490] env[61923]: DEBUG oslo_concurrency.lockutils [None req-74d3449c-2b90-473c-a8b0-7be80c06daed tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Lock "interface-6012d3b5-739c-4762-9bb4-09c51171dcd7-b395b183-1c30-42de-929f-690ea10d9d75" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 6.458s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 907.728926] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-15de031b-9972-4bde-a9c6-0e58042e23f7 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] [instance: bac71328-3796-4a65-aa41-ccda93bc7926] Unregistered the VM {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 907.729421] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-15de031b-9972-4bde-a9c6-0e58042e23f7 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] [instance: bac71328-3796-4a65-aa41-ccda93bc7926] Deleting contents of the VM from datastore datastore1 {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 907.729748] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-15de031b-9972-4bde-a9c6-0e58042e23f7 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Deleting the datastore file [datastore1] bac71328-3796-4a65-aa41-ccda93bc7926 {{(pid=61923) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 907.730321] env[61923]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-57bbc9b1-f560-4dec-b412-eb2f8812f30f {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.739312] env[61923]: DEBUG oslo_vmware.api [None req-15de031b-9972-4bde-a9c6-0e58042e23f7 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Waiting for the task: (returnval){ [ 907.739312] env[61923]: value = "task-1377765" [ 907.739312] env[61923]: _type = "Task" [ 907.739312] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 907.753880] env[61923]: DEBUG oslo_vmware.api [None req-15de031b-9972-4bde-a9c6-0e58042e23f7 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Task: {'id': task-1377765, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.955215] env[61923]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-6634c5cd-ad5a-47c9-ae73-537721e8912a {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.965110] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77f19a74-4ade-409c-94c1-fe3dd750b49a {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.999477] env[61923]: DEBUG nova.compute.manager [req-921b1676-2d6b-43e1-b976-ca5c3c9dc5ae req-602c27ca-8a3d-4849-b1bd-0cac7b9b9266 service nova] [instance: 263f0866-49af-4ab6-8132-d993642ebedb] Detach interface failed, port_id=a16defc4-1374-44de-a7c1-e2d75f27e954, reason: Instance 263f0866-49af-4ab6-8132-d993642ebedb could not be found. {{(pid=61923) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 907.999756] env[61923]: DEBUG nova.compute.manager [req-921b1676-2d6b-43e1-b976-ca5c3c9dc5ae req-602c27ca-8a3d-4849-b1bd-0cac7b9b9266 service nova] [instance: 263f0866-49af-4ab6-8132-d993642ebedb] Received event network-vif-deleted-99825a7a-e38f-449b-8633-318da7f11ab5 {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 907.999953] env[61923]: INFO nova.compute.manager [req-921b1676-2d6b-43e1-b976-ca5c3c9dc5ae req-602c27ca-8a3d-4849-b1bd-0cac7b9b9266 service nova] [instance: 263f0866-49af-4ab6-8132-d993642ebedb] Neutron deleted interface 99825a7a-e38f-449b-8633-318da7f11ab5; detaching it from the instance and deleting it from the info cache [ 908.000177] env[61923]: DEBUG nova.network.neutron [req-921b1676-2d6b-43e1-b976-ca5c3c9dc5ae req-602c27ca-8a3d-4849-b1bd-0cac7b9b9266 service nova] [instance: 263f0866-49af-4ab6-8132-d993642ebedb] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 908.045337] env[61923]: DEBUG oslo_concurrency.lockutils [None req-69c7da51-15c3-4307-92c9-fa8aea180f17 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 908.068852] env[61923]: DEBUG oslo_concurrency.lockutils [None req-4a664699-f279-4048-aa87-1277ae6e408d tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.856s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 908.074709] env[61923]: DEBUG oslo_concurrency.lockutils [None req-e1b75445-6f25-4dcb-9846-d40076ac5b42 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Lock "624e9207-7f24-4f45-8f52-0fc7a60c1e84" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 18.180s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 908.074918] env[61923]: DEBUG oslo_concurrency.lockutils [None req-9a89d0be-1b30-454c-a15f-f7048c8714c5 tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.591s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 908.076623] env[61923]: INFO nova.compute.claims [None req-9a89d0be-1b30-454c-a15f-f7048c8714c5 tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] [instance: 62c75dfd-6046-4de1-b1e3-13a307af1394] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 908.079830] env[61923]: DEBUG oslo_vmware.api [None req-cb709b57-999a-41a8-b165-7029c0613fc5 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Task: {'id': task-1377761, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.749749} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 908.079952] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-cb709b57-999a-41a8-b165-7029c0613fc5 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk to [datastore1] 03c1e8ea-a322-46fa-9b07-e283080a7871/03c1e8ea-a322-46fa-9b07-e283080a7871.vmdk {{(pid=61923) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 908.080173] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-cb709b57-999a-41a8-b165-7029c0613fc5 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: 03c1e8ea-a322-46fa-9b07-e283080a7871] Extending root virtual disk to 1048576 {{(pid=61923) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 908.080930] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-3831c3f4-805e-448e-8e9c-c23bb5a2be80 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.090071] env[61923]: DEBUG oslo_vmware.api [None req-cb709b57-999a-41a8-b165-7029c0613fc5 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Waiting for the task: (returnval){ [ 908.090071] env[61923]: value = "task-1377766" [ 908.090071] env[61923]: _type = "Task" [ 908.090071] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 908.104661] env[61923]: INFO nova.scheduler.client.report [None req-4a664699-f279-4048-aa87-1277ae6e408d tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] Deleted allocations for instance 263f0866-49af-4ab6-8132-d993642ebedb [ 908.109544] env[61923]: DEBUG oslo_vmware.api [None req-cb709b57-999a-41a8-b165-7029c0613fc5 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Task: {'id': task-1377766, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.206911] env[61923]: DEBUG oslo_vmware.api [None req-81989905-2c22-4589-a2f0-03b8d81c0d2f tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Task: {'id': task-1377764, 'name': CloneVM_Task} progress is 94%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.250348] env[61923]: DEBUG oslo_vmware.api [None req-15de031b-9972-4bde-a9c6-0e58042e23f7 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Task: {'id': task-1377765, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.504530] env[61923]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-fabf7fec-926b-44bf-b2ec-93a90191f44a {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.515528] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a29dbb47-d0ae-4020-9b26-ff2e5954d5af {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.548689] env[61923]: DEBUG nova.compute.manager [req-921b1676-2d6b-43e1-b976-ca5c3c9dc5ae req-602c27ca-8a3d-4849-b1bd-0cac7b9b9266 service nova] [instance: 263f0866-49af-4ab6-8132-d993642ebedb] Detach interface failed, port_id=99825a7a-e38f-449b-8633-318da7f11ab5, reason: Instance 263f0866-49af-4ab6-8132-d993642ebedb could not be found. {{(pid=61923) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 908.601272] env[61923]: INFO nova.compute.manager [None req-e087e202-bbb7-4067-8df4-dbcbdaf570ee tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] [instance: 624e9207-7f24-4f45-8f52-0fc7a60c1e84] Rebuilding instance [ 908.603298] env[61923]: DEBUG oslo_vmware.api [None req-cb709b57-999a-41a8-b165-7029c0613fc5 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Task: {'id': task-1377766, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.393341} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 908.603996] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-cb709b57-999a-41a8-b165-7029c0613fc5 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: 03c1e8ea-a322-46fa-9b07-e283080a7871] Extended root virtual disk {{(pid=61923) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 908.604568] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b9184d3-cd49-4674-a387-83d95b79ce4c {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.629701] env[61923]: DEBUG nova.virt.vmwareapi.volumeops [None req-cb709b57-999a-41a8-b165-7029c0613fc5 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: 03c1e8ea-a322-46fa-9b07-e283080a7871] Reconfiguring VM instance instance-00000056 to attach disk [datastore1] 03c1e8ea-a322-46fa-9b07-e283080a7871/03c1e8ea-a322-46fa-9b07-e283080a7871.vmdk or device None with type sparse {{(pid=61923) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 908.630273] env[61923]: DEBUG oslo_concurrency.lockutils [None req-4a664699-f279-4048-aa87-1277ae6e408d tempest-ServersTestMultiNic-960118741 tempest-ServersTestMultiNic-960118741-project-member] Lock "263f0866-49af-4ab6-8132-d993642ebedb" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.611s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 908.633785] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9b0c3805-1bcd-4e2f-b4d5-c9d3ad325913 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.656995] env[61923]: DEBUG oslo_vmware.api [None req-cb709b57-999a-41a8-b165-7029c0613fc5 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Waiting for the task: (returnval){ [ 908.656995] env[61923]: value = "task-1377767" [ 908.656995] env[61923]: _type = "Task" [ 908.656995] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 908.667485] env[61923]: DEBUG oslo_vmware.api [None req-cb709b57-999a-41a8-b165-7029c0613fc5 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Task: {'id': task-1377767, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.672637] env[61923]: DEBUG nova.compute.manager [None req-e087e202-bbb7-4067-8df4-dbcbdaf570ee tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] [instance: 624e9207-7f24-4f45-8f52-0fc7a60c1e84] Checking state {{(pid=61923) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 908.673435] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23988f35-2c55-4b24-914a-e8220f7774d2 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.707636] env[61923]: DEBUG oslo_vmware.api [None req-81989905-2c22-4589-a2f0-03b8d81c0d2f tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Task: {'id': task-1377764, 'name': CloneVM_Task} progress is 94%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 908.750785] env[61923]: DEBUG oslo_vmware.api [None req-15de031b-9972-4bde-a9c6-0e58042e23f7 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Task: {'id': task-1377765, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.752719} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 908.751676] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-15de031b-9972-4bde-a9c6-0e58042e23f7 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Deleted the datastore file {{(pid=61923) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 908.751676] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-15de031b-9972-4bde-a9c6-0e58042e23f7 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] [instance: bac71328-3796-4a65-aa41-ccda93bc7926] Deleted contents of the VM from datastore datastore1 {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 908.751676] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-15de031b-9972-4bde-a9c6-0e58042e23f7 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] [instance: bac71328-3796-4a65-aa41-ccda93bc7926] Instance destroyed {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 908.751853] env[61923]: INFO nova.compute.manager [None req-15de031b-9972-4bde-a9c6-0e58042e23f7 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] [instance: bac71328-3796-4a65-aa41-ccda93bc7926] Took 1.63 seconds to destroy the instance on the hypervisor. [ 908.751978] env[61923]: DEBUG oslo.service.loopingcall [None req-15de031b-9972-4bde-a9c6-0e58042e23f7 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61923) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 908.752212] env[61923]: DEBUG nova.compute.manager [-] [instance: bac71328-3796-4a65-aa41-ccda93bc7926] Deallocating network for instance {{(pid=61923) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 908.752347] env[61923]: DEBUG nova.network.neutron [-] [instance: bac71328-3796-4a65-aa41-ccda93bc7926] deallocate_for_instance() {{(pid=61923) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 909.055693] env[61923]: DEBUG oslo_concurrency.lockutils [None req-f37b2e0d-8f0c-42f5-bf08-4bc9d42bb688 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Acquiring lock "interface-6012d3b5-739c-4762-9bb4-09c51171dcd7-b395b183-1c30-42de-929f-690ea10d9d75" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 909.055693] env[61923]: DEBUG oslo_concurrency.lockutils [None req-f37b2e0d-8f0c-42f5-bf08-4bc9d42bb688 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Lock "interface-6012d3b5-739c-4762-9bb4-09c51171dcd7-b395b183-1c30-42de-929f-690ea10d9d75" acquired by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: waited 0.001s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 909.178754] env[61923]: DEBUG oslo_vmware.api [None req-cb709b57-999a-41a8-b165-7029c0613fc5 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Task: {'id': task-1377767, 'name': ReconfigVM_Task, 'duration_secs': 0.325662} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 909.180027] env[61923]: DEBUG nova.virt.vmwareapi.volumeops [None req-cb709b57-999a-41a8-b165-7029c0613fc5 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: 03c1e8ea-a322-46fa-9b07-e283080a7871] Reconfigured VM instance instance-00000056 to attach disk [datastore1] 03c1e8ea-a322-46fa-9b07-e283080a7871/03c1e8ea-a322-46fa-9b07-e283080a7871.vmdk or device None with type sparse {{(pid=61923) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 909.181025] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-1d079c41-56b7-425f-bfcd-b11f85195f34 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.184168] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-e087e202-bbb7-4067-8df4-dbcbdaf570ee tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] [instance: 624e9207-7f24-4f45-8f52-0fc7a60c1e84] Powering off the VM {{(pid=61923) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 909.184453] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-11f61990-da34-4746-a25a-e448e755d1fd {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.188183] env[61923]: DEBUG oslo_vmware.api [None req-cb709b57-999a-41a8-b165-7029c0613fc5 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Waiting for the task: (returnval){ [ 909.188183] env[61923]: value = "task-1377768" [ 909.188183] env[61923]: _type = "Task" [ 909.188183] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 909.195710] env[61923]: DEBUG oslo_vmware.api [None req-e087e202-bbb7-4067-8df4-dbcbdaf570ee tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Waiting for the task: (returnval){ [ 909.195710] env[61923]: value = "task-1377769" [ 909.195710] env[61923]: _type = "Task" [ 909.195710] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 909.202282] env[61923]: DEBUG oslo_vmware.api [None req-cb709b57-999a-41a8-b165-7029c0613fc5 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Task: {'id': task-1377768, 'name': Rename_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.212824] env[61923]: DEBUG oslo_vmware.api [None req-e087e202-bbb7-4067-8df4-dbcbdaf570ee tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Task: {'id': task-1377769, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.216196] env[61923]: DEBUG oslo_vmware.api [None req-81989905-2c22-4589-a2f0-03b8d81c0d2f tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Task: {'id': task-1377764, 'name': CloneVM_Task} progress is 100%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.306791] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25132664-4e47-4f0d-9eb6-70537042be6b {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.316830] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fcf5dd0e-dc4c-4cfb-a9a2-3fded5fef641 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.357298] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec695d96-8aef-491e-a253-aa40679c5a1d {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.367717] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64415a57-54af-4248-9c36-6aee144abd4d {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.382984] env[61923]: DEBUG nova.compute.provider_tree [None req-9a89d0be-1b30-454c-a15f-f7048c8714c5 tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 909.559737] env[61923]: DEBUG oslo_concurrency.lockutils [None req-f37b2e0d-8f0c-42f5-bf08-4bc9d42bb688 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Acquiring lock "6012d3b5-739c-4762-9bb4-09c51171dcd7" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 909.560088] env[61923]: DEBUG oslo_concurrency.lockutils [None req-f37b2e0d-8f0c-42f5-bf08-4bc9d42bb688 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Acquired lock "6012d3b5-739c-4762-9bb4-09c51171dcd7" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 909.560817] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fda5ce0d-4907-4072-a8a1-84740f00dd41 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.581419] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9b1abd2-63e0-41ba-94ae-7e0992e79922 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.609352] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-f37b2e0d-8f0c-42f5-bf08-4bc9d42bb688 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] [instance: 6012d3b5-739c-4762-9bb4-09c51171dcd7] Reconfiguring VM to detach interface {{(pid=61923) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1968}} [ 909.609670] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4dc27bd2-4f2a-4c10-bbfc-ec3f62a81c97 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.634067] env[61923]: DEBUG oslo_vmware.api [None req-f37b2e0d-8f0c-42f5-bf08-4bc9d42bb688 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Waiting for the task: (returnval){ [ 909.634067] env[61923]: value = "task-1377770" [ 909.634067] env[61923]: _type = "Task" [ 909.634067] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 909.641015] env[61923]: DEBUG oslo_vmware.api [None req-f37b2e0d-8f0c-42f5-bf08-4bc9d42bb688 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Task: {'id': task-1377770, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.673235] env[61923]: DEBUG nova.network.neutron [-] [instance: bac71328-3796-4a65-aa41-ccda93bc7926] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 909.700411] env[61923]: DEBUG oslo_vmware.api [None req-cb709b57-999a-41a8-b165-7029c0613fc5 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Task: {'id': task-1377768, 'name': Rename_Task, 'duration_secs': 0.151871} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 909.703512] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-cb709b57-999a-41a8-b165-7029c0613fc5 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: 03c1e8ea-a322-46fa-9b07-e283080a7871] Powering on the VM {{(pid=61923) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 909.704717] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-02cbf139-043f-431a-9f92-5da53428b964 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.717340] env[61923]: DEBUG oslo_vmware.api [None req-e087e202-bbb7-4067-8df4-dbcbdaf570ee tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Task: {'id': task-1377769, 'name': PowerOffVM_Task} progress is 100%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.718753] env[61923]: DEBUG oslo_vmware.api [None req-cb709b57-999a-41a8-b165-7029c0613fc5 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Waiting for the task: (returnval){ [ 909.718753] env[61923]: value = "task-1377771" [ 909.718753] env[61923]: _type = "Task" [ 909.718753] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 909.719014] env[61923]: DEBUG oslo_vmware.api [None req-81989905-2c22-4589-a2f0-03b8d81c0d2f tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Task: {'id': task-1377764, 'name': CloneVM_Task, 'duration_secs': 1.561005} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 909.719399] env[61923]: INFO nova.virt.vmwareapi.vmops [None req-81989905-2c22-4589-a2f0-03b8d81c0d2f tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] [instance: 87ae37dc-d0d0-4b76-afdd-0ba3e8f6ce0b] Created linked-clone VM from snapshot [ 909.722893] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1fae346-62b1-437a-95f1-49839d250bbf {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.734575] env[61923]: DEBUG oslo_vmware.api [None req-cb709b57-999a-41a8-b165-7029c0613fc5 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Task: {'id': task-1377771, 'name': PowerOnVM_Task} progress is 33%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.740372] env[61923]: DEBUG nova.compute.manager [req-17882b3b-15d4-4c17-aa18-465cd1b88dd7 req-a3c4bc86-d30b-43c2-8816-40d9def96885 service nova] [instance: bac71328-3796-4a65-aa41-ccda93bc7926] Received event network-vif-deleted-923d9288-8294-4410-b6e7-529fefc2c6a7 {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 909.741073] env[61923]: DEBUG nova.virt.vmwareapi.images [None req-81989905-2c22-4589-a2f0-03b8d81c0d2f tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] [instance: 87ae37dc-d0d0-4b76-afdd-0ba3e8f6ce0b] Uploading image ba0d8caf-3d48-4681-b6d3-c851741d738b {{(pid=61923) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 909.763548] env[61923]: DEBUG oslo_vmware.rw_handles [None req-81989905-2c22-4589-a2f0-03b8d81c0d2f tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 909.763548] env[61923]: value = "vm-292715" [ 909.763548] env[61923]: _type = "VirtualMachine" [ 909.763548] env[61923]: }. {{(pid=61923) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 909.763845] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-f005739c-5703-47ca-9591-cc63a9a4f005 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.773208] env[61923]: DEBUG oslo_vmware.rw_handles [None req-81989905-2c22-4589-a2f0-03b8d81c0d2f tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Lease: (returnval){ [ 909.773208] env[61923]: value = "session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]521651a1-df0b-404a-db41-538d58f8442e" [ 909.773208] env[61923]: _type = "HttpNfcLease" [ 909.773208] env[61923]: } obtained for exporting VM: (result){ [ 909.773208] env[61923]: value = "vm-292715" [ 909.773208] env[61923]: _type = "VirtualMachine" [ 909.773208] env[61923]: }. {{(pid=61923) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 909.773613] env[61923]: DEBUG oslo_vmware.api [None req-81989905-2c22-4589-a2f0-03b8d81c0d2f tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Waiting for the lease: (returnval){ [ 909.773613] env[61923]: value = "session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]521651a1-df0b-404a-db41-538d58f8442e" [ 909.773613] env[61923]: _type = "HttpNfcLease" [ 909.773613] env[61923]: } to be ready. {{(pid=61923) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 909.781843] env[61923]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 909.781843] env[61923]: value = "session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]521651a1-df0b-404a-db41-538d58f8442e" [ 909.781843] env[61923]: _type = "HttpNfcLease" [ 909.781843] env[61923]: } is initializing. {{(pid=61923) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 909.886209] env[61923]: DEBUG nova.scheduler.client.report [None req-9a89d0be-1b30-454c-a15f-f7048c8714c5 tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 177, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 910.144452] env[61923]: DEBUG oslo_vmware.api [None req-f37b2e0d-8f0c-42f5-bf08-4bc9d42bb688 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Task: {'id': task-1377770, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.175794] env[61923]: INFO nova.compute.manager [-] [instance: bac71328-3796-4a65-aa41-ccda93bc7926] Took 1.42 seconds to deallocate network for instance. [ 910.209952] env[61923]: DEBUG oslo_vmware.api [None req-e087e202-bbb7-4067-8df4-dbcbdaf570ee tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Task: {'id': task-1377769, 'name': PowerOffVM_Task, 'duration_secs': 0.564198} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 910.210310] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-e087e202-bbb7-4067-8df4-dbcbdaf570ee tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] [instance: 624e9207-7f24-4f45-8f52-0fc7a60c1e84] Powered off the VM {{(pid=61923) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 910.210559] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-e087e202-bbb7-4067-8df4-dbcbdaf570ee tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] [instance: 624e9207-7f24-4f45-8f52-0fc7a60c1e84] Destroying instance {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 910.211383] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-625a7fb2-db1f-40f3-86e6-425c5463cb99 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.219195] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-e087e202-bbb7-4067-8df4-dbcbdaf570ee tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] [instance: 624e9207-7f24-4f45-8f52-0fc7a60c1e84] Unregistering the VM {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 910.219599] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-38970486-3777-43d5-97c4-28b158ae716f {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.233485] env[61923]: DEBUG oslo_vmware.api [None req-cb709b57-999a-41a8-b165-7029c0613fc5 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Task: {'id': task-1377771, 'name': PowerOnVM_Task} progress is 33%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.252907] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-e087e202-bbb7-4067-8df4-dbcbdaf570ee tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] [instance: 624e9207-7f24-4f45-8f52-0fc7a60c1e84] Unregistered the VM {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 910.253288] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-e087e202-bbb7-4067-8df4-dbcbdaf570ee tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] [instance: 624e9207-7f24-4f45-8f52-0fc7a60c1e84] Deleting contents of the VM from datastore datastore1 {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 910.253547] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-e087e202-bbb7-4067-8df4-dbcbdaf570ee tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Deleting the datastore file [datastore1] 624e9207-7f24-4f45-8f52-0fc7a60c1e84 {{(pid=61923) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 910.253884] env[61923]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-98fd06f8-ea0b-4e51-92b2-7885e8774b31 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.261547] env[61923]: DEBUG oslo_vmware.api [None req-e087e202-bbb7-4067-8df4-dbcbdaf570ee tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Waiting for the task: (returnval){ [ 910.261547] env[61923]: value = "task-1377774" [ 910.261547] env[61923]: _type = "Task" [ 910.261547] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 910.270497] env[61923]: DEBUG oslo_vmware.api [None req-e087e202-bbb7-4067-8df4-dbcbdaf570ee tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Task: {'id': task-1377774, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.280764] env[61923]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 910.280764] env[61923]: value = "session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]521651a1-df0b-404a-db41-538d58f8442e" [ 910.280764] env[61923]: _type = "HttpNfcLease" [ 910.280764] env[61923]: } is ready. {{(pid=61923) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 910.281156] env[61923]: DEBUG oslo_vmware.rw_handles [None req-81989905-2c22-4589-a2f0-03b8d81c0d2f tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 910.281156] env[61923]: value = "session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]521651a1-df0b-404a-db41-538d58f8442e" [ 910.281156] env[61923]: _type = "HttpNfcLease" [ 910.281156] env[61923]: }. {{(pid=61923) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 910.282014] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db903410-cfa3-49ff-89d0-098553f00ba1 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.289337] env[61923]: DEBUG oslo_vmware.rw_handles [None req-81989905-2c22-4589-a2f0-03b8d81c0d2f tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/522995c9-80ac-b69a-62c7-1e4b40c5a0df/disk-0.vmdk from lease info. {{(pid=61923) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 910.289592] env[61923]: DEBUG oslo_vmware.rw_handles [None req-81989905-2c22-4589-a2f0-03b8d81c0d2f tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Opening URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/522995c9-80ac-b69a-62c7-1e4b40c5a0df/disk-0.vmdk for reading. {{(pid=61923) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 910.391155] env[61923]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-1e0bf086-5ccb-4a72-b98b-f400ffb54b4d {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.393552] env[61923]: DEBUG oslo_concurrency.lockutils [None req-9a89d0be-1b30-454c-a15f-f7048c8714c5 tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.318s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 910.394665] env[61923]: DEBUG nova.compute.manager [None req-9a89d0be-1b30-454c-a15f-f7048c8714c5 tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] [instance: 62c75dfd-6046-4de1-b1e3-13a307af1394] Start building networks asynchronously for instance. {{(pid=61923) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 910.396913] env[61923]: DEBUG oslo_concurrency.lockutils [None req-697de1ff-d6e2-4a1e-86f8-0419a0e8777a tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.736s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 910.399944] env[61923]: DEBUG nova.objects.instance [None req-697de1ff-d6e2-4a1e-86f8-0419a0e8777a tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Lazy-loading 'resources' on Instance uuid 59198f4d-4dde-4eaf-9f6c-a962cbe53c6e {{(pid=61923) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 910.645100] env[61923]: DEBUG oslo_vmware.api [None req-f37b2e0d-8f0c-42f5-bf08-4bc9d42bb688 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Task: {'id': task-1377770, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.682477] env[61923]: DEBUG oslo_concurrency.lockutils [None req-15de031b-9972-4bde-a9c6-0e58042e23f7 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 910.730242] env[61923]: DEBUG oslo_vmware.api [None req-cb709b57-999a-41a8-b165-7029c0613fc5 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Task: {'id': task-1377771, 'name': PowerOnVM_Task} progress is 33%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.774281] env[61923]: DEBUG oslo_vmware.api [None req-e087e202-bbb7-4067-8df4-dbcbdaf570ee tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Task: {'id': task-1377774, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.344163} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 910.774905] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-e087e202-bbb7-4067-8df4-dbcbdaf570ee tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Deleted the datastore file {{(pid=61923) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 910.774905] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-e087e202-bbb7-4067-8df4-dbcbdaf570ee tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] [instance: 624e9207-7f24-4f45-8f52-0fc7a60c1e84] Deleted contents of the VM from datastore datastore1 {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 910.774905] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-e087e202-bbb7-4067-8df4-dbcbdaf570ee tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] [instance: 624e9207-7f24-4f45-8f52-0fc7a60c1e84] Instance destroyed {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 910.901019] env[61923]: DEBUG nova.compute.utils [None req-9a89d0be-1b30-454c-a15f-f7048c8714c5 tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Using /dev/sd instead of None {{(pid=61923) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 910.907385] env[61923]: DEBUG nova.compute.manager [None req-9a89d0be-1b30-454c-a15f-f7048c8714c5 tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] [instance: 62c75dfd-6046-4de1-b1e3-13a307af1394] Allocating IP information in the background. {{(pid=61923) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 910.907883] env[61923]: DEBUG nova.network.neutron [None req-9a89d0be-1b30-454c-a15f-f7048c8714c5 tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] [instance: 62c75dfd-6046-4de1-b1e3-13a307af1394] allocate_for_instance() {{(pid=61923) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 910.964680] env[61923]: DEBUG nova.policy [None req-9a89d0be-1b30-454c-a15f-f7048c8714c5 tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'eb00f8aff92b44e9a0e739366a1d2662', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7abfe95660904cd4b8824257795d8e8a', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61923) authorize /opt/stack/nova/nova/policy.py:201}} [ 911.144658] env[61923]: DEBUG oslo_vmware.api [None req-f37b2e0d-8f0c-42f5-bf08-4bc9d42bb688 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Task: {'id': task-1377770, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.186743] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f490fe47-c059-4412-9f24-2790181a0012 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.198197] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f55b96ca-b35a-439c-9380-ac8cc1255c9b {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.243742] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b57370d8-ca05-45be-a9c1-a612ee5f4af6 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.256717] env[61923]: DEBUG oslo_vmware.api [None req-cb709b57-999a-41a8-b165-7029c0613fc5 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Task: {'id': task-1377771, 'name': PowerOnVM_Task, 'duration_secs': 1.494441} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 911.257411] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-cb709b57-999a-41a8-b165-7029c0613fc5 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: 03c1e8ea-a322-46fa-9b07-e283080a7871] Powered on the VM {{(pid=61923) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 911.257637] env[61923]: INFO nova.compute.manager [None req-cb709b57-999a-41a8-b165-7029c0613fc5 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: 03c1e8ea-a322-46fa-9b07-e283080a7871] Took 8.39 seconds to spawn the instance on the hypervisor. [ 911.257822] env[61923]: DEBUG nova.compute.manager [None req-cb709b57-999a-41a8-b165-7029c0613fc5 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: 03c1e8ea-a322-46fa-9b07-e283080a7871] Checking state {{(pid=61923) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 911.259336] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1af7346b-8002-47a1-b177-7eb66a7eb535 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.263902] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d79983e-e829-41c8-8afe-8cb3e0d87c29 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.278987] env[61923]: DEBUG nova.compute.provider_tree [None req-697de1ff-d6e2-4a1e-86f8-0419a0e8777a tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Updating inventory in ProviderTree for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 911.398635] env[61923]: DEBUG nova.network.neutron [None req-9a89d0be-1b30-454c-a15f-f7048c8714c5 tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] [instance: 62c75dfd-6046-4de1-b1e3-13a307af1394] Successfully created port: 2d89de21-00bd-46da-b7c1-6ac2cbba1982 {{(pid=61923) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 911.412022] env[61923]: DEBUG nova.compute.manager [None req-9a89d0be-1b30-454c-a15f-f7048c8714c5 tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] [instance: 62c75dfd-6046-4de1-b1e3-13a307af1394] Start building block device mappings for instance. {{(pid=61923) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 911.652551] env[61923]: DEBUG oslo_vmware.api [None req-f37b2e0d-8f0c-42f5-bf08-4bc9d42bb688 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Task: {'id': task-1377770, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.806167] env[61923]: INFO nova.compute.manager [None req-cb709b57-999a-41a8-b165-7029c0613fc5 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: 03c1e8ea-a322-46fa-9b07-e283080a7871] Took 17.70 seconds to build instance. [ 911.823423] env[61923]: ERROR nova.scheduler.client.report [None req-697de1ff-d6e2-4a1e-86f8-0419a0e8777a tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] [req-39a6fe24-e9d8-47dc-a7fb-b302c546eb7d] Failed to update inventory to [{'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}}] for resource provider with UUID f81803f2-f7f8-4939-a757-a77d34a1d0a2. Got 409: {"errors": [{"status": 409, "title": "Conflict", "detail": "There was a conflict when trying to complete your request.\n\n resource provider generation conflict ", "code": "placement.concurrent_update", "request_id": "req-39a6fe24-e9d8-47dc-a7fb-b302c546eb7d"}]} [ 911.840966] env[61923]: DEBUG nova.virt.hardware [None req-e087e202-bbb7-4067-8df4-dbcbdaf570ee tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-29T20:07:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-29T20:07:35Z,direct_url=,disk_format='vmdk',id=0f153f63-ae0a-45b1-b7f5-7f9b673c947f,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='4e7b5e3b3c3443c8bd5c70f8a15739f5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-29T20:07:36Z,virtual_size=,visibility=), allow threads: False {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 911.841765] env[61923]: DEBUG nova.virt.hardware [None req-e087e202-bbb7-4067-8df4-dbcbdaf570ee tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Flavor limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 911.842771] env[61923]: DEBUG nova.virt.hardware [None req-e087e202-bbb7-4067-8df4-dbcbdaf570ee tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Image limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 911.843061] env[61923]: DEBUG nova.virt.hardware [None req-e087e202-bbb7-4067-8df4-dbcbdaf570ee tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Flavor pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 911.843333] env[61923]: DEBUG nova.virt.hardware [None req-e087e202-bbb7-4067-8df4-dbcbdaf570ee tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Image pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 911.843620] env[61923]: DEBUG nova.virt.hardware [None req-e087e202-bbb7-4067-8df4-dbcbdaf570ee tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 911.843915] env[61923]: DEBUG nova.virt.hardware [None req-e087e202-bbb7-4067-8df4-dbcbdaf570ee tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 911.844154] env[61923]: DEBUG nova.virt.hardware [None req-e087e202-bbb7-4067-8df4-dbcbdaf570ee tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 911.844426] env[61923]: DEBUG nova.virt.hardware [None req-e087e202-bbb7-4067-8df4-dbcbdaf570ee tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Got 1 possible topologies {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 911.844657] env[61923]: DEBUG nova.virt.hardware [None req-e087e202-bbb7-4067-8df4-dbcbdaf570ee tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 911.845303] env[61923]: DEBUG nova.virt.hardware [None req-e087e202-bbb7-4067-8df4-dbcbdaf570ee tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 911.846569] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-457494a2-c99f-4365-a4e2-387626b3cba1 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.852158] env[61923]: DEBUG nova.scheduler.client.report [None req-697de1ff-d6e2-4a1e-86f8-0419a0e8777a tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Refreshing inventories for resource provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 911.861395] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77d6ecb8-5510-495d-b602-0dec5d6ee62c {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.879580] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-e087e202-bbb7-4067-8df4-dbcbdaf570ee tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] [instance: 624e9207-7f24-4f45-8f52-0fc7a60c1e84] Instance VIF info [] {{(pid=61923) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 911.885754] env[61923]: DEBUG oslo.service.loopingcall [None req-e087e202-bbb7-4067-8df4-dbcbdaf570ee tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61923) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 911.887185] env[61923]: DEBUG nova.scheduler.client.report [None req-697de1ff-d6e2-4a1e-86f8-0419a0e8777a tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Updating ProviderTree inventory for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 177, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 911.887455] env[61923]: DEBUG nova.compute.provider_tree [None req-697de1ff-d6e2-4a1e-86f8-0419a0e8777a tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Updating inventory in ProviderTree for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 177, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 911.890034] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 624e9207-7f24-4f45-8f52-0fc7a60c1e84] Creating VM on the ESX host {{(pid=61923) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 911.890718] env[61923]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d6892d23-2215-4022-a5a7-55855c352fd3 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.905520] env[61923]: DEBUG nova.scheduler.client.report [None req-697de1ff-d6e2-4a1e-86f8-0419a0e8777a tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Refreshing aggregate associations for resource provider f81803f2-f7f8-4939-a757-a77d34a1d0a2, aggregates: None {{(pid=61923) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 911.920081] env[61923]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 911.920081] env[61923]: value = "task-1377775" [ 911.920081] env[61923]: _type = "Task" [ 911.920081] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 911.928216] env[61923]: DEBUG nova.scheduler.client.report [None req-697de1ff-d6e2-4a1e-86f8-0419a0e8777a tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Refreshing trait associations for resource provider f81803f2-f7f8-4939-a757-a77d34a1d0a2, traits: COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_SAME_HOST_COLD_MIGRATE,COMPUTE_NODE,HW_ARCH_X86_64,COMPUTE_NET_ATTACH_INTERFACE,COMPUTE_IMAGE_TYPE_ISO {{(pid=61923) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 911.936013] env[61923]: DEBUG oslo_vmware.api [-] Task: {'id': task-1377775, 'name': CreateVM_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.144791] env[61923]: DEBUG oslo_vmware.api [None req-f37b2e0d-8f0c-42f5-bf08-4bc9d42bb688 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Task: {'id': task-1377770, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.146597] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b70d9900-4d5a-4a0c-bf7a-de5ed44335c5 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.154582] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a8456e1-e849-4984-81fe-680cb43d9da4 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.188255] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95b129d8-8437-4bf0-92dc-484f21a2337a {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.197137] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff503aec-3c63-453a-b846-169ca568907c {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.217937] env[61923]: DEBUG nova.compute.provider_tree [None req-697de1ff-d6e2-4a1e-86f8-0419a0e8777a tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Updating inventory in ProviderTree for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 912.310080] env[61923]: DEBUG oslo_concurrency.lockutils [None req-cb709b57-999a-41a8-b165-7029c0613fc5 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Lock "03c1e8ea-a322-46fa-9b07-e283080a7871" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 19.211s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 912.423471] env[61923]: DEBUG nova.compute.manager [None req-9a89d0be-1b30-454c-a15f-f7048c8714c5 tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] [instance: 62c75dfd-6046-4de1-b1e3-13a307af1394] Start spawning the instance on the hypervisor. {{(pid=61923) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 912.435456] env[61923]: DEBUG oslo_vmware.api [-] Task: {'id': task-1377775, 'name': CreateVM_Task, 'duration_secs': 0.392162} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 912.435744] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 624e9207-7f24-4f45-8f52-0fc7a60c1e84] Created VM on the ESX host {{(pid=61923) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 912.436366] env[61923]: DEBUG oslo_concurrency.lockutils [None req-e087e202-bbb7-4067-8df4-dbcbdaf570ee tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 912.436701] env[61923]: DEBUG oslo_concurrency.lockutils [None req-e087e202-bbb7-4067-8df4-dbcbdaf570ee tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 912.437022] env[61923]: DEBUG oslo_concurrency.lockutils [None req-e087e202-bbb7-4067-8df4-dbcbdaf570ee tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 912.437397] env[61923]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f696a415-b762-42b4-8b95-30c7705ff9e3 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.444466] env[61923]: DEBUG oslo_vmware.api [None req-e087e202-bbb7-4067-8df4-dbcbdaf570ee tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Waiting for the task: (returnval){ [ 912.444466] env[61923]: value = "session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]5223ada8-666f-1e49-dea6-c1bacd9569ab" [ 912.444466] env[61923]: _type = "Task" [ 912.444466] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 912.455694] env[61923]: DEBUG oslo_vmware.api [None req-e087e202-bbb7-4067-8df4-dbcbdaf570ee tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]5223ada8-666f-1e49-dea6-c1bacd9569ab, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.458244] env[61923]: DEBUG nova.virt.hardware [None req-9a89d0be-1b30-454c-a15f-f7048c8714c5 tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-29T20:07:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-29T20:07:35Z,direct_url=,disk_format='vmdk',id=0f153f63-ae0a-45b1-b7f5-7f9b673c947f,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='4e7b5e3b3c3443c8bd5c70f8a15739f5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-29T20:07:36Z,virtual_size=,visibility=), allow threads: False {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 912.458949] env[61923]: DEBUG nova.virt.hardware [None req-9a89d0be-1b30-454c-a15f-f7048c8714c5 tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Flavor limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 912.458949] env[61923]: DEBUG nova.virt.hardware [None req-9a89d0be-1b30-454c-a15f-f7048c8714c5 tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Image limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 912.459077] env[61923]: DEBUG nova.virt.hardware [None req-9a89d0be-1b30-454c-a15f-f7048c8714c5 tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Flavor pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 912.459222] env[61923]: DEBUG nova.virt.hardware [None req-9a89d0be-1b30-454c-a15f-f7048c8714c5 tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Image pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 912.459500] env[61923]: DEBUG nova.virt.hardware [None req-9a89d0be-1b30-454c-a15f-f7048c8714c5 tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 912.459781] env[61923]: DEBUG nova.virt.hardware [None req-9a89d0be-1b30-454c-a15f-f7048c8714c5 tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 912.460008] env[61923]: DEBUG nova.virt.hardware [None req-9a89d0be-1b30-454c-a15f-f7048c8714c5 tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 912.460243] env[61923]: DEBUG nova.virt.hardware [None req-9a89d0be-1b30-454c-a15f-f7048c8714c5 tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Got 1 possible topologies {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 912.460475] env[61923]: DEBUG nova.virt.hardware [None req-9a89d0be-1b30-454c-a15f-f7048c8714c5 tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 912.460775] env[61923]: DEBUG nova.virt.hardware [None req-9a89d0be-1b30-454c-a15f-f7048c8714c5 tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 912.461698] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc991c0b-4447-46bd-9ec6-4d16338c5d04 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.469875] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ecf9b877-a3dc-4020-8e50-8ea4ad8525b2 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.645796] env[61923]: DEBUG oslo_vmware.api [None req-f37b2e0d-8f0c-42f5-bf08-4bc9d42bb688 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Task: {'id': task-1377770, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.706587] env[61923]: DEBUG oslo_concurrency.lockutils [None req-85598005-1527-489f-b9ce-6a2b3821cde4 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Acquiring lock "03c1e8ea-a322-46fa-9b07-e283080a7871" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 912.706933] env[61923]: DEBUG oslo_concurrency.lockutils [None req-85598005-1527-489f-b9ce-6a2b3821cde4 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Lock "03c1e8ea-a322-46fa-9b07-e283080a7871" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 912.707111] env[61923]: DEBUG oslo_concurrency.lockutils [None req-85598005-1527-489f-b9ce-6a2b3821cde4 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Acquiring lock "03c1e8ea-a322-46fa-9b07-e283080a7871-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 912.707357] env[61923]: DEBUG oslo_concurrency.lockutils [None req-85598005-1527-489f-b9ce-6a2b3821cde4 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Lock "03c1e8ea-a322-46fa-9b07-e283080a7871-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 912.707577] env[61923]: DEBUG oslo_concurrency.lockutils [None req-85598005-1527-489f-b9ce-6a2b3821cde4 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Lock "03c1e8ea-a322-46fa-9b07-e283080a7871-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 912.710092] env[61923]: INFO nova.compute.manager [None req-85598005-1527-489f-b9ce-6a2b3821cde4 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: 03c1e8ea-a322-46fa-9b07-e283080a7871] Terminating instance [ 912.712251] env[61923]: DEBUG nova.compute.manager [None req-85598005-1527-489f-b9ce-6a2b3821cde4 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: 03c1e8ea-a322-46fa-9b07-e283080a7871] Start destroying the instance on the hypervisor. {{(pid=61923) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 912.712497] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-85598005-1527-489f-b9ce-6a2b3821cde4 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: 03c1e8ea-a322-46fa-9b07-e283080a7871] Destroying instance {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 912.713521] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c1f0be3-c50a-4555-8766-f307c56d115b {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.724819] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-85598005-1527-489f-b9ce-6a2b3821cde4 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: 03c1e8ea-a322-46fa-9b07-e283080a7871] Powering off the VM {{(pid=61923) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 912.725360] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-62c74e36-f57c-4bb6-a37b-bd154160de08 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.734688] env[61923]: DEBUG oslo_vmware.api [None req-85598005-1527-489f-b9ce-6a2b3821cde4 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Waiting for the task: (returnval){ [ 912.734688] env[61923]: value = "task-1377776" [ 912.734688] env[61923]: _type = "Task" [ 912.734688] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 912.745720] env[61923]: DEBUG oslo_vmware.api [None req-85598005-1527-489f-b9ce-6a2b3821cde4 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Task: {'id': task-1377776, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.754958] env[61923]: DEBUG nova.scheduler.client.report [None req-697de1ff-d6e2-4a1e-86f8-0419a0e8777a tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Updated inventory for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 with generation 98 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 912.755269] env[61923]: DEBUG nova.compute.provider_tree [None req-697de1ff-d6e2-4a1e-86f8-0419a0e8777a tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Updating resource provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 generation from 98 to 99 during operation: update_inventory {{(pid=61923) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 912.755459] env[61923]: DEBUG nova.compute.provider_tree [None req-697de1ff-d6e2-4a1e-86f8-0419a0e8777a tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Updating inventory in ProviderTree for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 912.959641] env[61923]: DEBUG oslo_vmware.api [None req-e087e202-bbb7-4067-8df4-dbcbdaf570ee tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]5223ada8-666f-1e49-dea6-c1bacd9569ab, 'name': SearchDatastore_Task, 'duration_secs': 0.02048} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 912.959922] env[61923]: DEBUG oslo_concurrency.lockutils [None req-e087e202-bbb7-4067-8df4-dbcbdaf570ee tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 912.960196] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-e087e202-bbb7-4067-8df4-dbcbdaf570ee tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] [instance: 624e9207-7f24-4f45-8f52-0fc7a60c1e84] Processing image 0f153f63-ae0a-45b1-b7f5-7f9b673c947f {{(pid=61923) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 912.960520] env[61923]: DEBUG oslo_concurrency.lockutils [None req-e087e202-bbb7-4067-8df4-dbcbdaf570ee tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 912.960641] env[61923]: DEBUG oslo_concurrency.lockutils [None req-e087e202-bbb7-4067-8df4-dbcbdaf570ee tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 912.960829] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-e087e202-bbb7-4067-8df4-dbcbdaf570ee tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61923) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 912.961139] env[61923]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-def0dc8d-e739-4323-9f89-21a967d711bb {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.973113] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-e087e202-bbb7-4067-8df4-dbcbdaf570ee tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61923) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 912.973113] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-e087e202-bbb7-4067-8df4-dbcbdaf570ee tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61923) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 912.974351] env[61923]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b021344a-a599-4a8f-8025-4586f2bf9c9e {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.978570] env[61923]: DEBUG nova.compute.manager [req-6a207006-9392-47d7-b015-e0e38c9d8993 req-4d3f045f-4fa3-40cd-90ae-6dd907897702 service nova] [instance: 62c75dfd-6046-4de1-b1e3-13a307af1394] Received event network-vif-plugged-2d89de21-00bd-46da-b7c1-6ac2cbba1982 {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 912.978570] env[61923]: DEBUG oslo_concurrency.lockutils [req-6a207006-9392-47d7-b015-e0e38c9d8993 req-4d3f045f-4fa3-40cd-90ae-6dd907897702 service nova] Acquiring lock "62c75dfd-6046-4de1-b1e3-13a307af1394-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 912.978748] env[61923]: DEBUG oslo_concurrency.lockutils [req-6a207006-9392-47d7-b015-e0e38c9d8993 req-4d3f045f-4fa3-40cd-90ae-6dd907897702 service nova] Lock "62c75dfd-6046-4de1-b1e3-13a307af1394-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 912.979969] env[61923]: DEBUG oslo_concurrency.lockutils [req-6a207006-9392-47d7-b015-e0e38c9d8993 req-4d3f045f-4fa3-40cd-90ae-6dd907897702 service nova] Lock "62c75dfd-6046-4de1-b1e3-13a307af1394-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 912.979969] env[61923]: DEBUG nova.compute.manager [req-6a207006-9392-47d7-b015-e0e38c9d8993 req-4d3f045f-4fa3-40cd-90ae-6dd907897702 service nova] [instance: 62c75dfd-6046-4de1-b1e3-13a307af1394] No waiting events found dispatching network-vif-plugged-2d89de21-00bd-46da-b7c1-6ac2cbba1982 {{(pid=61923) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 912.979969] env[61923]: WARNING nova.compute.manager [req-6a207006-9392-47d7-b015-e0e38c9d8993 req-4d3f045f-4fa3-40cd-90ae-6dd907897702 service nova] [instance: 62c75dfd-6046-4de1-b1e3-13a307af1394] Received unexpected event network-vif-plugged-2d89de21-00bd-46da-b7c1-6ac2cbba1982 for instance with vm_state building and task_state spawning. [ 912.984454] env[61923]: DEBUG oslo_vmware.api [None req-e087e202-bbb7-4067-8df4-dbcbdaf570ee tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Waiting for the task: (returnval){ [ 912.984454] env[61923]: value = "session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52f7b636-ac5e-3e2d-e948-bd0ef0c53bfb" [ 912.984454] env[61923]: _type = "Task" [ 912.984454] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 912.997584] env[61923]: DEBUG oslo_vmware.api [None req-e087e202-bbb7-4067-8df4-dbcbdaf570ee tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52f7b636-ac5e-3e2d-e948-bd0ef0c53bfb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.091389] env[61923]: DEBUG nova.network.neutron [None req-9a89d0be-1b30-454c-a15f-f7048c8714c5 tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] [instance: 62c75dfd-6046-4de1-b1e3-13a307af1394] Successfully updated port: 2d89de21-00bd-46da-b7c1-6ac2cbba1982 {{(pid=61923) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 913.146885] env[61923]: DEBUG oslo_vmware.api [None req-f37b2e0d-8f0c-42f5-bf08-4bc9d42bb688 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Task: {'id': task-1377770, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.246774] env[61923]: DEBUG oslo_vmware.api [None req-85598005-1527-489f-b9ce-6a2b3821cde4 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Task: {'id': task-1377776, 'name': PowerOffVM_Task} progress is 100%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.260924] env[61923]: DEBUG oslo_concurrency.lockutils [None req-697de1ff-d6e2-4a1e-86f8-0419a0e8777a tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.864s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 913.263154] env[61923]: DEBUG oslo_concurrency.lockutils [None req-69c7da51-15c3-4307-92c9-fa8aea180f17 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 5.218s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 913.263392] env[61923]: DEBUG nova.objects.instance [None req-69c7da51-15c3-4307-92c9-fa8aea180f17 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] Lazy-loading 'resources' on Instance uuid af3c317a-4007-4cea-a060-1e7dde5ce49e {{(pid=61923) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 913.285339] env[61923]: INFO nova.scheduler.client.report [None req-697de1ff-d6e2-4a1e-86f8-0419a0e8777a tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Deleted allocations for instance 59198f4d-4dde-4eaf-9f6c-a962cbe53c6e [ 913.495976] env[61923]: DEBUG oslo_vmware.api [None req-e087e202-bbb7-4067-8df4-dbcbdaf570ee tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52f7b636-ac5e-3e2d-e948-bd0ef0c53bfb, 'name': SearchDatastore_Task, 'duration_secs': 0.011276} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 913.496597] env[61923]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-54b035c3-389d-40e3-8714-ab0a0d7f44ad {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.502577] env[61923]: DEBUG oslo_vmware.api [None req-e087e202-bbb7-4067-8df4-dbcbdaf570ee tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Waiting for the task: (returnval){ [ 913.502577] env[61923]: value = "session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52a07b51-fc4a-1f01-1dc0-db0a000a3c34" [ 913.502577] env[61923]: _type = "Task" [ 913.502577] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 913.510419] env[61923]: DEBUG oslo_vmware.api [None req-e087e202-bbb7-4067-8df4-dbcbdaf570ee tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52a07b51-fc4a-1f01-1dc0-db0a000a3c34, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.595371] env[61923]: DEBUG oslo_concurrency.lockutils [None req-9a89d0be-1b30-454c-a15f-f7048c8714c5 tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Acquiring lock "refresh_cache-62c75dfd-6046-4de1-b1e3-13a307af1394" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 913.595483] env[61923]: DEBUG oslo_concurrency.lockutils [None req-9a89d0be-1b30-454c-a15f-f7048c8714c5 tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Acquired lock "refresh_cache-62c75dfd-6046-4de1-b1e3-13a307af1394" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 913.597308] env[61923]: DEBUG nova.network.neutron [None req-9a89d0be-1b30-454c-a15f-f7048c8714c5 tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] [instance: 62c75dfd-6046-4de1-b1e3-13a307af1394] Building network info cache for instance {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 913.636424] env[61923]: DEBUG oslo_concurrency.lockutils [None req-40624371-b6ea-4bd8-a64e-58d8640198f8 tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Acquiring lock "b779d183-89ae-4e4d-ae99-e514e145ed43" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 913.636778] env[61923]: DEBUG oslo_concurrency.lockutils [None req-40624371-b6ea-4bd8-a64e-58d8640198f8 tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Lock "b779d183-89ae-4e4d-ae99-e514e145ed43" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 913.650742] env[61923]: DEBUG oslo_vmware.api [None req-f37b2e0d-8f0c-42f5-bf08-4bc9d42bb688 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Task: {'id': task-1377770, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.746016] env[61923]: DEBUG oslo_vmware.api [None req-85598005-1527-489f-b9ce-6a2b3821cde4 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Task: {'id': task-1377776, 'name': PowerOffVM_Task} progress is 100%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.792989] env[61923]: DEBUG oslo_concurrency.lockutils [None req-697de1ff-d6e2-4a1e-86f8-0419a0e8777a tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Lock "59198f4d-4dde-4eaf-9f6c-a962cbe53c6e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 9.048s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 913.946446] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc872bf1-08ec-4d67-bdb4-ecf9d4bd317d {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.956027] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36674ce7-0c1c-4e30-bdb4-1a250d188482 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.988550] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b416291-8f27-421d-9e39-d87a96413bb9 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.997018] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64de1117-6c0b-48d3-9f1d-e2f75170b389 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.014325] env[61923]: DEBUG nova.compute.provider_tree [None req-69c7da51-15c3-4307-92c9-fa8aea180f17 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 914.021833] env[61923]: DEBUG oslo_vmware.api [None req-e087e202-bbb7-4067-8df4-dbcbdaf570ee tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52a07b51-fc4a-1f01-1dc0-db0a000a3c34, 'name': SearchDatastore_Task, 'duration_secs': 0.010873} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 914.022708] env[61923]: DEBUG oslo_concurrency.lockutils [None req-e087e202-bbb7-4067-8df4-dbcbdaf570ee tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 914.022978] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-e087e202-bbb7-4067-8df4-dbcbdaf570ee tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk to [datastore1] 624e9207-7f24-4f45-8f52-0fc7a60c1e84/624e9207-7f24-4f45-8f52-0fc7a60c1e84.vmdk {{(pid=61923) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 914.023266] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e31e9274-3eef-445e-b86f-6055908c57ce {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.031646] env[61923]: DEBUG oslo_vmware.api [None req-e087e202-bbb7-4067-8df4-dbcbdaf570ee tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Waiting for the task: (returnval){ [ 914.031646] env[61923]: value = "task-1377777" [ 914.031646] env[61923]: _type = "Task" [ 914.031646] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 914.040991] env[61923]: DEBUG oslo_vmware.api [None req-e087e202-bbb7-4067-8df4-dbcbdaf570ee tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Task: {'id': task-1377777, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.131982] env[61923]: DEBUG nova.network.neutron [None req-9a89d0be-1b30-454c-a15f-f7048c8714c5 tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] [instance: 62c75dfd-6046-4de1-b1e3-13a307af1394] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 914.140776] env[61923]: DEBUG nova.compute.utils [None req-40624371-b6ea-4bd8-a64e-58d8640198f8 tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Using /dev/sd instead of None {{(pid=61923) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 914.153222] env[61923]: DEBUG oslo_vmware.api [None req-f37b2e0d-8f0c-42f5-bf08-4bc9d42bb688 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Task: {'id': task-1377770, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.246690] env[61923]: DEBUG oslo_vmware.api [None req-85598005-1527-489f-b9ce-6a2b3821cde4 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Task: {'id': task-1377776, 'name': PowerOffVM_Task, 'duration_secs': 1.191817} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 914.246989] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-85598005-1527-489f-b9ce-6a2b3821cde4 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: 03c1e8ea-a322-46fa-9b07-e283080a7871] Powered off the VM {{(pid=61923) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 914.247192] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-85598005-1527-489f-b9ce-6a2b3821cde4 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: 03c1e8ea-a322-46fa-9b07-e283080a7871] Unregistering the VM {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 914.247619] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-61dd7181-fa4e-4052-9fff-e4141be51cab {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.298164] env[61923]: DEBUG nova.network.neutron [None req-9a89d0be-1b30-454c-a15f-f7048c8714c5 tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] [instance: 62c75dfd-6046-4de1-b1e3-13a307af1394] Updating instance_info_cache with network_info: [{"id": "2d89de21-00bd-46da-b7c1-6ac2cbba1982", "address": "fa:16:3e:31:cd:0e", "network": {"id": "b8c2e837-0d64-4b07-9cc5-a66c8e8db1e9", "bridge": "br-int", "label": "tempest-ImagesTestJSON-133215580-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7abfe95660904cd4b8824257795d8e8a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f2c424c9-6446-4b2a-af8c-4d9c29117c39", "external-id": "nsx-vlan-transportzone-437", "segmentation_id": 437, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2d89de21-00", "ovs_interfaceid": "2d89de21-00bd-46da-b7c1-6ac2cbba1982", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 914.339319] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-85598005-1527-489f-b9ce-6a2b3821cde4 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: 03c1e8ea-a322-46fa-9b07-e283080a7871] Unregistered the VM {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 914.339600] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-85598005-1527-489f-b9ce-6a2b3821cde4 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: 03c1e8ea-a322-46fa-9b07-e283080a7871] Deleting contents of the VM from datastore datastore1 {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 914.339891] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-85598005-1527-489f-b9ce-6a2b3821cde4 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Deleting the datastore file [datastore1] 03c1e8ea-a322-46fa-9b07-e283080a7871 {{(pid=61923) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 914.340081] env[61923]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-aef79f3e-b19a-4873-9303-6f83de275ba7 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.347823] env[61923]: DEBUG oslo_vmware.api [None req-85598005-1527-489f-b9ce-6a2b3821cde4 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Waiting for the task: (returnval){ [ 914.347823] env[61923]: value = "task-1377779" [ 914.347823] env[61923]: _type = "Task" [ 914.347823] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 914.356131] env[61923]: DEBUG oslo_vmware.api [None req-85598005-1527-489f-b9ce-6a2b3821cde4 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Task: {'id': task-1377779, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.517666] env[61923]: DEBUG nova.scheduler.client.report [None req-69c7da51-15c3-4307-92c9-fa8aea180f17 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 914.544461] env[61923]: DEBUG oslo_vmware.api [None req-e087e202-bbb7-4067-8df4-dbcbdaf570ee tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Task: {'id': task-1377777, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.648207] env[61923]: DEBUG oslo_concurrency.lockutils [None req-40624371-b6ea-4bd8-a64e-58d8640198f8 tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Lock "b779d183-89ae-4e4d-ae99-e514e145ed43" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.011s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 914.653855] env[61923]: DEBUG oslo_vmware.api [None req-f37b2e0d-8f0c-42f5-bf08-4bc9d42bb688 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Task: {'id': task-1377770, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.801492] env[61923]: DEBUG oslo_concurrency.lockutils [None req-9a89d0be-1b30-454c-a15f-f7048c8714c5 tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Releasing lock "refresh_cache-62c75dfd-6046-4de1-b1e3-13a307af1394" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 914.801956] env[61923]: DEBUG nova.compute.manager [None req-9a89d0be-1b30-454c-a15f-f7048c8714c5 tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] [instance: 62c75dfd-6046-4de1-b1e3-13a307af1394] Instance network_info: |[{"id": "2d89de21-00bd-46da-b7c1-6ac2cbba1982", "address": "fa:16:3e:31:cd:0e", "network": {"id": "b8c2e837-0d64-4b07-9cc5-a66c8e8db1e9", "bridge": "br-int", "label": "tempest-ImagesTestJSON-133215580-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7abfe95660904cd4b8824257795d8e8a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f2c424c9-6446-4b2a-af8c-4d9c29117c39", "external-id": "nsx-vlan-transportzone-437", "segmentation_id": 437, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2d89de21-00", "ovs_interfaceid": "2d89de21-00bd-46da-b7c1-6ac2cbba1982", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61923) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 914.802439] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-9a89d0be-1b30-454c-a15f-f7048c8714c5 tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] [instance: 62c75dfd-6046-4de1-b1e3-13a307af1394] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:31:cd:0e', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'f2c424c9-6446-4b2a-af8c-4d9c29117c39', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '2d89de21-00bd-46da-b7c1-6ac2cbba1982', 'vif_model': 'vmxnet3'}] {{(pid=61923) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 914.810642] env[61923]: DEBUG oslo.service.loopingcall [None req-9a89d0be-1b30-454c-a15f-f7048c8714c5 tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61923) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 914.810912] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 62c75dfd-6046-4de1-b1e3-13a307af1394] Creating VM on the ESX host {{(pid=61923) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 914.811160] env[61923]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-2b5e4b64-8123-4c34-bd08-1b10c7e7449f {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.834157] env[61923]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 914.834157] env[61923]: value = "task-1377780" [ 914.834157] env[61923]: _type = "Task" [ 914.834157] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 914.842653] env[61923]: DEBUG oslo_vmware.api [-] Task: {'id': task-1377780, 'name': CreateVM_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.857865] env[61923]: DEBUG oslo_vmware.api [None req-85598005-1527-489f-b9ce-6a2b3821cde4 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Task: {'id': task-1377779, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.002723] env[61923]: DEBUG nova.compute.manager [req-92c72b97-70ae-4fac-aad3-9024b33d34ad req-b7ca80f0-f8ab-4f6f-8f5d-0951a2af6a44 service nova] [instance: 62c75dfd-6046-4de1-b1e3-13a307af1394] Received event network-changed-2d89de21-00bd-46da-b7c1-6ac2cbba1982 {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 915.002995] env[61923]: DEBUG nova.compute.manager [req-92c72b97-70ae-4fac-aad3-9024b33d34ad req-b7ca80f0-f8ab-4f6f-8f5d-0951a2af6a44 service nova] [instance: 62c75dfd-6046-4de1-b1e3-13a307af1394] Refreshing instance network info cache due to event network-changed-2d89de21-00bd-46da-b7c1-6ac2cbba1982. {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 915.003278] env[61923]: DEBUG oslo_concurrency.lockutils [req-92c72b97-70ae-4fac-aad3-9024b33d34ad req-b7ca80f0-f8ab-4f6f-8f5d-0951a2af6a44 service nova] Acquiring lock "refresh_cache-62c75dfd-6046-4de1-b1e3-13a307af1394" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 915.003444] env[61923]: DEBUG oslo_concurrency.lockutils [req-92c72b97-70ae-4fac-aad3-9024b33d34ad req-b7ca80f0-f8ab-4f6f-8f5d-0951a2af6a44 service nova] Acquired lock "refresh_cache-62c75dfd-6046-4de1-b1e3-13a307af1394" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 915.003658] env[61923]: DEBUG nova.network.neutron [req-92c72b97-70ae-4fac-aad3-9024b33d34ad req-b7ca80f0-f8ab-4f6f-8f5d-0951a2af6a44 service nova] [instance: 62c75dfd-6046-4de1-b1e3-13a307af1394] Refreshing network info cache for port 2d89de21-00bd-46da-b7c1-6ac2cbba1982 {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 915.022510] env[61923]: DEBUG oslo_concurrency.lockutils [None req-69c7da51-15c3-4307-92c9-fa8aea180f17 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.759s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 915.025038] env[61923]: DEBUG oslo_concurrency.lockutils [None req-15de031b-9972-4bde-a9c6-0e58042e23f7 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 4.343s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 915.025301] env[61923]: DEBUG nova.objects.instance [None req-15de031b-9972-4bde-a9c6-0e58042e23f7 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Lazy-loading 'resources' on Instance uuid bac71328-3796-4a65-aa41-ccda93bc7926 {{(pid=61923) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 915.045211] env[61923]: DEBUG oslo_vmware.api [None req-e087e202-bbb7-4067-8df4-dbcbdaf570ee tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Task: {'id': task-1377777, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.758985} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 915.046228] env[61923]: INFO nova.scheduler.client.report [None req-69c7da51-15c3-4307-92c9-fa8aea180f17 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] Deleted allocations for instance af3c317a-4007-4cea-a060-1e7dde5ce49e [ 915.047338] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-e087e202-bbb7-4067-8df4-dbcbdaf570ee tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk to [datastore1] 624e9207-7f24-4f45-8f52-0fc7a60c1e84/624e9207-7f24-4f45-8f52-0fc7a60c1e84.vmdk {{(pid=61923) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 915.047599] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-e087e202-bbb7-4067-8df4-dbcbdaf570ee tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] [instance: 624e9207-7f24-4f45-8f52-0fc7a60c1e84] Extending root virtual disk to 1048576 {{(pid=61923) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 915.050132] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-311391df-419a-44c7-a4a3-142786df267a {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.058939] env[61923]: DEBUG oslo_vmware.api [None req-e087e202-bbb7-4067-8df4-dbcbdaf570ee tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Waiting for the task: (returnval){ [ 915.058939] env[61923]: value = "task-1377781" [ 915.058939] env[61923]: _type = "Task" [ 915.058939] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 915.067358] env[61923]: DEBUG oslo_vmware.api [None req-e087e202-bbb7-4067-8df4-dbcbdaf570ee tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Task: {'id': task-1377781, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.151040] env[61923]: DEBUG oslo_vmware.api [None req-f37b2e0d-8f0c-42f5-bf08-4bc9d42bb688 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Task: {'id': task-1377770, 'name': ReconfigVM_Task} progress is 18%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.347219] env[61923]: DEBUG oslo_vmware.api [-] Task: {'id': task-1377780, 'name': CreateVM_Task, 'duration_secs': 0.401178} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 915.347550] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 62c75dfd-6046-4de1-b1e3-13a307af1394] Created VM on the ESX host {{(pid=61923) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 915.348376] env[61923]: DEBUG oslo_concurrency.lockutils [None req-9a89d0be-1b30-454c-a15f-f7048c8714c5 tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 915.348577] env[61923]: DEBUG oslo_concurrency.lockutils [None req-9a89d0be-1b30-454c-a15f-f7048c8714c5 tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 915.349527] env[61923]: DEBUG oslo_concurrency.lockutils [None req-9a89d0be-1b30-454c-a15f-f7048c8714c5 tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 915.350269] env[61923]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3e346bf9-4602-48e3-b3fd-d9cda83cca6f {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.359073] env[61923]: DEBUG oslo_vmware.api [None req-9a89d0be-1b30-454c-a15f-f7048c8714c5 tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Waiting for the task: (returnval){ [ 915.359073] env[61923]: value = "session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52f3b625-39bf-fdd1-58fd-ee6826274275" [ 915.359073] env[61923]: _type = "Task" [ 915.359073] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 915.362411] env[61923]: DEBUG oslo_vmware.api [None req-85598005-1527-489f-b9ce-6a2b3821cde4 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Task: {'id': task-1377779, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.582565} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 915.365570] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-85598005-1527-489f-b9ce-6a2b3821cde4 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Deleted the datastore file {{(pid=61923) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 915.365793] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-85598005-1527-489f-b9ce-6a2b3821cde4 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: 03c1e8ea-a322-46fa-9b07-e283080a7871] Deleted contents of the VM from datastore datastore1 {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 915.366039] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-85598005-1527-489f-b9ce-6a2b3821cde4 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: 03c1e8ea-a322-46fa-9b07-e283080a7871] Instance destroyed {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 915.366220] env[61923]: INFO nova.compute.manager [None req-85598005-1527-489f-b9ce-6a2b3821cde4 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: 03c1e8ea-a322-46fa-9b07-e283080a7871] Took 2.65 seconds to destroy the instance on the hypervisor. [ 915.366490] env[61923]: DEBUG oslo.service.loopingcall [None req-85598005-1527-489f-b9ce-6a2b3821cde4 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61923) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 915.366699] env[61923]: DEBUG nova.compute.manager [-] [instance: 03c1e8ea-a322-46fa-9b07-e283080a7871] Deallocating network for instance {{(pid=61923) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 915.366795] env[61923]: DEBUG nova.network.neutron [-] [instance: 03c1e8ea-a322-46fa-9b07-e283080a7871] deallocate_for_instance() {{(pid=61923) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 915.380704] env[61923]: DEBUG oslo_vmware.api [None req-9a89d0be-1b30-454c-a15f-f7048c8714c5 tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52f3b625-39bf-fdd1-58fd-ee6826274275, 'name': SearchDatastore_Task, 'duration_secs': 0.017362} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 915.381052] env[61923]: DEBUG oslo_concurrency.lockutils [None req-9a89d0be-1b30-454c-a15f-f7048c8714c5 tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 915.381307] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-9a89d0be-1b30-454c-a15f-f7048c8714c5 tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] [instance: 62c75dfd-6046-4de1-b1e3-13a307af1394] Processing image 0f153f63-ae0a-45b1-b7f5-7f9b673c947f {{(pid=61923) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 915.381549] env[61923]: DEBUG oslo_concurrency.lockutils [None req-9a89d0be-1b30-454c-a15f-f7048c8714c5 tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 915.381736] env[61923]: DEBUG oslo_concurrency.lockutils [None req-9a89d0be-1b30-454c-a15f-f7048c8714c5 tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 915.381950] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-9a89d0be-1b30-454c-a15f-f7048c8714c5 tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61923) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 915.382252] env[61923]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-4ef9f7a6-c997-49e0-9dac-14174f1884c0 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.391698] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-9a89d0be-1b30-454c-a15f-f7048c8714c5 tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61923) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 915.391920] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-9a89d0be-1b30-454c-a15f-f7048c8714c5 tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61923) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 915.392695] env[61923]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c937b64d-16f4-4b1e-9235-a6a2ab03db55 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.398794] env[61923]: DEBUG oslo_vmware.api [None req-9a89d0be-1b30-454c-a15f-f7048c8714c5 tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Waiting for the task: (returnval){ [ 915.398794] env[61923]: value = "session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52c16982-0cce-ddb0-a7e8-210f120422b9" [ 915.398794] env[61923]: _type = "Task" [ 915.398794] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 915.410558] env[61923]: DEBUG oslo_vmware.api [None req-9a89d0be-1b30-454c-a15f-f7048c8714c5 tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52c16982-0cce-ddb0-a7e8-210f120422b9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.559335] env[61923]: DEBUG oslo_concurrency.lockutils [None req-69c7da51-15c3-4307-92c9-fa8aea180f17 tempest-AttachVolumeNegativeTest-1855469338 tempest-AttachVolumeNegativeTest-1855469338-project-member] Lock "af3c317a-4007-4cea-a060-1e7dde5ce49e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 11.070s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 915.579800] env[61923]: DEBUG oslo_vmware.api [None req-e087e202-bbb7-4067-8df4-dbcbdaf570ee tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Task: {'id': task-1377781, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.080264} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 915.580085] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-e087e202-bbb7-4067-8df4-dbcbdaf570ee tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] [instance: 624e9207-7f24-4f45-8f52-0fc7a60c1e84] Extended root virtual disk {{(pid=61923) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 915.581072] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b95d1d3-866d-4e9a-9a31-f5bb969b60b7 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.616563] env[61923]: DEBUG nova.virt.vmwareapi.volumeops [None req-e087e202-bbb7-4067-8df4-dbcbdaf570ee tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] [instance: 624e9207-7f24-4f45-8f52-0fc7a60c1e84] Reconfiguring VM instance instance-00000055 to attach disk [datastore1] 624e9207-7f24-4f45-8f52-0fc7a60c1e84/624e9207-7f24-4f45-8f52-0fc7a60c1e84.vmdk or device None with type sparse {{(pid=61923) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 915.621691] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3b28d490-f37f-4625-8856-0b00e7ace50e {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.651311] env[61923]: DEBUG oslo_vmware.api [None req-e087e202-bbb7-4067-8df4-dbcbdaf570ee tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Waiting for the task: (returnval){ [ 915.651311] env[61923]: value = "task-1377782" [ 915.651311] env[61923]: _type = "Task" [ 915.651311] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 915.655066] env[61923]: DEBUG oslo_vmware.api [None req-f37b2e0d-8f0c-42f5-bf08-4bc9d42bb688 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Task: {'id': task-1377770, 'name': ReconfigVM_Task, 'duration_secs': 5.79416} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 915.661084] env[61923]: DEBUG oslo_concurrency.lockutils [None req-f37b2e0d-8f0c-42f5-bf08-4bc9d42bb688 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Releasing lock "6012d3b5-739c-4762-9bb4-09c51171dcd7" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 915.661341] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-f37b2e0d-8f0c-42f5-bf08-4bc9d42bb688 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] [instance: 6012d3b5-739c-4762-9bb4-09c51171dcd7] Reconfigured VM to detach interface {{(pid=61923) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1978}} [ 915.670069] env[61923]: DEBUG oslo_vmware.api [None req-e087e202-bbb7-4067-8df4-dbcbdaf570ee tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Task: {'id': task-1377782, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.708738] env[61923]: DEBUG oslo_concurrency.lockutils [None req-40624371-b6ea-4bd8-a64e-58d8640198f8 tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Acquiring lock "b779d183-89ae-4e4d-ae99-e514e145ed43" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 915.711268] env[61923]: DEBUG oslo_concurrency.lockutils [None req-40624371-b6ea-4bd8-a64e-58d8640198f8 tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Lock "b779d183-89ae-4e4d-ae99-e514e145ed43" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 915.711268] env[61923]: INFO nova.compute.manager [None req-40624371-b6ea-4bd8-a64e-58d8640198f8 tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] [instance: b779d183-89ae-4e4d-ae99-e514e145ed43] Attaching volume 84fb906f-352f-4eae-b044-c86e8dda3254 to /dev/sdb [ 915.751462] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44d677bf-d2f4-4672-9e4a-ff33b248cd2b {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.761107] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79b67558-8975-4723-a08b-12d14dd8c677 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.778643] env[61923]: DEBUG nova.virt.block_device [None req-40624371-b6ea-4bd8-a64e-58d8640198f8 tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] [instance: b779d183-89ae-4e4d-ae99-e514e145ed43] Updating existing volume attachment record: 9d9b269c-b7f3-4468-bf90-7935a7a3fd68 {{(pid=61923) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 915.790855] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8d73a02-035a-483a-99a8-0a0d07e2f32d {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.802200] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f22c9bdf-5922-4482-abc9-8fbb95133a41 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.838345] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fdca8a97-d73c-41a8-9dd7-3c5e73247de2 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.846776] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d30b9e2-42ed-4923-addb-59deff2315c0 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.851564] env[61923]: DEBUG nova.network.neutron [req-92c72b97-70ae-4fac-aad3-9024b33d34ad req-b7ca80f0-f8ab-4f6f-8f5d-0951a2af6a44 service nova] [instance: 62c75dfd-6046-4de1-b1e3-13a307af1394] Updated VIF entry in instance network info cache for port 2d89de21-00bd-46da-b7c1-6ac2cbba1982. {{(pid=61923) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 915.851908] env[61923]: DEBUG nova.network.neutron [req-92c72b97-70ae-4fac-aad3-9024b33d34ad req-b7ca80f0-f8ab-4f6f-8f5d-0951a2af6a44 service nova] [instance: 62c75dfd-6046-4de1-b1e3-13a307af1394] Updating instance_info_cache with network_info: [{"id": "2d89de21-00bd-46da-b7c1-6ac2cbba1982", "address": "fa:16:3e:31:cd:0e", "network": {"id": "b8c2e837-0d64-4b07-9cc5-a66c8e8db1e9", "bridge": "br-int", "label": "tempest-ImagesTestJSON-133215580-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7abfe95660904cd4b8824257795d8e8a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "f2c424c9-6446-4b2a-af8c-4d9c29117c39", "external-id": "nsx-vlan-transportzone-437", "segmentation_id": 437, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2d89de21-00", "ovs_interfaceid": "2d89de21-00bd-46da-b7c1-6ac2cbba1982", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 915.867173] env[61923]: DEBUG nova.compute.provider_tree [None req-15de031b-9972-4bde-a9c6-0e58042e23f7 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 915.911550] env[61923]: DEBUG oslo_vmware.api [None req-9a89d0be-1b30-454c-a15f-f7048c8714c5 tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52c16982-0cce-ddb0-a7e8-210f120422b9, 'name': SearchDatastore_Task, 'duration_secs': 0.010652} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 915.912421] env[61923]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5c7bacd6-4d80-4a83-9aee-0786d547cb35 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.918627] env[61923]: DEBUG oslo_vmware.api [None req-9a89d0be-1b30-454c-a15f-f7048c8714c5 tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Waiting for the task: (returnval){ [ 915.918627] env[61923]: value = "session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52d9118f-ae6d-bc86-db58-d930530513eb" [ 915.918627] env[61923]: _type = "Task" [ 915.918627] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 915.927250] env[61923]: DEBUG oslo_vmware.api [None req-9a89d0be-1b30-454c-a15f-f7048c8714c5 tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52d9118f-ae6d-bc86-db58-d930530513eb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.121751] env[61923]: DEBUG nova.network.neutron [-] [instance: 03c1e8ea-a322-46fa-9b07-e283080a7871] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 916.167673] env[61923]: DEBUG oslo_vmware.api [None req-e087e202-bbb7-4067-8df4-dbcbdaf570ee tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Task: {'id': task-1377782, 'name': ReconfigVM_Task, 'duration_secs': 0.274275} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 916.168402] env[61923]: DEBUG nova.virt.vmwareapi.volumeops [None req-e087e202-bbb7-4067-8df4-dbcbdaf570ee tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] [instance: 624e9207-7f24-4f45-8f52-0fc7a60c1e84] Reconfigured VM instance instance-00000055 to attach disk [datastore1] 624e9207-7f24-4f45-8f52-0fc7a60c1e84/624e9207-7f24-4f45-8f52-0fc7a60c1e84.vmdk or device None with type sparse {{(pid=61923) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 916.169101] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b535ecbb-d7f0-47c5-96cf-a0dfe54cef31 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.176267] env[61923]: DEBUG oslo_vmware.api [None req-e087e202-bbb7-4067-8df4-dbcbdaf570ee tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Waiting for the task: (returnval){ [ 916.176267] env[61923]: value = "task-1377787" [ 916.176267] env[61923]: _type = "Task" [ 916.176267] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 916.185565] env[61923]: DEBUG oslo_vmware.api [None req-e087e202-bbb7-4067-8df4-dbcbdaf570ee tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Task: {'id': task-1377787, 'name': Rename_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.354976] env[61923]: DEBUG oslo_concurrency.lockutils [req-92c72b97-70ae-4fac-aad3-9024b33d34ad req-b7ca80f0-f8ab-4f6f-8f5d-0951a2af6a44 service nova] Releasing lock "refresh_cache-62c75dfd-6046-4de1-b1e3-13a307af1394" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 916.370548] env[61923]: DEBUG nova.scheduler.client.report [None req-15de031b-9972-4bde-a9c6-0e58042e23f7 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 916.430480] env[61923]: DEBUG oslo_vmware.api [None req-9a89d0be-1b30-454c-a15f-f7048c8714c5 tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52d9118f-ae6d-bc86-db58-d930530513eb, 'name': SearchDatastore_Task, 'duration_secs': 0.017078} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 916.430713] env[61923]: DEBUG oslo_concurrency.lockutils [None req-9a89d0be-1b30-454c-a15f-f7048c8714c5 tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 916.431042] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-9a89d0be-1b30-454c-a15f-f7048c8714c5 tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk to [datastore1] 62c75dfd-6046-4de1-b1e3-13a307af1394/62c75dfd-6046-4de1-b1e3-13a307af1394.vmdk {{(pid=61923) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 916.431348] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-6b6a9330-59f2-4e89-9b03-c6bec3bd8120 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.439498] env[61923]: DEBUG oslo_vmware.api [None req-9a89d0be-1b30-454c-a15f-f7048c8714c5 tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Waiting for the task: (returnval){ [ 916.439498] env[61923]: value = "task-1377788" [ 916.439498] env[61923]: _type = "Task" [ 916.439498] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 916.449670] env[61923]: DEBUG oslo_vmware.api [None req-9a89d0be-1b30-454c-a15f-f7048c8714c5 tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Task: {'id': task-1377788, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.625654] env[61923]: INFO nova.compute.manager [-] [instance: 03c1e8ea-a322-46fa-9b07-e283080a7871] Took 1.26 seconds to deallocate network for instance. [ 916.687866] env[61923]: DEBUG oslo_vmware.api [None req-e087e202-bbb7-4067-8df4-dbcbdaf570ee tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Task: {'id': task-1377787, 'name': Rename_Task, 'duration_secs': 0.223222} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 916.688202] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-e087e202-bbb7-4067-8df4-dbcbdaf570ee tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] [instance: 624e9207-7f24-4f45-8f52-0fc7a60c1e84] Powering on the VM {{(pid=61923) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 916.688502] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-53d9dddb-6517-4442-bc77-2694c222603f {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.696565] env[61923]: DEBUG oslo_vmware.api [None req-e087e202-bbb7-4067-8df4-dbcbdaf570ee tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Waiting for the task: (returnval){ [ 916.696565] env[61923]: value = "task-1377789" [ 916.696565] env[61923]: _type = "Task" [ 916.696565] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 916.706130] env[61923]: DEBUG oslo_vmware.api [None req-e087e202-bbb7-4067-8df4-dbcbdaf570ee tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Task: {'id': task-1377789, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.875947] env[61923]: DEBUG oslo_concurrency.lockutils [None req-15de031b-9972-4bde-a9c6-0e58042e23f7 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.851s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 916.904875] env[61923]: INFO nova.scheduler.client.report [None req-15de031b-9972-4bde-a9c6-0e58042e23f7 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Deleted allocations for instance bac71328-3796-4a65-aa41-ccda93bc7926 [ 916.950933] env[61923]: DEBUG oslo_vmware.api [None req-9a89d0be-1b30-454c-a15f-f7048c8714c5 tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Task: {'id': task-1377788, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.033474] env[61923]: DEBUG nova.compute.manager [req-fcfc2be6-957c-4530-ae4e-f59b57e78a28 req-b196646c-14f7-46e7-abcb-55f9679794d0 service nova] [instance: 03c1e8ea-a322-46fa-9b07-e283080a7871] Received event network-vif-deleted-37dd225f-bb7f-471b-8484-5387c0e22183 {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 917.043612] env[61923]: DEBUG oslo_concurrency.lockutils [None req-f37b2e0d-8f0c-42f5-bf08-4bc9d42bb688 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Acquiring lock "refresh_cache-6012d3b5-739c-4762-9bb4-09c51171dcd7" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 917.043907] env[61923]: DEBUG oslo_concurrency.lockutils [None req-f37b2e0d-8f0c-42f5-bf08-4bc9d42bb688 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Acquired lock "refresh_cache-6012d3b5-739c-4762-9bb4-09c51171dcd7" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 917.044152] env[61923]: DEBUG nova.network.neutron [None req-f37b2e0d-8f0c-42f5-bf08-4bc9d42bb688 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] [instance: 6012d3b5-739c-4762-9bb4-09c51171dcd7] Building network info cache for instance {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 917.132341] env[61923]: DEBUG oslo_concurrency.lockutils [None req-85598005-1527-489f-b9ce-6a2b3821cde4 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 917.132754] env[61923]: DEBUG oslo_concurrency.lockutils [None req-85598005-1527-489f-b9ce-6a2b3821cde4 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 917.133150] env[61923]: DEBUG nova.objects.instance [None req-85598005-1527-489f-b9ce-6a2b3821cde4 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Lazy-loading 'resources' on Instance uuid 03c1e8ea-a322-46fa-9b07-e283080a7871 {{(pid=61923) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 917.208671] env[61923]: DEBUG oslo_vmware.api [None req-e087e202-bbb7-4067-8df4-dbcbdaf570ee tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Task: {'id': task-1377789, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.413457] env[61923]: DEBUG oslo_concurrency.lockutils [None req-15de031b-9972-4bde-a9c6-0e58042e23f7 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Lock "bac71328-3796-4a65-aa41-ccda93bc7926" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 10.297s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 917.454262] env[61923]: DEBUG oslo_vmware.api [None req-9a89d0be-1b30-454c-a15f-f7048c8714c5 tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Task: {'id': task-1377788, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.649001} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 917.454744] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-9a89d0be-1b30-454c-a15f-f7048c8714c5 tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk to [datastore1] 62c75dfd-6046-4de1-b1e3-13a307af1394/62c75dfd-6046-4de1-b1e3-13a307af1394.vmdk {{(pid=61923) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 917.454960] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-9a89d0be-1b30-454c-a15f-f7048c8714c5 tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] [instance: 62c75dfd-6046-4de1-b1e3-13a307af1394] Extending root virtual disk to 1048576 {{(pid=61923) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 917.455324] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-d9700beb-e444-453f-a43f-9fa79758628f {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.464709] env[61923]: DEBUG oslo_vmware.api [None req-9a89d0be-1b30-454c-a15f-f7048c8714c5 tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Waiting for the task: (returnval){ [ 917.464709] env[61923]: value = "task-1377790" [ 917.464709] env[61923]: _type = "Task" [ 917.464709] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 917.477239] env[61923]: DEBUG oslo_vmware.api [None req-9a89d0be-1b30-454c-a15f-f7048c8714c5 tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Task: {'id': task-1377790, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.534650] env[61923]: DEBUG oslo_concurrency.lockutils [None req-518b7c19-1088-43f5-9762-d7516e0ed822 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Acquiring lock "6012d3b5-739c-4762-9bb4-09c51171dcd7" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 917.534899] env[61923]: DEBUG oslo_concurrency.lockutils [None req-518b7c19-1088-43f5-9762-d7516e0ed822 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Lock "6012d3b5-739c-4762-9bb4-09c51171dcd7" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 917.535144] env[61923]: DEBUG oslo_concurrency.lockutils [None req-518b7c19-1088-43f5-9762-d7516e0ed822 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Acquiring lock "6012d3b5-739c-4762-9bb4-09c51171dcd7-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 917.535341] env[61923]: DEBUG oslo_concurrency.lockutils [None req-518b7c19-1088-43f5-9762-d7516e0ed822 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Lock "6012d3b5-739c-4762-9bb4-09c51171dcd7-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 917.535551] env[61923]: DEBUG oslo_concurrency.lockutils [None req-518b7c19-1088-43f5-9762-d7516e0ed822 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Lock "6012d3b5-739c-4762-9bb4-09c51171dcd7-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 917.537873] env[61923]: INFO nova.compute.manager [None req-518b7c19-1088-43f5-9762-d7516e0ed822 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] [instance: 6012d3b5-739c-4762-9bb4-09c51171dcd7] Terminating instance [ 917.539792] env[61923]: DEBUG nova.compute.manager [None req-518b7c19-1088-43f5-9762-d7516e0ed822 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] [instance: 6012d3b5-739c-4762-9bb4-09c51171dcd7] Start destroying the instance on the hypervisor. {{(pid=61923) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 917.540014] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-518b7c19-1088-43f5-9762-d7516e0ed822 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] [instance: 6012d3b5-739c-4762-9bb4-09c51171dcd7] Destroying instance {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 917.541058] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c155a947-1040-4309-9f07-7dc14f838b1b {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.551360] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-518b7c19-1088-43f5-9762-d7516e0ed822 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] [instance: 6012d3b5-739c-4762-9bb4-09c51171dcd7] Powering off the VM {{(pid=61923) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 917.551994] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3241435f-c855-44c7-9c69-228a25ca3f44 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.560453] env[61923]: DEBUG oslo_vmware.api [None req-518b7c19-1088-43f5-9762-d7516e0ed822 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Waiting for the task: (returnval){ [ 917.560453] env[61923]: value = "task-1377791" [ 917.560453] env[61923]: _type = "Task" [ 917.560453] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 917.571461] env[61923]: DEBUG oslo_vmware.api [None req-518b7c19-1088-43f5-9762-d7516e0ed822 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Task: {'id': task-1377791, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.710058] env[61923]: DEBUG oslo_vmware.api [None req-e087e202-bbb7-4067-8df4-dbcbdaf570ee tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Task: {'id': task-1377789, 'name': PowerOnVM_Task, 'duration_secs': 0.544323} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 917.712759] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-e087e202-bbb7-4067-8df4-dbcbdaf570ee tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] [instance: 624e9207-7f24-4f45-8f52-0fc7a60c1e84] Powered on the VM {{(pid=61923) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 917.713027] env[61923]: DEBUG nova.compute.manager [None req-e087e202-bbb7-4067-8df4-dbcbdaf570ee tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] [instance: 624e9207-7f24-4f45-8f52-0fc7a60c1e84] Checking state {{(pid=61923) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 917.713906] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9acc938-7485-4783-8c05-a68a0e3940cc {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.859369] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86971224-b047-490d-9d95-2643fc559000 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.870363] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4be1b869-65c3-4d7c-9998-50e1c7966944 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.905345] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27cda57c-6a68-4337-901c-2205a60eabc8 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.914886] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06b9db0b-7cb5-476a-a2c5-bca089bf8368 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.934187] env[61923]: DEBUG nova.compute.provider_tree [None req-85598005-1527-489f-b9ce-6a2b3821cde4 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Updating inventory in ProviderTree for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 177, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 917.963441] env[61923]: INFO nova.network.neutron [None req-f37b2e0d-8f0c-42f5-bf08-4bc9d42bb688 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] [instance: 6012d3b5-739c-4762-9bb4-09c51171dcd7] Port b395b183-1c30-42de-929f-690ea10d9d75 from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 917.963820] env[61923]: DEBUG nova.network.neutron [None req-f37b2e0d-8f0c-42f5-bf08-4bc9d42bb688 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] [instance: 6012d3b5-739c-4762-9bb4-09c51171dcd7] Updating instance_info_cache with network_info: [{"id": "269a41b0-75e7-46af-9497-8eae3102df11", "address": "fa:16:3e:e1:02:b8", "network": {"id": "a8b7e1b4-1ec6-4226-bdd6-535dbce450d2", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1928758125-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.146", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "912300a09b80452e85e3cd13c4a644cc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "7806fe18-2b89-4386-87b1-f22876f82af2", "external-id": "nsx-vlan-transportzone-727", "segmentation_id": 727, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap269a41b0-75", "ovs_interfaceid": "269a41b0-75e7-46af-9497-8eae3102df11", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 917.975940] env[61923]: DEBUG oslo_vmware.api [None req-9a89d0be-1b30-454c-a15f-f7048c8714c5 tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Task: {'id': task-1377790, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.115925} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 917.976245] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-9a89d0be-1b30-454c-a15f-f7048c8714c5 tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] [instance: 62c75dfd-6046-4de1-b1e3-13a307af1394] Extended root virtual disk {{(pid=61923) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 917.977131] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83e32dca-d81c-4319-9942-6aeab6c65072 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.001638] env[61923]: DEBUG nova.virt.vmwareapi.volumeops [None req-9a89d0be-1b30-454c-a15f-f7048c8714c5 tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] [instance: 62c75dfd-6046-4de1-b1e3-13a307af1394] Reconfiguring VM instance instance-00000057 to attach disk [datastore1] 62c75dfd-6046-4de1-b1e3-13a307af1394/62c75dfd-6046-4de1-b1e3-13a307af1394.vmdk or device None with type sparse {{(pid=61923) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 918.002655] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6a1df848-b4e6-4995-a510-eaf09afc6308 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.026058] env[61923]: DEBUG oslo_vmware.api [None req-9a89d0be-1b30-454c-a15f-f7048c8714c5 tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Waiting for the task: (returnval){ [ 918.026058] env[61923]: value = "task-1377792" [ 918.026058] env[61923]: _type = "Task" [ 918.026058] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 918.035424] env[61923]: DEBUG oslo_vmware.api [None req-9a89d0be-1b30-454c-a15f-f7048c8714c5 tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Task: {'id': task-1377792, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.070897] env[61923]: DEBUG oslo_vmware.api [None req-518b7c19-1088-43f5-9762-d7516e0ed822 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Task: {'id': task-1377791, 'name': PowerOffVM_Task, 'duration_secs': 0.268281} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 918.071212] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-518b7c19-1088-43f5-9762-d7516e0ed822 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] [instance: 6012d3b5-739c-4762-9bb4-09c51171dcd7] Powered off the VM {{(pid=61923) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 918.071385] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-518b7c19-1088-43f5-9762-d7516e0ed822 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] [instance: 6012d3b5-739c-4762-9bb4-09c51171dcd7] Unregistering the VM {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 918.071663] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-7e8c2500-3d92-48fa-b81b-425ff6e4a71c {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.144556] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-518b7c19-1088-43f5-9762-d7516e0ed822 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] [instance: 6012d3b5-739c-4762-9bb4-09c51171dcd7] Unregistered the VM {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 918.144771] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-518b7c19-1088-43f5-9762-d7516e0ed822 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] [instance: 6012d3b5-739c-4762-9bb4-09c51171dcd7] Deleting contents of the VM from datastore datastore2 {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 918.144963] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-518b7c19-1088-43f5-9762-d7516e0ed822 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Deleting the datastore file [datastore2] 6012d3b5-739c-4762-9bb4-09c51171dcd7 {{(pid=61923) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 918.145263] env[61923]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-4e2a8796-1d8f-4846-9f6f-4283f5d2b60e {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.153028] env[61923]: DEBUG oslo_vmware.api [None req-518b7c19-1088-43f5-9762-d7516e0ed822 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Waiting for the task: (returnval){ [ 918.153028] env[61923]: value = "task-1377794" [ 918.153028] env[61923]: _type = "Task" [ 918.153028] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 918.161675] env[61923]: DEBUG oslo_vmware.api [None req-518b7c19-1088-43f5-9762-d7516e0ed822 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Task: {'id': task-1377794, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.238935] env[61923]: DEBUG oslo_concurrency.lockutils [None req-e087e202-bbb7-4067-8df4-dbcbdaf570ee tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 918.377668] env[61923]: DEBUG oslo_vmware.rw_handles [None req-81989905-2c22-4589-a2f0-03b8d81c0d2f tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/522995c9-80ac-b69a-62c7-1e4b40c5a0df/disk-0.vmdk. {{(pid=61923) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 918.378768] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf65c6ea-003c-4bbd-9f43-837fac8f993e {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.386295] env[61923]: DEBUG oslo_vmware.rw_handles [None req-81989905-2c22-4589-a2f0-03b8d81c0d2f tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/522995c9-80ac-b69a-62c7-1e4b40c5a0df/disk-0.vmdk is in state: ready. {{(pid=61923) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 918.386484] env[61923]: ERROR oslo_vmware.rw_handles [None req-81989905-2c22-4589-a2f0-03b8d81c0d2f tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Aborting lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/522995c9-80ac-b69a-62c7-1e4b40c5a0df/disk-0.vmdk due to incomplete transfer. [ 918.386731] env[61923]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-cb6aa28e-a1cb-4845-a665-2f50179f6357 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.396621] env[61923]: DEBUG oslo_vmware.rw_handles [None req-81989905-2c22-4589-a2f0-03b8d81c0d2f tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Closed VMDK read handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/522995c9-80ac-b69a-62c7-1e4b40c5a0df/disk-0.vmdk. {{(pid=61923) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 918.396869] env[61923]: DEBUG nova.virt.vmwareapi.images [None req-81989905-2c22-4589-a2f0-03b8d81c0d2f tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] [instance: 87ae37dc-d0d0-4b76-afdd-0ba3e8f6ce0b] Uploaded image ba0d8caf-3d48-4681-b6d3-c851741d738b to the Glance image server {{(pid=61923) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 918.399993] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-81989905-2c22-4589-a2f0-03b8d81c0d2f tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] [instance: 87ae37dc-d0d0-4b76-afdd-0ba3e8f6ce0b] Destroying the VM {{(pid=61923) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 918.400455] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-f16ae835-918d-4e87-a082-839399b0bc46 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.409471] env[61923]: DEBUG oslo_vmware.api [None req-81989905-2c22-4589-a2f0-03b8d81c0d2f tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Waiting for the task: (returnval){ [ 918.409471] env[61923]: value = "task-1377796" [ 918.409471] env[61923]: _type = "Task" [ 918.409471] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 918.421207] env[61923]: DEBUG oslo_vmware.api [None req-81989905-2c22-4589-a2f0-03b8d81c0d2f tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Task: {'id': task-1377796, 'name': Destroy_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.466695] env[61923]: DEBUG oslo_concurrency.lockutils [None req-f37b2e0d-8f0c-42f5-bf08-4bc9d42bb688 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Releasing lock "refresh_cache-6012d3b5-739c-4762-9bb4-09c51171dcd7" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 918.471288] env[61923]: DEBUG nova.scheduler.client.report [None req-85598005-1527-489f-b9ce-6a2b3821cde4 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Updated inventory for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 with generation 99 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 177, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 918.471288] env[61923]: DEBUG nova.compute.provider_tree [None req-85598005-1527-489f-b9ce-6a2b3821cde4 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Updating resource provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 generation from 99 to 100 during operation: update_inventory {{(pid=61923) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 918.471288] env[61923]: DEBUG nova.compute.provider_tree [None req-85598005-1527-489f-b9ce-6a2b3821cde4 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Updating inventory in ProviderTree for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 177, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 918.479828] env[61923]: DEBUG oslo_concurrency.lockutils [None req-f37b2e0d-8f0c-42f5-bf08-4bc9d42bb688 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Lock "interface-6012d3b5-739c-4762-9bb4-09c51171dcd7-b395b183-1c30-42de-929f-690ea10d9d75" "released" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: held 9.426s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 918.536778] env[61923]: DEBUG oslo_vmware.api [None req-9a89d0be-1b30-454c-a15f-f7048c8714c5 tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Task: {'id': task-1377792, 'name': ReconfigVM_Task, 'duration_secs': 0.355999} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 918.537336] env[61923]: DEBUG nova.virt.vmwareapi.volumeops [None req-9a89d0be-1b30-454c-a15f-f7048c8714c5 tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] [instance: 62c75dfd-6046-4de1-b1e3-13a307af1394] Reconfigured VM instance instance-00000057 to attach disk [datastore1] 62c75dfd-6046-4de1-b1e3-13a307af1394/62c75dfd-6046-4de1-b1e3-13a307af1394.vmdk or device None with type sparse {{(pid=61923) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 918.538123] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-3c4c49f8-5cd2-490c-b806-17792c3a286a {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.546444] env[61923]: DEBUG oslo_vmware.api [None req-9a89d0be-1b30-454c-a15f-f7048c8714c5 tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Waiting for the task: (returnval){ [ 918.546444] env[61923]: value = "task-1377797" [ 918.546444] env[61923]: _type = "Task" [ 918.546444] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 918.554908] env[61923]: DEBUG oslo_vmware.api [None req-9a89d0be-1b30-454c-a15f-f7048c8714c5 tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Task: {'id': task-1377797, 'name': Rename_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.664166] env[61923]: DEBUG oslo_vmware.api [None req-518b7c19-1088-43f5-9762-d7516e0ed822 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Task: {'id': task-1377794, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.192939} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 918.664469] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-518b7c19-1088-43f5-9762-d7516e0ed822 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Deleted the datastore file {{(pid=61923) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 918.664689] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-518b7c19-1088-43f5-9762-d7516e0ed822 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] [instance: 6012d3b5-739c-4762-9bb4-09c51171dcd7] Deleted contents of the VM from datastore datastore2 {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 918.664892] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-518b7c19-1088-43f5-9762-d7516e0ed822 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] [instance: 6012d3b5-739c-4762-9bb4-09c51171dcd7] Instance destroyed {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 918.665112] env[61923]: INFO nova.compute.manager [None req-518b7c19-1088-43f5-9762-d7516e0ed822 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] [instance: 6012d3b5-739c-4762-9bb4-09c51171dcd7] Took 1.13 seconds to destroy the instance on the hypervisor. [ 918.665384] env[61923]: DEBUG oslo.service.loopingcall [None req-518b7c19-1088-43f5-9762-d7516e0ed822 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61923) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 918.665617] env[61923]: DEBUG nova.compute.manager [-] [instance: 6012d3b5-739c-4762-9bb4-09c51171dcd7] Deallocating network for instance {{(pid=61923) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 918.665715] env[61923]: DEBUG nova.network.neutron [-] [instance: 6012d3b5-739c-4762-9bb4-09c51171dcd7] deallocate_for_instance() {{(pid=61923) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 918.685126] env[61923]: DEBUG oslo_concurrency.lockutils [None req-2c23dec8-473d-46b7-9285-b6dfd62f1983 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Acquiring lock "624e9207-7f24-4f45-8f52-0fc7a60c1e84" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 918.685391] env[61923]: DEBUG oslo_concurrency.lockutils [None req-2c23dec8-473d-46b7-9285-b6dfd62f1983 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Lock "624e9207-7f24-4f45-8f52-0fc7a60c1e84" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 918.685659] env[61923]: DEBUG oslo_concurrency.lockutils [None req-2c23dec8-473d-46b7-9285-b6dfd62f1983 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Acquiring lock "624e9207-7f24-4f45-8f52-0fc7a60c1e84-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 918.685929] env[61923]: DEBUG oslo_concurrency.lockutils [None req-2c23dec8-473d-46b7-9285-b6dfd62f1983 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Lock "624e9207-7f24-4f45-8f52-0fc7a60c1e84-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 918.686165] env[61923]: DEBUG oslo_concurrency.lockutils [None req-2c23dec8-473d-46b7-9285-b6dfd62f1983 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Lock "624e9207-7f24-4f45-8f52-0fc7a60c1e84-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 918.688332] env[61923]: INFO nova.compute.manager [None req-2c23dec8-473d-46b7-9285-b6dfd62f1983 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] [instance: 624e9207-7f24-4f45-8f52-0fc7a60c1e84] Terminating instance [ 918.690298] env[61923]: DEBUG oslo_concurrency.lockutils [None req-2c23dec8-473d-46b7-9285-b6dfd62f1983 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Acquiring lock "refresh_cache-624e9207-7f24-4f45-8f52-0fc7a60c1e84" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 918.690464] env[61923]: DEBUG oslo_concurrency.lockutils [None req-2c23dec8-473d-46b7-9285-b6dfd62f1983 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Acquired lock "refresh_cache-624e9207-7f24-4f45-8f52-0fc7a60c1e84" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 918.690640] env[61923]: DEBUG nova.network.neutron [None req-2c23dec8-473d-46b7-9285-b6dfd62f1983 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] [instance: 624e9207-7f24-4f45-8f52-0fc7a60c1e84] Building network info cache for instance {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 918.920393] env[61923]: DEBUG oslo_vmware.api [None req-81989905-2c22-4589-a2f0-03b8d81c0d2f tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Task: {'id': task-1377796, 'name': Destroy_Task, 'duration_secs': 0.342393} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 918.920720] env[61923]: INFO nova.virt.vmwareapi.vm_util [None req-81989905-2c22-4589-a2f0-03b8d81c0d2f tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] [instance: 87ae37dc-d0d0-4b76-afdd-0ba3e8f6ce0b] Destroyed the VM [ 918.920971] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-81989905-2c22-4589-a2f0-03b8d81c0d2f tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] [instance: 87ae37dc-d0d0-4b76-afdd-0ba3e8f6ce0b] Deleting Snapshot of the VM instance {{(pid=61923) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 918.921250] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-1559c1ae-d9a8-4772-a05f-a728d746a782 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.928356] env[61923]: DEBUG oslo_vmware.api [None req-81989905-2c22-4589-a2f0-03b8d81c0d2f tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Waiting for the task: (returnval){ [ 918.928356] env[61923]: value = "task-1377798" [ 918.928356] env[61923]: _type = "Task" [ 918.928356] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 918.936720] env[61923]: DEBUG oslo_vmware.api [None req-81989905-2c22-4589-a2f0-03b8d81c0d2f tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Task: {'id': task-1377798, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.984089] env[61923]: DEBUG oslo_concurrency.lockutils [None req-85598005-1527-489f-b9ce-6a2b3821cde4 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.851s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 918.986051] env[61923]: DEBUG oslo_concurrency.lockutils [None req-e087e202-bbb7-4067-8df4-dbcbdaf570ee tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 0.747s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 918.986249] env[61923]: DEBUG nova.objects.instance [None req-e087e202-bbb7-4067-8df4-dbcbdaf570ee tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] [instance: 624e9207-7f24-4f45-8f52-0fc7a60c1e84] Trying to apply a migration context that does not seem to be set for this instance {{(pid=61923) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 919.009252] env[61923]: INFO nova.scheduler.client.report [None req-85598005-1527-489f-b9ce-6a2b3821cde4 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Deleted allocations for instance 03c1e8ea-a322-46fa-9b07-e283080a7871 [ 919.059083] env[61923]: DEBUG oslo_vmware.api [None req-9a89d0be-1b30-454c-a15f-f7048c8714c5 tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Task: {'id': task-1377797, 'name': Rename_Task, 'duration_secs': 0.289765} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 919.059569] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-9a89d0be-1b30-454c-a15f-f7048c8714c5 tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] [instance: 62c75dfd-6046-4de1-b1e3-13a307af1394] Powering on the VM {{(pid=61923) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 919.059822] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-501b4206-f0dc-41ac-add8-927d5aef8f2f {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.069056] env[61923]: DEBUG oslo_vmware.api [None req-9a89d0be-1b30-454c-a15f-f7048c8714c5 tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Waiting for the task: (returnval){ [ 919.069056] env[61923]: value = "task-1377799" [ 919.069056] env[61923]: _type = "Task" [ 919.069056] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 919.080919] env[61923]: DEBUG oslo_vmware.api [None req-9a89d0be-1b30-454c-a15f-f7048c8714c5 tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Task: {'id': task-1377799, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.221136] env[61923]: DEBUG nova.network.neutron [None req-2c23dec8-473d-46b7-9285-b6dfd62f1983 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] [instance: 624e9207-7f24-4f45-8f52-0fc7a60c1e84] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 919.299373] env[61923]: DEBUG nova.network.neutron [None req-2c23dec8-473d-46b7-9285-b6dfd62f1983 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] [instance: 624e9207-7f24-4f45-8f52-0fc7a60c1e84] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 919.439950] env[61923]: DEBUG oslo_vmware.api [None req-81989905-2c22-4589-a2f0-03b8d81c0d2f tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Task: {'id': task-1377798, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.525525] env[61923]: DEBUG oslo_concurrency.lockutils [None req-85598005-1527-489f-b9ce-6a2b3821cde4 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Lock "03c1e8ea-a322-46fa-9b07-e283080a7871" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.818s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 919.580683] env[61923]: DEBUG oslo_vmware.api [None req-9a89d0be-1b30-454c-a15f-f7048c8714c5 tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Task: {'id': task-1377799, 'name': PowerOnVM_Task, 'duration_secs': 0.503705} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 919.580683] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-9a89d0be-1b30-454c-a15f-f7048c8714c5 tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] [instance: 62c75dfd-6046-4de1-b1e3-13a307af1394] Powered on the VM {{(pid=61923) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 919.580683] env[61923]: INFO nova.compute.manager [None req-9a89d0be-1b30-454c-a15f-f7048c8714c5 tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] [instance: 62c75dfd-6046-4de1-b1e3-13a307af1394] Took 7.16 seconds to spawn the instance on the hypervisor. [ 919.580683] env[61923]: DEBUG nova.compute.manager [None req-9a89d0be-1b30-454c-a15f-f7048c8714c5 tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] [instance: 62c75dfd-6046-4de1-b1e3-13a307af1394] Checking state {{(pid=61923) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 919.581224] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e40c7668-ad30-43e2-87f2-01e31314abb1 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.802014] env[61923]: DEBUG oslo_concurrency.lockutils [None req-2c23dec8-473d-46b7-9285-b6dfd62f1983 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Releasing lock "refresh_cache-624e9207-7f24-4f45-8f52-0fc7a60c1e84" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 919.802479] env[61923]: DEBUG nova.compute.manager [None req-2c23dec8-473d-46b7-9285-b6dfd62f1983 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] [instance: 624e9207-7f24-4f45-8f52-0fc7a60c1e84] Start destroying the instance on the hypervisor. {{(pid=61923) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 919.802677] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-2c23dec8-473d-46b7-9285-b6dfd62f1983 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] [instance: 624e9207-7f24-4f45-8f52-0fc7a60c1e84] Destroying instance {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 919.803703] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0ae158c-3262-425f-b04a-a34f6ff31ac7 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.815417] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-2c23dec8-473d-46b7-9285-b6dfd62f1983 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] [instance: 624e9207-7f24-4f45-8f52-0fc7a60c1e84] Powering off the VM {{(pid=61923) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 919.815885] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c8e7b31f-c94f-43d6-8022-86e78fdd47e1 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.831613] env[61923]: DEBUG oslo_vmware.api [None req-2c23dec8-473d-46b7-9285-b6dfd62f1983 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Waiting for the task: (returnval){ [ 919.831613] env[61923]: value = "task-1377800" [ 919.831613] env[61923]: _type = "Task" [ 919.831613] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 919.844351] env[61923]: DEBUG oslo_vmware.api [None req-2c23dec8-473d-46b7-9285-b6dfd62f1983 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Task: {'id': task-1377800, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.869156] env[61923]: DEBUG nova.compute.manager [req-6045fecf-1b7b-4ce8-b065-ff256a0092c0 req-eae84ff1-8a77-41ba-b81c-b0e4f750aff4 service nova] [instance: 6012d3b5-739c-4762-9bb4-09c51171dcd7] Received event network-vif-deleted-269a41b0-75e7-46af-9497-8eae3102df11 {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 919.869423] env[61923]: INFO nova.compute.manager [req-6045fecf-1b7b-4ce8-b065-ff256a0092c0 req-eae84ff1-8a77-41ba-b81c-b0e4f750aff4 service nova] [instance: 6012d3b5-739c-4762-9bb4-09c51171dcd7] Neutron deleted interface 269a41b0-75e7-46af-9497-8eae3102df11; detaching it from the instance and deleting it from the info cache [ 919.869625] env[61923]: DEBUG nova.network.neutron [req-6045fecf-1b7b-4ce8-b065-ff256a0092c0 req-eae84ff1-8a77-41ba-b81c-b0e4f750aff4 service nova] [instance: 6012d3b5-739c-4762-9bb4-09c51171dcd7] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 919.939974] env[61923]: DEBUG oslo_vmware.api [None req-81989905-2c22-4589-a2f0-03b8d81c0d2f tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Task: {'id': task-1377798, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.998100] env[61923]: DEBUG oslo_concurrency.lockutils [None req-e087e202-bbb7-4067-8df4-dbcbdaf570ee tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.010s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 920.103016] env[61923]: INFO nova.compute.manager [None req-9a89d0be-1b30-454c-a15f-f7048c8714c5 tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] [instance: 62c75dfd-6046-4de1-b1e3-13a307af1394] Took 13.66 seconds to build instance. [ 920.333352] env[61923]: DEBUG nova.network.neutron [-] [instance: 6012d3b5-739c-4762-9bb4-09c51171dcd7] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 920.343057] env[61923]: DEBUG nova.virt.vmwareapi.volumeops [None req-40624371-b6ea-4bd8-a64e-58d8640198f8 tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] [instance: b779d183-89ae-4e4d-ae99-e514e145ed43] Volume attach. Driver type: vmdk {{(pid=61923) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 920.343377] env[61923]: DEBUG nova.virt.vmwareapi.volumeops [None req-40624371-b6ea-4bd8-a64e-58d8640198f8 tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] [instance: b779d183-89ae-4e4d-ae99-e514e145ed43] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-292719', 'volume_id': '84fb906f-352f-4eae-b044-c86e8dda3254', 'name': 'volume-84fb906f-352f-4eae-b044-c86e8dda3254', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'b779d183-89ae-4e4d-ae99-e514e145ed43', 'attached_at': '', 'detached_at': '', 'volume_id': '84fb906f-352f-4eae-b044-c86e8dda3254', 'serial': '84fb906f-352f-4eae-b044-c86e8dda3254'} {{(pid=61923) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 920.344494] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8cee1a0-8fca-4096-931c-01bb60ad0b71 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.355352] env[61923]: DEBUG oslo_vmware.api [None req-2c23dec8-473d-46b7-9285-b6dfd62f1983 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Task: {'id': task-1377800, 'name': PowerOffVM_Task, 'duration_secs': 0.244673} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 920.368633] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-2c23dec8-473d-46b7-9285-b6dfd62f1983 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] [instance: 624e9207-7f24-4f45-8f52-0fc7a60c1e84] Powered off the VM {{(pid=61923) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 920.368633] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-2c23dec8-473d-46b7-9285-b6dfd62f1983 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] [instance: 624e9207-7f24-4f45-8f52-0fc7a60c1e84] Unregistering the VM {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 920.371524] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-59ea60de-adf5-4a8c-b9c0-e4218f237916 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.373648] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33f3ac43-9a61-4894-a634-347a69e5e4fe {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.379403] env[61923]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-9db80af2-19ae-46d1-aa5c-d0939092101e {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.411428] env[61923]: DEBUG nova.virt.vmwareapi.volumeops [None req-40624371-b6ea-4bd8-a64e-58d8640198f8 tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] [instance: b779d183-89ae-4e4d-ae99-e514e145ed43] Reconfiguring VM instance instance-0000004d to attach disk [datastore1] volume-84fb906f-352f-4eae-b044-c86e8dda3254/volume-84fb906f-352f-4eae-b044-c86e8dda3254.vmdk or device None with type thin {{(pid=61923) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 920.413829] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7d6dcf68-2a21-4e0f-807e-645b00b985c4 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.430614] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-2c23dec8-473d-46b7-9285-b6dfd62f1983 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] [instance: 624e9207-7f24-4f45-8f52-0fc7a60c1e84] Unregistered the VM {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 920.431069] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-2c23dec8-473d-46b7-9285-b6dfd62f1983 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] [instance: 624e9207-7f24-4f45-8f52-0fc7a60c1e84] Deleting contents of the VM from datastore datastore1 {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 920.431069] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-2c23dec8-473d-46b7-9285-b6dfd62f1983 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Deleting the datastore file [datastore1] 624e9207-7f24-4f45-8f52-0fc7a60c1e84 {{(pid=61923) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 920.433050] env[61923]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ba31b590-eafb-4d45-bac0-ea861deee480 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.438681] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e06ad7d9-1f0f-408d-89c1-ebd1f7851377 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.461146] env[61923]: DEBUG oslo_vmware.api [None req-81989905-2c22-4589-a2f0-03b8d81c0d2f tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Task: {'id': task-1377798, 'name': RemoveSnapshot_Task, 'duration_secs': 1.36286} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 920.463711] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-81989905-2c22-4589-a2f0-03b8d81c0d2f tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] [instance: 87ae37dc-d0d0-4b76-afdd-0ba3e8f6ce0b] Deleted Snapshot of the VM instance {{(pid=61923) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 920.463989] env[61923]: INFO nova.compute.manager [None req-81989905-2c22-4589-a2f0-03b8d81c0d2f tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] [instance: 87ae37dc-d0d0-4b76-afdd-0ba3e8f6ce0b] Took 15.36 seconds to snapshot the instance on the hypervisor. [ 920.466535] env[61923]: DEBUG oslo_vmware.api [None req-2c23dec8-473d-46b7-9285-b6dfd62f1983 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Waiting for the task: (returnval){ [ 920.466535] env[61923]: value = "task-1377802" [ 920.466535] env[61923]: _type = "Task" [ 920.466535] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 920.468745] env[61923]: DEBUG oslo_vmware.api [None req-40624371-b6ea-4bd8-a64e-58d8640198f8 tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Waiting for the task: (returnval){ [ 920.468745] env[61923]: value = "task-1377803" [ 920.468745] env[61923]: _type = "Task" [ 920.468745] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 920.477730] env[61923]: DEBUG nova.compute.manager [req-6045fecf-1b7b-4ce8-b065-ff256a0092c0 req-eae84ff1-8a77-41ba-b81c-b0e4f750aff4 service nova] [instance: 6012d3b5-739c-4762-9bb4-09c51171dcd7] Detach interface failed, port_id=269a41b0-75e7-46af-9497-8eae3102df11, reason: Instance 6012d3b5-739c-4762-9bb4-09c51171dcd7 could not be found. {{(pid=61923) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 920.494363] env[61923]: DEBUG oslo_vmware.api [None req-40624371-b6ea-4bd8-a64e-58d8640198f8 tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Task: {'id': task-1377803, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.494363] env[61923]: DEBUG oslo_vmware.api [None req-2c23dec8-473d-46b7-9285-b6dfd62f1983 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Task: {'id': task-1377802, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.566715] env[61923]: DEBUG oslo_concurrency.lockutils [None req-91fc90ee-9f70-48af-84f3-b299c046606b tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Acquiring lock "4c2c973a-5d23-4b8a-8b99-6b634e337054" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 920.567066] env[61923]: DEBUG oslo_concurrency.lockutils [None req-91fc90ee-9f70-48af-84f3-b299c046606b tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Lock "4c2c973a-5d23-4b8a-8b99-6b634e337054" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 920.567337] env[61923]: DEBUG oslo_concurrency.lockutils [None req-91fc90ee-9f70-48af-84f3-b299c046606b tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Acquiring lock "4c2c973a-5d23-4b8a-8b99-6b634e337054-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 920.567587] env[61923]: DEBUG oslo_concurrency.lockutils [None req-91fc90ee-9f70-48af-84f3-b299c046606b tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Lock "4c2c973a-5d23-4b8a-8b99-6b634e337054-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 920.567786] env[61923]: DEBUG oslo_concurrency.lockutils [None req-91fc90ee-9f70-48af-84f3-b299c046606b tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Lock "4c2c973a-5d23-4b8a-8b99-6b634e337054-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 920.570313] env[61923]: INFO nova.compute.manager [None req-91fc90ee-9f70-48af-84f3-b299c046606b tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: 4c2c973a-5d23-4b8a-8b99-6b634e337054] Terminating instance [ 920.574666] env[61923]: DEBUG nova.compute.manager [None req-91fc90ee-9f70-48af-84f3-b299c046606b tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: 4c2c973a-5d23-4b8a-8b99-6b634e337054] Start destroying the instance on the hypervisor. {{(pid=61923) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 920.574871] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-91fc90ee-9f70-48af-84f3-b299c046606b tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: 4c2c973a-5d23-4b8a-8b99-6b634e337054] Destroying instance {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 920.575826] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95e05b41-02af-40c0-abf2-e5dcbbd28783 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.585448] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-91fc90ee-9f70-48af-84f3-b299c046606b tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: 4c2c973a-5d23-4b8a-8b99-6b634e337054] Powering off the VM {{(pid=61923) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 920.585729] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1848d845-901c-415b-af5e-599704d47d65 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.593522] env[61923]: DEBUG oslo_vmware.api [None req-91fc90ee-9f70-48af-84f3-b299c046606b tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Waiting for the task: (returnval){ [ 920.593522] env[61923]: value = "task-1377804" [ 920.593522] env[61923]: _type = "Task" [ 920.593522] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 920.603263] env[61923]: DEBUG oslo_vmware.api [None req-91fc90ee-9f70-48af-84f3-b299c046606b tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Task: {'id': task-1377804, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.604848] env[61923]: DEBUG oslo_concurrency.lockutils [None req-9a89d0be-1b30-454c-a15f-f7048c8714c5 tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Lock "62c75dfd-6046-4de1-b1e3-13a307af1394" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.176s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 920.845224] env[61923]: INFO nova.compute.manager [-] [instance: 6012d3b5-739c-4762-9bb4-09c51171dcd7] Took 2.18 seconds to deallocate network for instance. [ 920.994306] env[61923]: DEBUG oslo_vmware.api [None req-40624371-b6ea-4bd8-a64e-58d8640198f8 tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Task: {'id': task-1377803, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.994629] env[61923]: DEBUG oslo_vmware.api [None req-2c23dec8-473d-46b7-9285-b6dfd62f1983 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Task: {'id': task-1377802, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.201482} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 920.994893] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-2c23dec8-473d-46b7-9285-b6dfd62f1983 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Deleted the datastore file {{(pid=61923) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 920.995134] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-2c23dec8-473d-46b7-9285-b6dfd62f1983 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] [instance: 624e9207-7f24-4f45-8f52-0fc7a60c1e84] Deleted contents of the VM from datastore datastore1 {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 920.995363] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-2c23dec8-473d-46b7-9285-b6dfd62f1983 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] [instance: 624e9207-7f24-4f45-8f52-0fc7a60c1e84] Instance destroyed {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 920.995579] env[61923]: INFO nova.compute.manager [None req-2c23dec8-473d-46b7-9285-b6dfd62f1983 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] [instance: 624e9207-7f24-4f45-8f52-0fc7a60c1e84] Took 1.19 seconds to destroy the instance on the hypervisor. [ 920.995886] env[61923]: DEBUG oslo.service.loopingcall [None req-2c23dec8-473d-46b7-9285-b6dfd62f1983 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61923) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 920.996145] env[61923]: DEBUG nova.compute.manager [-] [instance: 624e9207-7f24-4f45-8f52-0fc7a60c1e84] Deallocating network for instance {{(pid=61923) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 920.996432] env[61923]: DEBUG nova.network.neutron [-] [instance: 624e9207-7f24-4f45-8f52-0fc7a60c1e84] deallocate_for_instance() {{(pid=61923) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 921.012218] env[61923]: DEBUG nova.network.neutron [-] [instance: 624e9207-7f24-4f45-8f52-0fc7a60c1e84] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 921.106553] env[61923]: DEBUG oslo_vmware.api [None req-91fc90ee-9f70-48af-84f3-b299c046606b tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Task: {'id': task-1377804, 'name': PowerOffVM_Task, 'duration_secs': 0.220811} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 921.106846] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-91fc90ee-9f70-48af-84f3-b299c046606b tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: 4c2c973a-5d23-4b8a-8b99-6b634e337054] Powered off the VM {{(pid=61923) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 921.107084] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-91fc90ee-9f70-48af-84f3-b299c046606b tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: 4c2c973a-5d23-4b8a-8b99-6b634e337054] Unregistering the VM {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 921.107421] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-eb8734c3-49dc-45c7-9533-1ae2adc1172f {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.177689] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-91fc90ee-9f70-48af-84f3-b299c046606b tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: 4c2c973a-5d23-4b8a-8b99-6b634e337054] Unregistered the VM {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 921.177942] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-91fc90ee-9f70-48af-84f3-b299c046606b tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: 4c2c973a-5d23-4b8a-8b99-6b634e337054] Deleting contents of the VM from datastore datastore1 {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 921.178260] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-91fc90ee-9f70-48af-84f3-b299c046606b tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Deleting the datastore file [datastore1] 4c2c973a-5d23-4b8a-8b99-6b634e337054 {{(pid=61923) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 921.179039] env[61923]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b3d67a79-733e-43d4-ba7a-8db22acf8d73 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.187922] env[61923]: DEBUG oslo_vmware.api [None req-91fc90ee-9f70-48af-84f3-b299c046606b tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Waiting for the task: (returnval){ [ 921.187922] env[61923]: value = "task-1377806" [ 921.187922] env[61923]: _type = "Task" [ 921.187922] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 921.197871] env[61923]: DEBUG oslo_vmware.api [None req-91fc90ee-9f70-48af-84f3-b299c046606b tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Task: {'id': task-1377806, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.352517] env[61923]: DEBUG oslo_concurrency.lockutils [None req-518b7c19-1088-43f5-9762-d7516e0ed822 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 921.352517] env[61923]: DEBUG oslo_concurrency.lockutils [None req-518b7c19-1088-43f5-9762-d7516e0ed822 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 921.352660] env[61923]: DEBUG nova.objects.instance [None req-518b7c19-1088-43f5-9762-d7516e0ed822 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Lazy-loading 'resources' on Instance uuid 6012d3b5-739c-4762-9bb4-09c51171dcd7 {{(pid=61923) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 921.487086] env[61923]: DEBUG oslo_vmware.api [None req-40624371-b6ea-4bd8-a64e-58d8640198f8 tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Task: {'id': task-1377803, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.514484] env[61923]: DEBUG nova.network.neutron [-] [instance: 624e9207-7f24-4f45-8f52-0fc7a60c1e84] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 921.698571] env[61923]: DEBUG oslo_vmware.api [None req-91fc90ee-9f70-48af-84f3-b299c046606b tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Task: {'id': task-1377806, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.991254] env[61923]: DEBUG oslo_vmware.api [None req-40624371-b6ea-4bd8-a64e-58d8640198f8 tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Task: {'id': task-1377803, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.995313] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-440caed6-fd56-4b85-b954-e66dcdebc537 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.002826] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea1372ff-7ec1-4482-9ccb-979d3ce074db {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.035030] env[61923]: INFO nova.compute.manager [-] [instance: 624e9207-7f24-4f45-8f52-0fc7a60c1e84] Took 1.04 seconds to deallocate network for instance. [ 922.037283] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c636fb08-ed65-4e6a-8ef6-0352b29d6c8a {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.049705] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-452fd4ec-c6c5-4e22-9f8f-f1820d49aa1f {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.064697] env[61923]: DEBUG nova.compute.provider_tree [None req-518b7c19-1088-43f5-9762-d7516e0ed822 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 922.177104] env[61923]: DEBUG nova.compute.manager [None req-edec20df-41c5-4b44-b67e-1c51779999ca tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] [instance: 62c75dfd-6046-4de1-b1e3-13a307af1394] Checking state {{(pid=61923) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 922.178266] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b2c3e5c-60d6-4d5d-b751-d24b55ca3aef {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.197969] env[61923]: DEBUG oslo_vmware.api [None req-91fc90ee-9f70-48af-84f3-b299c046606b tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Task: {'id': task-1377806, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.490800] env[61923]: DEBUG oslo_vmware.api [None req-40624371-b6ea-4bd8-a64e-58d8640198f8 tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Task: {'id': task-1377803, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.544975] env[61923]: DEBUG oslo_concurrency.lockutils [None req-2c23dec8-473d-46b7-9285-b6dfd62f1983 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 922.568197] env[61923]: DEBUG nova.scheduler.client.report [None req-518b7c19-1088-43f5-9762-d7516e0ed822 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 177, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 922.690357] env[61923]: INFO nova.compute.manager [None req-edec20df-41c5-4b44-b67e-1c51779999ca tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] [instance: 62c75dfd-6046-4de1-b1e3-13a307af1394] instance snapshotting [ 922.695981] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb242087-a7f9-47d7-ba52-8e214a9ef92a {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.704752] env[61923]: DEBUG oslo_vmware.api [None req-91fc90ee-9f70-48af-84f3-b299c046606b tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Task: {'id': task-1377806, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.717879] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa0edd32-aba3-4fad-816b-f054eb97daa9 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.989564] env[61923]: DEBUG oslo_vmware.api [None req-40624371-b6ea-4bd8-a64e-58d8640198f8 tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Task: {'id': task-1377803, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.073382] env[61923]: DEBUG oslo_concurrency.lockutils [None req-518b7c19-1088-43f5-9762-d7516e0ed822 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.721s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 923.075589] env[61923]: DEBUG oslo_concurrency.lockutils [None req-2c23dec8-473d-46b7-9285-b6dfd62f1983 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.531s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 923.075877] env[61923]: DEBUG nova.objects.instance [None req-2c23dec8-473d-46b7-9285-b6dfd62f1983 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Lazy-loading 'resources' on Instance uuid 624e9207-7f24-4f45-8f52-0fc7a60c1e84 {{(pid=61923) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 923.096884] env[61923]: INFO nova.scheduler.client.report [None req-518b7c19-1088-43f5-9762-d7516e0ed822 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Deleted allocations for instance 6012d3b5-739c-4762-9bb4-09c51171dcd7 [ 923.202779] env[61923]: DEBUG oslo_vmware.api [None req-91fc90ee-9f70-48af-84f3-b299c046606b tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Task: {'id': task-1377806, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.229754] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-edec20df-41c5-4b44-b67e-1c51779999ca tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] [instance: 62c75dfd-6046-4de1-b1e3-13a307af1394] Creating Snapshot of the VM instance {{(pid=61923) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 923.230128] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-55ae5ab7-2940-426f-ae54-52a85992d330 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.238310] env[61923]: DEBUG oslo_vmware.api [None req-edec20df-41c5-4b44-b67e-1c51779999ca tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Waiting for the task: (returnval){ [ 923.238310] env[61923]: value = "task-1377807" [ 923.238310] env[61923]: _type = "Task" [ 923.238310] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 923.246820] env[61923]: DEBUG oslo_vmware.api [None req-edec20df-41c5-4b44-b67e-1c51779999ca tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Task: {'id': task-1377807, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.490104] env[61923]: DEBUG oslo_vmware.api [None req-40624371-b6ea-4bd8-a64e-58d8640198f8 tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Task: {'id': task-1377803, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.567900] env[61923]: DEBUG oslo_concurrency.lockutils [None req-3856851e-8c34-4281-b1ef-81601f186c03 tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Acquiring lock "2719569b-8572-4199-8158-7bb367d17dc5" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 923.568211] env[61923]: DEBUG oslo_concurrency.lockutils [None req-3856851e-8c34-4281-b1ef-81601f186c03 tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Lock "2719569b-8572-4199-8158-7bb367d17dc5" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 923.568432] env[61923]: DEBUG oslo_concurrency.lockutils [None req-3856851e-8c34-4281-b1ef-81601f186c03 tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Acquiring lock "2719569b-8572-4199-8158-7bb367d17dc5-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 923.568616] env[61923]: DEBUG oslo_concurrency.lockutils [None req-3856851e-8c34-4281-b1ef-81601f186c03 tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Lock "2719569b-8572-4199-8158-7bb367d17dc5-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 923.568789] env[61923]: DEBUG oslo_concurrency.lockutils [None req-3856851e-8c34-4281-b1ef-81601f186c03 tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Lock "2719569b-8572-4199-8158-7bb367d17dc5-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 923.571145] env[61923]: INFO nova.compute.manager [None req-3856851e-8c34-4281-b1ef-81601f186c03 tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] [instance: 2719569b-8572-4199-8158-7bb367d17dc5] Terminating instance [ 923.573115] env[61923]: DEBUG nova.compute.manager [None req-3856851e-8c34-4281-b1ef-81601f186c03 tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] [instance: 2719569b-8572-4199-8158-7bb367d17dc5] Start destroying the instance on the hypervisor. {{(pid=61923) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 923.573325] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-3856851e-8c34-4281-b1ef-81601f186c03 tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] [instance: 2719569b-8572-4199-8158-7bb367d17dc5] Destroying instance {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 923.574163] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46352754-c621-4837-bf6a-a70a2d24eeff {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.590366] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-3856851e-8c34-4281-b1ef-81601f186c03 tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] [instance: 2719569b-8572-4199-8158-7bb367d17dc5] Powering off the VM {{(pid=61923) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 923.590648] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c8325e2d-afdc-4cdf-8cd4-3974d0ef22f5 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.606842] env[61923]: DEBUG oslo_vmware.api [None req-3856851e-8c34-4281-b1ef-81601f186c03 tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Waiting for the task: (returnval){ [ 923.606842] env[61923]: value = "task-1377808" [ 923.606842] env[61923]: _type = "Task" [ 923.606842] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 923.607433] env[61923]: DEBUG oslo_concurrency.lockutils [None req-518b7c19-1088-43f5-9762-d7516e0ed822 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Lock "6012d3b5-739c-4762-9bb4-09c51171dcd7" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.072s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 923.618925] env[61923]: DEBUG oslo_vmware.api [None req-3856851e-8c34-4281-b1ef-81601f186c03 tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Task: {'id': task-1377808, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.706673] env[61923]: DEBUG oslo_vmware.api [None req-91fc90ee-9f70-48af-84f3-b299c046606b tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Task: {'id': task-1377806, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.734674] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2ae4ff6-6042-4c86-ae87-faf6fed6fab2 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.745186] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb6cfbb1-6540-4cf3-9b23-3437d585eea4 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.752446] env[61923]: DEBUG oslo_vmware.api [None req-edec20df-41c5-4b44-b67e-1c51779999ca tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Task: {'id': task-1377807, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.782620] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87a3d4d0-c675-47e8-ab4c-b770a721ad15 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.792059] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c3c67ba-5591-4cc8-9ccd-5204855de00d {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.807739] env[61923]: DEBUG nova.compute.provider_tree [None req-2c23dec8-473d-46b7-9285-b6dfd62f1983 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 923.990859] env[61923]: DEBUG oslo_vmware.api [None req-40624371-b6ea-4bd8-a64e-58d8640198f8 tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Task: {'id': task-1377803, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.120051] env[61923]: DEBUG oslo_vmware.api [None req-3856851e-8c34-4281-b1ef-81601f186c03 tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Task: {'id': task-1377808, 'name': PowerOffVM_Task, 'duration_secs': 0.233395} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 924.120051] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-3856851e-8c34-4281-b1ef-81601f186c03 tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] [instance: 2719569b-8572-4199-8158-7bb367d17dc5] Powered off the VM {{(pid=61923) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 924.120051] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-3856851e-8c34-4281-b1ef-81601f186c03 tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] [instance: 2719569b-8572-4199-8158-7bb367d17dc5] Unregistering the VM {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 924.120051] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-8fb02089-5fe9-47cc-8ccd-1836b7181ad2 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.205086] env[61923]: DEBUG oslo_vmware.api [None req-91fc90ee-9f70-48af-84f3-b299c046606b tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Task: {'id': task-1377806, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.248885] env[61923]: DEBUG oslo_vmware.api [None req-edec20df-41c5-4b44-b67e-1c51779999ca tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Task: {'id': task-1377807, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.312268] env[61923]: DEBUG nova.scheduler.client.report [None req-2c23dec8-473d-46b7-9285-b6dfd62f1983 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 177, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 924.491919] env[61923]: DEBUG oslo_vmware.api [None req-40624371-b6ea-4bd8-a64e-58d8640198f8 tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Task: {'id': task-1377803, 'name': ReconfigVM_Task, 'duration_secs': 3.971794} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 924.491919] env[61923]: DEBUG nova.virt.vmwareapi.volumeops [None req-40624371-b6ea-4bd8-a64e-58d8640198f8 tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] [instance: b779d183-89ae-4e4d-ae99-e514e145ed43] Reconfigured VM instance instance-0000004d to attach disk [datastore1] volume-84fb906f-352f-4eae-b044-c86e8dda3254/volume-84fb906f-352f-4eae-b044-c86e8dda3254.vmdk or device None with type thin {{(pid=61923) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 924.496309] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0e5312bd-df50-4a66-9e57-0b1ef42660e0 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.512614] env[61923]: DEBUG oslo_vmware.api [None req-40624371-b6ea-4bd8-a64e-58d8640198f8 tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Waiting for the task: (returnval){ [ 924.512614] env[61923]: value = "task-1377810" [ 924.512614] env[61923]: _type = "Task" [ 924.512614] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 924.526539] env[61923]: DEBUG oslo_vmware.api [None req-40624371-b6ea-4bd8-a64e-58d8640198f8 tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Task: {'id': task-1377810, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.652290] env[61923]: DEBUG oslo_concurrency.lockutils [None req-b04cd48e-3d8d-4840-b909-cd681fdc6a00 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Acquiring lock "422d06d6-6932-46c3-bb25-841e6f65c028" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 924.652431] env[61923]: DEBUG oslo_concurrency.lockutils [None req-b04cd48e-3d8d-4840-b909-cd681fdc6a00 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Lock "422d06d6-6932-46c3-bb25-841e6f65c028" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 924.652712] env[61923]: DEBUG oslo_concurrency.lockutils [None req-b04cd48e-3d8d-4840-b909-cd681fdc6a00 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Acquiring lock "422d06d6-6932-46c3-bb25-841e6f65c028-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 924.652920] env[61923]: DEBUG oslo_concurrency.lockutils [None req-b04cd48e-3d8d-4840-b909-cd681fdc6a00 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Lock "422d06d6-6932-46c3-bb25-841e6f65c028-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 924.653123] env[61923]: DEBUG oslo_concurrency.lockutils [None req-b04cd48e-3d8d-4840-b909-cd681fdc6a00 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Lock "422d06d6-6932-46c3-bb25-841e6f65c028-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 924.656197] env[61923]: INFO nova.compute.manager [None req-b04cd48e-3d8d-4840-b909-cd681fdc6a00 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] [instance: 422d06d6-6932-46c3-bb25-841e6f65c028] Terminating instance [ 924.657597] env[61923]: DEBUG nova.compute.manager [None req-b04cd48e-3d8d-4840-b909-cd681fdc6a00 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] [instance: 422d06d6-6932-46c3-bb25-841e6f65c028] Start destroying the instance on the hypervisor. {{(pid=61923) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 924.657803] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-b04cd48e-3d8d-4840-b909-cd681fdc6a00 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] [instance: 422d06d6-6932-46c3-bb25-841e6f65c028] Destroying instance {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 924.658651] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2832d3dd-6154-43ca-8d0e-7a0f2793a96c {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.667032] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-b04cd48e-3d8d-4840-b909-cd681fdc6a00 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] [instance: 422d06d6-6932-46c3-bb25-841e6f65c028] Powering off the VM {{(pid=61923) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 924.667353] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c1627e5a-b8e4-4495-93c7-3ddf70c2ff42 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.675264] env[61923]: DEBUG oslo_vmware.api [None req-b04cd48e-3d8d-4840-b909-cd681fdc6a00 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Waiting for the task: (returnval){ [ 924.675264] env[61923]: value = "task-1377811" [ 924.675264] env[61923]: _type = "Task" [ 924.675264] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 924.684862] env[61923]: DEBUG oslo_vmware.api [None req-b04cd48e-3d8d-4840-b909-cd681fdc6a00 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Task: {'id': task-1377811, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.708114] env[61923]: DEBUG oslo_vmware.api [None req-91fc90ee-9f70-48af-84f3-b299c046606b tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Task: {'id': task-1377806, 'name': DeleteDatastoreFile_Task, 'duration_secs': 3.162445} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 924.708499] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-91fc90ee-9f70-48af-84f3-b299c046606b tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Deleted the datastore file {{(pid=61923) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 924.708631] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-91fc90ee-9f70-48af-84f3-b299c046606b tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: 4c2c973a-5d23-4b8a-8b99-6b634e337054] Deleted contents of the VM from datastore datastore1 {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 924.708825] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-91fc90ee-9f70-48af-84f3-b299c046606b tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: 4c2c973a-5d23-4b8a-8b99-6b634e337054] Instance destroyed {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 924.709018] env[61923]: INFO nova.compute.manager [None req-91fc90ee-9f70-48af-84f3-b299c046606b tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: 4c2c973a-5d23-4b8a-8b99-6b634e337054] Took 4.13 seconds to destroy the instance on the hypervisor. [ 924.709276] env[61923]: DEBUG oslo.service.loopingcall [None req-91fc90ee-9f70-48af-84f3-b299c046606b tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61923) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 924.709524] env[61923]: DEBUG nova.compute.manager [-] [instance: 4c2c973a-5d23-4b8a-8b99-6b634e337054] Deallocating network for instance {{(pid=61923) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 924.709609] env[61923]: DEBUG nova.network.neutron [-] [instance: 4c2c973a-5d23-4b8a-8b99-6b634e337054] deallocate_for_instance() {{(pid=61923) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 924.750669] env[61923]: DEBUG oslo_vmware.api [None req-edec20df-41c5-4b44-b67e-1c51779999ca tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Task: {'id': task-1377807, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.816733] env[61923]: DEBUG oslo_concurrency.lockutils [None req-2c23dec8-473d-46b7-9285-b6dfd62f1983 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.741s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 924.844338] env[61923]: INFO nova.scheduler.client.report [None req-2c23dec8-473d-46b7-9285-b6dfd62f1983 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Deleted allocations for instance 624e9207-7f24-4f45-8f52-0fc7a60c1e84 [ 925.023459] env[61923]: DEBUG oslo_vmware.api [None req-40624371-b6ea-4bd8-a64e-58d8640198f8 tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Task: {'id': task-1377810, 'name': ReconfigVM_Task, 'duration_secs': 0.143652} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 925.023786] env[61923]: DEBUG nova.virt.vmwareapi.volumeops [None req-40624371-b6ea-4bd8-a64e-58d8640198f8 tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] [instance: b779d183-89ae-4e4d-ae99-e514e145ed43] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-292719', 'volume_id': '84fb906f-352f-4eae-b044-c86e8dda3254', 'name': 'volume-84fb906f-352f-4eae-b044-c86e8dda3254', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'b779d183-89ae-4e4d-ae99-e514e145ed43', 'attached_at': '', 'detached_at': '', 'volume_id': '84fb906f-352f-4eae-b044-c86e8dda3254', 'serial': '84fb906f-352f-4eae-b044-c86e8dda3254'} {{(pid=61923) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 925.128858] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-3856851e-8c34-4281-b1ef-81601f186c03 tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] [instance: 2719569b-8572-4199-8158-7bb367d17dc5] Unregistered the VM {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 925.129174] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-3856851e-8c34-4281-b1ef-81601f186c03 tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] [instance: 2719569b-8572-4199-8158-7bb367d17dc5] Deleting contents of the VM from datastore datastore1 {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 925.129422] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-3856851e-8c34-4281-b1ef-81601f186c03 tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Deleting the datastore file [datastore1] 2719569b-8572-4199-8158-7bb367d17dc5 {{(pid=61923) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 925.129711] env[61923]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ba7d8369-be5c-4f9a-a747-8eae716cc76e {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.134494] env[61923]: DEBUG nova.compute.manager [req-9cac4d2d-5df8-4fde-84c0-b72ec40d56a5 req-fa9ca3db-f1b9-4d41-a1f7-31a78610a222 service nova] [instance: 4c2c973a-5d23-4b8a-8b99-6b634e337054] Received event network-vif-deleted-9584dacf-1833-4a40-bfbe-56d7d5ea9998 {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 925.134711] env[61923]: INFO nova.compute.manager [req-9cac4d2d-5df8-4fde-84c0-b72ec40d56a5 req-fa9ca3db-f1b9-4d41-a1f7-31a78610a222 service nova] [instance: 4c2c973a-5d23-4b8a-8b99-6b634e337054] Neutron deleted interface 9584dacf-1833-4a40-bfbe-56d7d5ea9998; detaching it from the instance and deleting it from the info cache [ 925.134866] env[61923]: DEBUG nova.network.neutron [req-9cac4d2d-5df8-4fde-84c0-b72ec40d56a5 req-fa9ca3db-f1b9-4d41-a1f7-31a78610a222 service nova] [instance: 4c2c973a-5d23-4b8a-8b99-6b634e337054] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 925.140271] env[61923]: DEBUG oslo_vmware.api [None req-3856851e-8c34-4281-b1ef-81601f186c03 tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Waiting for the task: (returnval){ [ 925.140271] env[61923]: value = "task-1377812" [ 925.140271] env[61923]: _type = "Task" [ 925.140271] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 925.149085] env[61923]: DEBUG oslo_vmware.api [None req-3856851e-8c34-4281-b1ef-81601f186c03 tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Task: {'id': task-1377812, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.185806] env[61923]: DEBUG oslo_vmware.api [None req-b04cd48e-3d8d-4840-b909-cd681fdc6a00 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Task: {'id': task-1377811, 'name': PowerOffVM_Task, 'duration_secs': 0.207061} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 925.186103] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-b04cd48e-3d8d-4840-b909-cd681fdc6a00 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] [instance: 422d06d6-6932-46c3-bb25-841e6f65c028] Powered off the VM {{(pid=61923) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 925.186282] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-b04cd48e-3d8d-4840-b909-cd681fdc6a00 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] [instance: 422d06d6-6932-46c3-bb25-841e6f65c028] Unregistering the VM {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 925.186543] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-fedbca3a-7238-43c5-ae9d-686bfa7da468 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.248321] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-b04cd48e-3d8d-4840-b909-cd681fdc6a00 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] [instance: 422d06d6-6932-46c3-bb25-841e6f65c028] Unregistered the VM {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 925.248658] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-b04cd48e-3d8d-4840-b909-cd681fdc6a00 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] [instance: 422d06d6-6932-46c3-bb25-841e6f65c028] Deleting contents of the VM from datastore datastore2 {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 925.248742] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-b04cd48e-3d8d-4840-b909-cd681fdc6a00 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Deleting the datastore file [datastore2] 422d06d6-6932-46c3-bb25-841e6f65c028 {{(pid=61923) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 925.249028] env[61923]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e74e5c72-f133-4ce5-bcb3-fe960f686d36 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.254265] env[61923]: DEBUG oslo_vmware.api [None req-edec20df-41c5-4b44-b67e-1c51779999ca tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Task: {'id': task-1377807, 'name': CreateSnapshot_Task, 'duration_secs': 1.576737} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 925.254956] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-edec20df-41c5-4b44-b67e-1c51779999ca tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] [instance: 62c75dfd-6046-4de1-b1e3-13a307af1394] Created Snapshot of the VM instance {{(pid=61923) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 925.255692] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09ef2c97-ad7b-4ab1-8eec-42f5a3a59a8c {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.260217] env[61923]: DEBUG oslo_vmware.api [None req-b04cd48e-3d8d-4840-b909-cd681fdc6a00 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Waiting for the task: (returnval){ [ 925.260217] env[61923]: value = "task-1377814" [ 925.260217] env[61923]: _type = "Task" [ 925.260217] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 925.275052] env[61923]: DEBUG oslo_vmware.api [None req-b04cd48e-3d8d-4840-b909-cd681fdc6a00 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Task: {'id': task-1377814, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.354693] env[61923]: DEBUG oslo_concurrency.lockutils [None req-2c23dec8-473d-46b7-9285-b6dfd62f1983 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Lock "624e9207-7f24-4f45-8f52-0fc7a60c1e84" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.669s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 925.545766] env[61923]: DEBUG oslo_concurrency.lockutils [None req-f3aa73ee-e2c2-44a8-b10f-5d20f9062f30 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Acquiring lock "c98c96a5-d18e-4519-996a-949ded16225f" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 925.546154] env[61923]: DEBUG oslo_concurrency.lockutils [None req-f3aa73ee-e2c2-44a8-b10f-5d20f9062f30 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Lock "c98c96a5-d18e-4519-996a-949ded16225f" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 925.546408] env[61923]: DEBUG oslo_concurrency.lockutils [None req-f3aa73ee-e2c2-44a8-b10f-5d20f9062f30 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Acquiring lock "c98c96a5-d18e-4519-996a-949ded16225f-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 925.546634] env[61923]: DEBUG oslo_concurrency.lockutils [None req-f3aa73ee-e2c2-44a8-b10f-5d20f9062f30 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Lock "c98c96a5-d18e-4519-996a-949ded16225f-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 925.546843] env[61923]: DEBUG oslo_concurrency.lockutils [None req-f3aa73ee-e2c2-44a8-b10f-5d20f9062f30 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Lock "c98c96a5-d18e-4519-996a-949ded16225f-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 925.549589] env[61923]: INFO nova.compute.manager [None req-f3aa73ee-e2c2-44a8-b10f-5d20f9062f30 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] [instance: c98c96a5-d18e-4519-996a-949ded16225f] Terminating instance [ 925.551396] env[61923]: DEBUG oslo_concurrency.lockutils [None req-f3aa73ee-e2c2-44a8-b10f-5d20f9062f30 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Acquiring lock "refresh_cache-c98c96a5-d18e-4519-996a-949ded16225f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 925.551559] env[61923]: DEBUG oslo_concurrency.lockutils [None req-f3aa73ee-e2c2-44a8-b10f-5d20f9062f30 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Acquired lock "refresh_cache-c98c96a5-d18e-4519-996a-949ded16225f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 925.552442] env[61923]: DEBUG nova.network.neutron [None req-f3aa73ee-e2c2-44a8-b10f-5d20f9062f30 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] [instance: c98c96a5-d18e-4519-996a-949ded16225f] Building network info cache for instance {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 925.578243] env[61923]: DEBUG nova.network.neutron [-] [instance: 4c2c973a-5d23-4b8a-8b99-6b634e337054] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 925.637628] env[61923]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ea2d7a82-96f5-4919-a88a-a870dfef27f0 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.648515] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a9c9dae-db9f-415e-8881-804604689557 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.662090] env[61923]: DEBUG oslo_vmware.api [None req-3856851e-8c34-4281-b1ef-81601f186c03 tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Task: {'id': task-1377812, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.37146} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 925.662702] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-3856851e-8c34-4281-b1ef-81601f186c03 tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Deleted the datastore file {{(pid=61923) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 925.662933] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-3856851e-8c34-4281-b1ef-81601f186c03 tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] [instance: 2719569b-8572-4199-8158-7bb367d17dc5] Deleted contents of the VM from datastore datastore1 {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 925.663180] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-3856851e-8c34-4281-b1ef-81601f186c03 tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] [instance: 2719569b-8572-4199-8158-7bb367d17dc5] Instance destroyed {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 925.663380] env[61923]: INFO nova.compute.manager [None req-3856851e-8c34-4281-b1ef-81601f186c03 tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] [instance: 2719569b-8572-4199-8158-7bb367d17dc5] Took 2.09 seconds to destroy the instance on the hypervisor. [ 925.663683] env[61923]: DEBUG oslo.service.loopingcall [None req-3856851e-8c34-4281-b1ef-81601f186c03 tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61923) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 925.663908] env[61923]: DEBUG nova.compute.manager [-] [instance: 2719569b-8572-4199-8158-7bb367d17dc5] Deallocating network for instance {{(pid=61923) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 925.664051] env[61923]: DEBUG nova.network.neutron [-] [instance: 2719569b-8572-4199-8158-7bb367d17dc5] deallocate_for_instance() {{(pid=61923) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 925.680836] env[61923]: DEBUG nova.compute.manager [req-9cac4d2d-5df8-4fde-84c0-b72ec40d56a5 req-fa9ca3db-f1b9-4d41-a1f7-31a78610a222 service nova] [instance: 4c2c973a-5d23-4b8a-8b99-6b634e337054] Detach interface failed, port_id=9584dacf-1833-4a40-bfbe-56d7d5ea9998, reason: Instance 4c2c973a-5d23-4b8a-8b99-6b634e337054 could not be found. {{(pid=61923) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 925.775974] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-edec20df-41c5-4b44-b67e-1c51779999ca tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] [instance: 62c75dfd-6046-4de1-b1e3-13a307af1394] Creating linked-clone VM from snapshot {{(pid=61923) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 925.781382] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-77c6c5e9-9d9d-4153-b263-04e8f19d16e1 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.788114] env[61923]: DEBUG oslo_vmware.api [None req-b04cd48e-3d8d-4840-b909-cd681fdc6a00 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Task: {'id': task-1377814, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.417345} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 925.788114] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-b04cd48e-3d8d-4840-b909-cd681fdc6a00 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Deleted the datastore file {{(pid=61923) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 925.788114] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-b04cd48e-3d8d-4840-b909-cd681fdc6a00 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] [instance: 422d06d6-6932-46c3-bb25-841e6f65c028] Deleted contents of the VM from datastore datastore2 {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 925.788114] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-b04cd48e-3d8d-4840-b909-cd681fdc6a00 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] [instance: 422d06d6-6932-46c3-bb25-841e6f65c028] Instance destroyed {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 925.788114] env[61923]: INFO nova.compute.manager [None req-b04cd48e-3d8d-4840-b909-cd681fdc6a00 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] [instance: 422d06d6-6932-46c3-bb25-841e6f65c028] Took 1.13 seconds to destroy the instance on the hypervisor. [ 925.788114] env[61923]: DEBUG oslo.service.loopingcall [None req-b04cd48e-3d8d-4840-b909-cd681fdc6a00 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61923) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 925.788537] env[61923]: DEBUG nova.compute.manager [-] [instance: 422d06d6-6932-46c3-bb25-841e6f65c028] Deallocating network for instance {{(pid=61923) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 925.788537] env[61923]: DEBUG nova.network.neutron [-] [instance: 422d06d6-6932-46c3-bb25-841e6f65c028] deallocate_for_instance() {{(pid=61923) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 925.795526] env[61923]: DEBUG oslo_vmware.api [None req-edec20df-41c5-4b44-b67e-1c51779999ca tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Waiting for the task: (returnval){ [ 925.795526] env[61923]: value = "task-1377815" [ 925.795526] env[61923]: _type = "Task" [ 925.795526] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 925.806540] env[61923]: DEBUG oslo_vmware.api [None req-edec20df-41c5-4b44-b67e-1c51779999ca tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Task: {'id': task-1377815, 'name': CloneVM_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.066458] env[61923]: DEBUG nova.objects.instance [None req-40624371-b6ea-4bd8-a64e-58d8640198f8 tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Lazy-loading 'flavor' on Instance uuid b779d183-89ae-4e4d-ae99-e514e145ed43 {{(pid=61923) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 926.084331] env[61923]: INFO nova.compute.manager [-] [instance: 4c2c973a-5d23-4b8a-8b99-6b634e337054] Took 1.37 seconds to deallocate network for instance. [ 926.100641] env[61923]: DEBUG nova.network.neutron [None req-f3aa73ee-e2c2-44a8-b10f-5d20f9062f30 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] [instance: c98c96a5-d18e-4519-996a-949ded16225f] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 926.202977] env[61923]: DEBUG nova.network.neutron [None req-f3aa73ee-e2c2-44a8-b10f-5d20f9062f30 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] [instance: c98c96a5-d18e-4519-996a-949ded16225f] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 926.206920] env[61923]: DEBUG nova.compute.manager [req-73f830da-764e-489b-ad17-affc1109c1cf req-a0620548-b99f-45c8-949c-1119410fd12b service nova] [instance: 422d06d6-6932-46c3-bb25-841e6f65c028] Received event network-vif-deleted-97caa1b8-427c-46b0-be7a-d59b932900c3 {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 926.208088] env[61923]: INFO nova.compute.manager [req-73f830da-764e-489b-ad17-affc1109c1cf req-a0620548-b99f-45c8-949c-1119410fd12b service nova] [instance: 422d06d6-6932-46c3-bb25-841e6f65c028] Neutron deleted interface 97caa1b8-427c-46b0-be7a-d59b932900c3; detaching it from the instance and deleting it from the info cache [ 926.208088] env[61923]: DEBUG nova.network.neutron [req-73f830da-764e-489b-ad17-affc1109c1cf req-a0620548-b99f-45c8-949c-1119410fd12b service nova] [instance: 422d06d6-6932-46c3-bb25-841e6f65c028] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 926.287354] env[61923]: DEBUG oslo_concurrency.lockutils [None req-bc2e42ae-45d5-481f-a5ba-4796c19cafe7 tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Acquiring lock "b779d183-89ae-4e4d-ae99-e514e145ed43" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 926.305852] env[61923]: DEBUG oslo_vmware.api [None req-edec20df-41c5-4b44-b67e-1c51779999ca tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Task: {'id': task-1377815, 'name': CloneVM_Task} progress is 94%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.572544] env[61923]: DEBUG oslo_concurrency.lockutils [None req-40624371-b6ea-4bd8-a64e-58d8640198f8 tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Lock "b779d183-89ae-4e4d-ae99-e514e145ed43" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 10.863s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 926.572814] env[61923]: DEBUG oslo_concurrency.lockutils [None req-bc2e42ae-45d5-481f-a5ba-4796c19cafe7 tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Lock "b779d183-89ae-4e4d-ae99-e514e145ed43" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.286s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 926.573217] env[61923]: DEBUG nova.compute.manager [None req-bc2e42ae-45d5-481f-a5ba-4796c19cafe7 tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] [instance: b779d183-89ae-4e4d-ae99-e514e145ed43] Checking state {{(pid=61923) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 926.574138] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50571ade-5632-4bad-b9fe-2043e34c41e0 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.580916] env[61923]: DEBUG nova.compute.manager [None req-bc2e42ae-45d5-481f-a5ba-4796c19cafe7 tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] [instance: b779d183-89ae-4e4d-ae99-e514e145ed43] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=61923) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 926.581494] env[61923]: DEBUG nova.objects.instance [None req-bc2e42ae-45d5-481f-a5ba-4796c19cafe7 tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Lazy-loading 'flavor' on Instance uuid b779d183-89ae-4e4d-ae99-e514e145ed43 {{(pid=61923) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 926.586174] env[61923]: DEBUG nova.network.neutron [-] [instance: 422d06d6-6932-46c3-bb25-841e6f65c028] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 926.589893] env[61923]: DEBUG oslo_concurrency.lockutils [None req-91fc90ee-9f70-48af-84f3-b299c046606b tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 926.590156] env[61923]: DEBUG oslo_concurrency.lockutils [None req-91fc90ee-9f70-48af-84f3-b299c046606b tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 926.590532] env[61923]: DEBUG nova.objects.instance [None req-91fc90ee-9f70-48af-84f3-b299c046606b tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Lazy-loading 'resources' on Instance uuid 4c2c973a-5d23-4b8a-8b99-6b634e337054 {{(pid=61923) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 926.607800] env[61923]: DEBUG nova.network.neutron [-] [instance: 2719569b-8572-4199-8158-7bb367d17dc5] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 926.710809] env[61923]: DEBUG oslo_concurrency.lockutils [None req-f3aa73ee-e2c2-44a8-b10f-5d20f9062f30 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Releasing lock "refresh_cache-c98c96a5-d18e-4519-996a-949ded16225f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 926.711293] env[61923]: DEBUG nova.compute.manager [None req-f3aa73ee-e2c2-44a8-b10f-5d20f9062f30 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] [instance: c98c96a5-d18e-4519-996a-949ded16225f] Start destroying the instance on the hypervisor. {{(pid=61923) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 926.711488] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-f3aa73ee-e2c2-44a8-b10f-5d20f9062f30 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] [instance: c98c96a5-d18e-4519-996a-949ded16225f] Destroying instance {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 926.711777] env[61923]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e78061b0-202d-448a-a40e-a1a037949f8c {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.714945] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1084563b-9285-4b68-8171-a4d1e594ea91 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.723531] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-f3aa73ee-e2c2-44a8-b10f-5d20f9062f30 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] [instance: c98c96a5-d18e-4519-996a-949ded16225f] Powering off the VM {{(pid=61923) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 926.724758] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2a82bdcd-8b19-4056-b093-b743ee1cc4ef {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.729390] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9632626f-4097-403f-83c0-3f33692764fc {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.749464] env[61923]: DEBUG oslo_vmware.api [None req-f3aa73ee-e2c2-44a8-b10f-5d20f9062f30 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Waiting for the task: (returnval){ [ 926.749464] env[61923]: value = "task-1377816" [ 926.749464] env[61923]: _type = "Task" [ 926.749464] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 926.759851] env[61923]: DEBUG nova.compute.manager [req-73f830da-764e-489b-ad17-affc1109c1cf req-a0620548-b99f-45c8-949c-1119410fd12b service nova] [instance: 422d06d6-6932-46c3-bb25-841e6f65c028] Detach interface failed, port_id=97caa1b8-427c-46b0-be7a-d59b932900c3, reason: Instance 422d06d6-6932-46c3-bb25-841e6f65c028 could not be found. {{(pid=61923) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 926.765274] env[61923]: DEBUG oslo_vmware.api [None req-f3aa73ee-e2c2-44a8-b10f-5d20f9062f30 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Task: {'id': task-1377816, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.809325] env[61923]: DEBUG oslo_vmware.api [None req-edec20df-41c5-4b44-b67e-1c51779999ca tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Task: {'id': task-1377815, 'name': CloneVM_Task} progress is 95%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.085854] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-bc2e42ae-45d5-481f-a5ba-4796c19cafe7 tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] [instance: b779d183-89ae-4e4d-ae99-e514e145ed43] Powering off the VM {{(pid=61923) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 927.086142] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9ea42a22-c77c-4342-9a25-003a61dd867a {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.088042] env[61923]: INFO nova.compute.manager [-] [instance: 422d06d6-6932-46c3-bb25-841e6f65c028] Took 1.30 seconds to deallocate network for instance. [ 927.095901] env[61923]: DEBUG oslo_vmware.api [None req-bc2e42ae-45d5-481f-a5ba-4796c19cafe7 tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Waiting for the task: (returnval){ [ 927.095901] env[61923]: value = "task-1377817" [ 927.095901] env[61923]: _type = "Task" [ 927.095901] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 927.104570] env[61923]: DEBUG oslo_vmware.api [None req-bc2e42ae-45d5-481f-a5ba-4796c19cafe7 tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Task: {'id': task-1377817, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.110377] env[61923]: INFO nova.compute.manager [-] [instance: 2719569b-8572-4199-8158-7bb367d17dc5] Took 1.45 seconds to deallocate network for instance. [ 927.160086] env[61923]: DEBUG nova.compute.manager [req-a60ed81b-5954-42c0-b3e3-948fe4119520 req-d1ede622-3036-4200-babe-831f677cab9d service nova] [instance: 2719569b-8572-4199-8158-7bb367d17dc5] Received event network-vif-deleted-e092aadb-6904-43bc-b14c-0b1e010c4b2f {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 927.220971] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93588b4b-273a-4646-a2e7-caa1128a237f {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.229275] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f593cee-5dc4-46ca-8f55-77547ef1c855 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.263968] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0930503d-392d-4b05-aa10-c5919bcd6f14 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.274307] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5250dc1e-e062-42f2-8664-748253a3f724 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.278040] env[61923]: DEBUG oslo_vmware.api [None req-f3aa73ee-e2c2-44a8-b10f-5d20f9062f30 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Task: {'id': task-1377816, 'name': PowerOffVM_Task, 'duration_secs': 0.236057} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 927.278311] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-f3aa73ee-e2c2-44a8-b10f-5d20f9062f30 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] [instance: c98c96a5-d18e-4519-996a-949ded16225f] Powered off the VM {{(pid=61923) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 927.278504] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-f3aa73ee-e2c2-44a8-b10f-5d20f9062f30 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] [instance: c98c96a5-d18e-4519-996a-949ded16225f] Unregistering the VM {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 927.279067] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-85421c50-67fb-4360-a58a-6b5dd6ae69f8 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.288911] env[61923]: DEBUG nova.compute.provider_tree [None req-91fc90ee-9f70-48af-84f3-b299c046606b tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Updating inventory in ProviderTree for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 927.305036] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-f3aa73ee-e2c2-44a8-b10f-5d20f9062f30 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] [instance: c98c96a5-d18e-4519-996a-949ded16225f] Unregistered the VM {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 927.305282] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-f3aa73ee-e2c2-44a8-b10f-5d20f9062f30 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] [instance: c98c96a5-d18e-4519-996a-949ded16225f] Deleting contents of the VM from datastore datastore2 {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 927.305506] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-f3aa73ee-e2c2-44a8-b10f-5d20f9062f30 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Deleting the datastore file [datastore2] c98c96a5-d18e-4519-996a-949ded16225f {{(pid=61923) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 927.306017] env[61923]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-12b58866-acc2-4429-b57a-db64ee34fe94 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.311549] env[61923]: DEBUG oslo_vmware.api [None req-edec20df-41c5-4b44-b67e-1c51779999ca tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Task: {'id': task-1377815, 'name': CloneVM_Task, 'duration_secs': 1.330101} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 927.312151] env[61923]: INFO nova.virt.vmwareapi.vmops [None req-edec20df-41c5-4b44-b67e-1c51779999ca tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] [instance: 62c75dfd-6046-4de1-b1e3-13a307af1394] Created linked-clone VM from snapshot [ 927.312884] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b6dd546-7bfb-4b27-a81e-2b4e76fbe980 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.316727] env[61923]: DEBUG oslo_vmware.api [None req-f3aa73ee-e2c2-44a8-b10f-5d20f9062f30 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Waiting for the task: (returnval){ [ 927.316727] env[61923]: value = "task-1377819" [ 927.316727] env[61923]: _type = "Task" [ 927.316727] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 927.323313] env[61923]: DEBUG nova.virt.vmwareapi.images [None req-edec20df-41c5-4b44-b67e-1c51779999ca tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] [instance: 62c75dfd-6046-4de1-b1e3-13a307af1394] Uploading image dab00e2d-38e9-4263-8fcb-5192869ee6d8 {{(pid=61923) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 927.331357] env[61923]: DEBUG oslo_vmware.api [None req-f3aa73ee-e2c2-44a8-b10f-5d20f9062f30 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Task: {'id': task-1377819, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.338060] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-edec20df-41c5-4b44-b67e-1c51779999ca tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] [instance: 62c75dfd-6046-4de1-b1e3-13a307af1394] Destroying the VM {{(pid=61923) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 927.338325] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-bac50da7-59f9-4778-80f9-2fc7fd891ae6 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.344695] env[61923]: DEBUG oslo_vmware.api [None req-edec20df-41c5-4b44-b67e-1c51779999ca tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Waiting for the task: (returnval){ [ 927.344695] env[61923]: value = "task-1377820" [ 927.344695] env[61923]: _type = "Task" [ 927.344695] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 927.354907] env[61923]: DEBUG oslo_vmware.api [None req-edec20df-41c5-4b44-b67e-1c51779999ca tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Task: {'id': task-1377820, 'name': Destroy_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.596759] env[61923]: DEBUG oslo_concurrency.lockutils [None req-b04cd48e-3d8d-4840-b909-cd681fdc6a00 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 927.605625] env[61923]: DEBUG oslo_vmware.api [None req-bc2e42ae-45d5-481f-a5ba-4796c19cafe7 tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Task: {'id': task-1377817, 'name': PowerOffVM_Task, 'duration_secs': 0.317405} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 927.606103] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-bc2e42ae-45d5-481f-a5ba-4796c19cafe7 tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] [instance: b779d183-89ae-4e4d-ae99-e514e145ed43] Powered off the VM {{(pid=61923) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 927.606182] env[61923]: DEBUG nova.compute.manager [None req-bc2e42ae-45d5-481f-a5ba-4796c19cafe7 tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] [instance: b779d183-89ae-4e4d-ae99-e514e145ed43] Checking state {{(pid=61923) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 927.606878] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72064765-be7a-4fef-8278-72e77088801b {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.619860] env[61923]: DEBUG oslo_concurrency.lockutils [None req-3856851e-8c34-4281-b1ef-81601f186c03 tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 927.821505] env[61923]: DEBUG nova.scheduler.client.report [None req-91fc90ee-9f70-48af-84f3-b299c046606b tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Updated inventory for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 with generation 100 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 927.821776] env[61923]: DEBUG nova.compute.provider_tree [None req-91fc90ee-9f70-48af-84f3-b299c046606b tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Updating resource provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 generation from 100 to 101 during operation: update_inventory {{(pid=61923) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 927.821958] env[61923]: DEBUG nova.compute.provider_tree [None req-91fc90ee-9f70-48af-84f3-b299c046606b tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Updating inventory in ProviderTree for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 927.831567] env[61923]: DEBUG oslo_vmware.api [None req-f3aa73ee-e2c2-44a8-b10f-5d20f9062f30 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Task: {'id': task-1377819, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.128891} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 927.831822] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-f3aa73ee-e2c2-44a8-b10f-5d20f9062f30 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Deleted the datastore file {{(pid=61923) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 927.832020] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-f3aa73ee-e2c2-44a8-b10f-5d20f9062f30 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] [instance: c98c96a5-d18e-4519-996a-949ded16225f] Deleted contents of the VM from datastore datastore2 {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 927.835313] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-f3aa73ee-e2c2-44a8-b10f-5d20f9062f30 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] [instance: c98c96a5-d18e-4519-996a-949ded16225f] Instance destroyed {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 927.835520] env[61923]: INFO nova.compute.manager [None req-f3aa73ee-e2c2-44a8-b10f-5d20f9062f30 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] [instance: c98c96a5-d18e-4519-996a-949ded16225f] Took 1.12 seconds to destroy the instance on the hypervisor. [ 927.835864] env[61923]: DEBUG oslo.service.loopingcall [None req-f3aa73ee-e2c2-44a8-b10f-5d20f9062f30 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61923) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 927.836152] env[61923]: DEBUG nova.compute.manager [-] [instance: c98c96a5-d18e-4519-996a-949ded16225f] Deallocating network for instance {{(pid=61923) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 927.838530] env[61923]: DEBUG nova.network.neutron [-] [instance: c98c96a5-d18e-4519-996a-949ded16225f] deallocate_for_instance() {{(pid=61923) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 927.854862] env[61923]: DEBUG oslo_vmware.api [None req-edec20df-41c5-4b44-b67e-1c51779999ca tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Task: {'id': task-1377820, 'name': Destroy_Task} progress is 33%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.907159] env[61923]: DEBUG nova.network.neutron [-] [instance: c98c96a5-d18e-4519-996a-949ded16225f] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 928.118304] env[61923]: DEBUG oslo_concurrency.lockutils [None req-bc2e42ae-45d5-481f-a5ba-4796c19cafe7 tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Lock "b779d183-89ae-4e4d-ae99-e514e145ed43" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.545s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 928.330061] env[61923]: DEBUG oslo_concurrency.lockutils [None req-91fc90ee-9f70-48af-84f3-b299c046606b tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.738s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 928.332029] env[61923]: DEBUG oslo_concurrency.lockutils [None req-b04cd48e-3d8d-4840-b909-cd681fdc6a00 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.734s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 928.336564] env[61923]: DEBUG nova.objects.instance [None req-b04cd48e-3d8d-4840-b909-cd681fdc6a00 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Lazy-loading 'resources' on Instance uuid 422d06d6-6932-46c3-bb25-841e6f65c028 {{(pid=61923) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 928.355724] env[61923]: INFO nova.scheduler.client.report [None req-91fc90ee-9f70-48af-84f3-b299c046606b tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Deleted allocations for instance 4c2c973a-5d23-4b8a-8b99-6b634e337054 [ 928.356872] env[61923]: DEBUG oslo_vmware.api [None req-edec20df-41c5-4b44-b67e-1c51779999ca tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Task: {'id': task-1377820, 'name': Destroy_Task, 'duration_secs': 0.666992} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 928.359160] env[61923]: INFO nova.virt.vmwareapi.vm_util [None req-edec20df-41c5-4b44-b67e-1c51779999ca tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] [instance: 62c75dfd-6046-4de1-b1e3-13a307af1394] Destroyed the VM [ 928.359442] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-edec20df-41c5-4b44-b67e-1c51779999ca tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] [instance: 62c75dfd-6046-4de1-b1e3-13a307af1394] Deleting Snapshot of the VM instance {{(pid=61923) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 928.360065] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-e726d7cb-8811-4761-93b0-1ef8d1bdc3c0 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.366569] env[61923]: DEBUG oslo_vmware.api [None req-edec20df-41c5-4b44-b67e-1c51779999ca tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Waiting for the task: (returnval){ [ 928.366569] env[61923]: value = "task-1377821" [ 928.366569] env[61923]: _type = "Task" [ 928.366569] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 928.374830] env[61923]: DEBUG oslo_vmware.api [None req-edec20df-41c5-4b44-b67e-1c51779999ca tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Task: {'id': task-1377821, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.409958] env[61923]: DEBUG nova.network.neutron [-] [instance: c98c96a5-d18e-4519-996a-949ded16225f] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 928.795979] env[61923]: DEBUG nova.objects.instance [None req-e18f4a96-a9c1-4347-ad97-ec2c6f662dce tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Lazy-loading 'flavor' on Instance uuid b779d183-89ae-4e4d-ae99-e514e145ed43 {{(pid=61923) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 928.863188] env[61923]: DEBUG oslo_concurrency.lockutils [None req-91fc90ee-9f70-48af-84f3-b299c046606b tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Lock "4c2c973a-5d23-4b8a-8b99-6b634e337054" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 8.296s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 928.882629] env[61923]: DEBUG oslo_vmware.api [None req-edec20df-41c5-4b44-b67e-1c51779999ca tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Task: {'id': task-1377821, 'name': RemoveSnapshot_Task, 'duration_secs': 0.491572} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 928.886976] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-edec20df-41c5-4b44-b67e-1c51779999ca tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] [instance: 62c75dfd-6046-4de1-b1e3-13a307af1394] Deleted Snapshot of the VM instance {{(pid=61923) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 928.914729] env[61923]: INFO nova.compute.manager [-] [instance: c98c96a5-d18e-4519-996a-949ded16225f] Took 1.08 seconds to deallocate network for instance. [ 928.963030] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4002e6ef-ef48-4633-ba1c-6b1fea3dce8a {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.969372] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cabd50b7-2d2c-4922-b596-cc20a35031c5 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.003471] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5dedb30b-e6f8-422f-bc8e-c020cbec2199 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.011674] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9071a672-5767-4160-b0b7-f687b72be99b {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.027396] env[61923]: DEBUG nova.compute.provider_tree [None req-b04cd48e-3d8d-4840-b909-cd681fdc6a00 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 929.301359] env[61923]: DEBUG oslo_concurrency.lockutils [None req-e18f4a96-a9c1-4347-ad97-ec2c6f662dce tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Acquiring lock "refresh_cache-b779d183-89ae-4e4d-ae99-e514e145ed43" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 929.301542] env[61923]: DEBUG oslo_concurrency.lockutils [None req-e18f4a96-a9c1-4347-ad97-ec2c6f662dce tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Acquired lock "refresh_cache-b779d183-89ae-4e4d-ae99-e514e145ed43" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 929.301722] env[61923]: DEBUG nova.network.neutron [None req-e18f4a96-a9c1-4347-ad97-ec2c6f662dce tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] [instance: b779d183-89ae-4e4d-ae99-e514e145ed43] Building network info cache for instance {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 929.301900] env[61923]: DEBUG nova.objects.instance [None req-e18f4a96-a9c1-4347-ad97-ec2c6f662dce tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Lazy-loading 'info_cache' on Instance uuid b779d183-89ae-4e4d-ae99-e514e145ed43 {{(pid=61923) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 929.393051] env[61923]: WARNING nova.compute.manager [None req-edec20df-41c5-4b44-b67e-1c51779999ca tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] [instance: 62c75dfd-6046-4de1-b1e3-13a307af1394] Image not found during snapshot: nova.exception.ImageNotFound: Image dab00e2d-38e9-4263-8fcb-5192869ee6d8 could not be found. [ 929.422424] env[61923]: DEBUG oslo_concurrency.lockutils [None req-f3aa73ee-e2c2-44a8-b10f-5d20f9062f30 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 929.530340] env[61923]: DEBUG nova.scheduler.client.report [None req-b04cd48e-3d8d-4840-b909-cd681fdc6a00 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 929.588634] env[61923]: DEBUG oslo_concurrency.lockutils [None req-a9734b29-9738-4ea3-b52c-d0194f5d6fd9 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Acquiring lock "02a8f197-80bb-4cee-bdd6-b07705759986" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 929.588904] env[61923]: DEBUG oslo_concurrency.lockutils [None req-a9734b29-9738-4ea3-b52c-d0194f5d6fd9 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Lock "02a8f197-80bb-4cee-bdd6-b07705759986" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 929.806169] env[61923]: DEBUG nova.objects.base [None req-e18f4a96-a9c1-4347-ad97-ec2c6f662dce tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Object Instance lazy-loaded attributes: flavor,info_cache {{(pid=61923) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 930.035075] env[61923]: DEBUG oslo_concurrency.lockutils [None req-b04cd48e-3d8d-4840-b909-cd681fdc6a00 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.705s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 930.037419] env[61923]: DEBUG oslo_concurrency.lockutils [None req-3856851e-8c34-4281-b1ef-81601f186c03 tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.418s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 930.038255] env[61923]: DEBUG nova.objects.instance [None req-3856851e-8c34-4281-b1ef-81601f186c03 tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Lazy-loading 'resources' on Instance uuid 2719569b-8572-4199-8158-7bb367d17dc5 {{(pid=61923) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 930.056417] env[61923]: INFO nova.scheduler.client.report [None req-b04cd48e-3d8d-4840-b909-cd681fdc6a00 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Deleted allocations for instance 422d06d6-6932-46c3-bb25-841e6f65c028 [ 930.091720] env[61923]: DEBUG nova.compute.manager [None req-a9734b29-9738-4ea3-b52c-d0194f5d6fd9 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: 02a8f197-80bb-4cee-bdd6-b07705759986] Starting instance... {{(pid=61923) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 930.261638] env[61923]: DEBUG oslo_concurrency.lockutils [None req-cc72dab8-aeca-42e3-aa89-07d278f05c01 tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Acquiring lock "62c75dfd-6046-4de1-b1e3-13a307af1394" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 930.262037] env[61923]: DEBUG oslo_concurrency.lockutils [None req-cc72dab8-aeca-42e3-aa89-07d278f05c01 tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Lock "62c75dfd-6046-4de1-b1e3-13a307af1394" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 930.262290] env[61923]: DEBUG oslo_concurrency.lockutils [None req-cc72dab8-aeca-42e3-aa89-07d278f05c01 tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Acquiring lock "62c75dfd-6046-4de1-b1e3-13a307af1394-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 930.262604] env[61923]: DEBUG oslo_concurrency.lockutils [None req-cc72dab8-aeca-42e3-aa89-07d278f05c01 tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Lock "62c75dfd-6046-4de1-b1e3-13a307af1394-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 930.262695] env[61923]: DEBUG oslo_concurrency.lockutils [None req-cc72dab8-aeca-42e3-aa89-07d278f05c01 tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Lock "62c75dfd-6046-4de1-b1e3-13a307af1394-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 930.265255] env[61923]: INFO nova.compute.manager [None req-cc72dab8-aeca-42e3-aa89-07d278f05c01 tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] [instance: 62c75dfd-6046-4de1-b1e3-13a307af1394] Terminating instance [ 930.267203] env[61923]: DEBUG nova.compute.manager [None req-cc72dab8-aeca-42e3-aa89-07d278f05c01 tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] [instance: 62c75dfd-6046-4de1-b1e3-13a307af1394] Start destroying the instance on the hypervisor. {{(pid=61923) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 930.267425] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-cc72dab8-aeca-42e3-aa89-07d278f05c01 tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] [instance: 62c75dfd-6046-4de1-b1e3-13a307af1394] Destroying instance {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 930.268262] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4bb019f0-930d-4f59-b2d2-9c888883528c {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.276285] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-cc72dab8-aeca-42e3-aa89-07d278f05c01 tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] [instance: 62c75dfd-6046-4de1-b1e3-13a307af1394] Powering off the VM {{(pid=61923) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 930.276522] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-99cff92e-8435-48fe-b74f-d2bb67f0d77b {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.283491] env[61923]: DEBUG oslo_vmware.api [None req-cc72dab8-aeca-42e3-aa89-07d278f05c01 tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Waiting for the task: (returnval){ [ 930.283491] env[61923]: value = "task-1377822" [ 930.283491] env[61923]: _type = "Task" [ 930.283491] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 930.290927] env[61923]: DEBUG oslo_vmware.api [None req-cc72dab8-aeca-42e3-aa89-07d278f05c01 tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Task: {'id': task-1377822, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 930.543857] env[61923]: DEBUG nova.network.neutron [None req-e18f4a96-a9c1-4347-ad97-ec2c6f662dce tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] [instance: b779d183-89ae-4e4d-ae99-e514e145ed43] Updating instance_info_cache with network_info: [{"id": "d49a3010-c84a-4bef-beed-9ca4aa7d5dc2", "address": "fa:16:3e:48:e6:a2", "network": {"id": "d2905082-7ca4-4f92-a142-bb85211bb7c4", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-885218622-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.251", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a13d827933fa4597984afdb91dbbdd39", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "03ac2c9c-6ad2-4a85-bfab-c7e336df859a", "external-id": "nsx-vlan-transportzone-379", "segmentation_id": 379, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd49a3010-c8", "ovs_interfaceid": "d49a3010-c84a-4bef-beed-9ca4aa7d5dc2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 930.563895] env[61923]: DEBUG oslo_concurrency.lockutils [None req-b04cd48e-3d8d-4840-b909-cd681fdc6a00 tempest-AttachInterfacesTestJSON-1269801032 tempest-AttachInterfacesTestJSON-1269801032-project-member] Lock "422d06d6-6932-46c3-bb25-841e6f65c028" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.911s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 930.615082] env[61923]: DEBUG oslo_concurrency.lockutils [None req-a9734b29-9738-4ea3-b52c-d0194f5d6fd9 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 930.624872] env[61923]: DEBUG oslo_concurrency.lockutils [None req-29a01eb2-e144-458f-9b46-c49e64b03e88 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Acquiring lock "a511392e-9ab6-42fb-b07c-f90e9786dcc1" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 930.624872] env[61923]: DEBUG oslo_concurrency.lockutils [None req-29a01eb2-e144-458f-9b46-c49e64b03e88 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Lock "a511392e-9ab6-42fb-b07c-f90e9786dcc1" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 930.625016] env[61923]: INFO nova.compute.manager [None req-29a01eb2-e144-458f-9b46-c49e64b03e88 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] [instance: a511392e-9ab6-42fb-b07c-f90e9786dcc1] Shelving [ 930.642905] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d2f6477-b23c-4cf6-a972-f3072d112ed4 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.651128] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1589ad1e-ecb7-4f68-9a5d-bb1f5572ef40 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.685679] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96603cd3-ca59-44d7-a57a-c4190691e6be {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.693017] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af536d1d-a99c-4362-a84e-600015606f98 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.711164] env[61923]: DEBUG nova.compute.provider_tree [None req-3856851e-8c34-4281-b1ef-81601f186c03 tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 930.793637] env[61923]: DEBUG oslo_vmware.api [None req-cc72dab8-aeca-42e3-aa89-07d278f05c01 tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Task: {'id': task-1377822, 'name': PowerOffVM_Task, 'duration_secs': 0.22325} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 930.794131] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-cc72dab8-aeca-42e3-aa89-07d278f05c01 tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] [instance: 62c75dfd-6046-4de1-b1e3-13a307af1394] Powered off the VM {{(pid=61923) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 930.794359] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-cc72dab8-aeca-42e3-aa89-07d278f05c01 tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] [instance: 62c75dfd-6046-4de1-b1e3-13a307af1394] Unregistering the VM {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 930.794625] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-60b61c47-2770-4892-9390-f35b2909a4d9 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.863821] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-cc72dab8-aeca-42e3-aa89-07d278f05c01 tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] [instance: 62c75dfd-6046-4de1-b1e3-13a307af1394] Unregistered the VM {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 930.864066] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-cc72dab8-aeca-42e3-aa89-07d278f05c01 tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] [instance: 62c75dfd-6046-4de1-b1e3-13a307af1394] Deleting contents of the VM from datastore datastore1 {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 930.864273] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-cc72dab8-aeca-42e3-aa89-07d278f05c01 tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Deleting the datastore file [datastore1] 62c75dfd-6046-4de1-b1e3-13a307af1394 {{(pid=61923) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 930.864586] env[61923]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-7d419b12-9ec0-4085-bc0e-2db19cd126a9 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.870563] env[61923]: DEBUG oslo_vmware.api [None req-cc72dab8-aeca-42e3-aa89-07d278f05c01 tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Waiting for the task: (returnval){ [ 930.870563] env[61923]: value = "task-1377824" [ 930.870563] env[61923]: _type = "Task" [ 930.870563] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 930.877974] env[61923]: DEBUG oslo_vmware.api [None req-cc72dab8-aeca-42e3-aa89-07d278f05c01 tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Task: {'id': task-1377824, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.046929] env[61923]: DEBUG oslo_concurrency.lockutils [None req-e18f4a96-a9c1-4347-ad97-ec2c6f662dce tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Releasing lock "refresh_cache-b779d183-89ae-4e4d-ae99-e514e145ed43" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 931.133476] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-29a01eb2-e144-458f-9b46-c49e64b03e88 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] [instance: a511392e-9ab6-42fb-b07c-f90e9786dcc1] Powering off the VM {{(pid=61923) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 931.133774] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c88cbdba-bd1b-4a13-9c48-0d25f363ee83 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.141499] env[61923]: DEBUG oslo_vmware.api [None req-29a01eb2-e144-458f-9b46-c49e64b03e88 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Waiting for the task: (returnval){ [ 931.141499] env[61923]: value = "task-1377825" [ 931.141499] env[61923]: _type = "Task" [ 931.141499] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 931.149842] env[61923]: DEBUG oslo_vmware.api [None req-29a01eb2-e144-458f-9b46-c49e64b03e88 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Task: {'id': task-1377825, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.214251] env[61923]: DEBUG nova.scheduler.client.report [None req-3856851e-8c34-4281-b1ef-81601f186c03 tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 931.383305] env[61923]: DEBUG oslo_vmware.api [None req-cc72dab8-aeca-42e3-aa89-07d278f05c01 tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Task: {'id': task-1377824, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.150167} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 931.383603] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-cc72dab8-aeca-42e3-aa89-07d278f05c01 tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Deleted the datastore file {{(pid=61923) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 931.383797] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-cc72dab8-aeca-42e3-aa89-07d278f05c01 tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] [instance: 62c75dfd-6046-4de1-b1e3-13a307af1394] Deleted contents of the VM from datastore datastore1 {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 931.383976] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-cc72dab8-aeca-42e3-aa89-07d278f05c01 tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] [instance: 62c75dfd-6046-4de1-b1e3-13a307af1394] Instance destroyed {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 931.384173] env[61923]: INFO nova.compute.manager [None req-cc72dab8-aeca-42e3-aa89-07d278f05c01 tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] [instance: 62c75dfd-6046-4de1-b1e3-13a307af1394] Took 1.12 seconds to destroy the instance on the hypervisor. [ 931.384417] env[61923]: DEBUG oslo.service.loopingcall [None req-cc72dab8-aeca-42e3-aa89-07d278f05c01 tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61923) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 931.384617] env[61923]: DEBUG nova.compute.manager [-] [instance: 62c75dfd-6046-4de1-b1e3-13a307af1394] Deallocating network for instance {{(pid=61923) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 931.384711] env[61923]: DEBUG nova.network.neutron [-] [instance: 62c75dfd-6046-4de1-b1e3-13a307af1394] deallocate_for_instance() {{(pid=61923) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 931.550592] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-e18f4a96-a9c1-4347-ad97-ec2c6f662dce tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] [instance: b779d183-89ae-4e4d-ae99-e514e145ed43] Powering on the VM {{(pid=61923) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 931.550943] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5687bfd1-f1d6-4920-8e17-4b9df282e590 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.558860] env[61923]: DEBUG oslo_vmware.api [None req-e18f4a96-a9c1-4347-ad97-ec2c6f662dce tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Waiting for the task: (returnval){ [ 931.558860] env[61923]: value = "task-1377826" [ 931.558860] env[61923]: _type = "Task" [ 931.558860] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 931.566479] env[61923]: DEBUG oslo_vmware.api [None req-e18f4a96-a9c1-4347-ad97-ec2c6f662dce tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Task: {'id': task-1377826, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 931.570022] env[61923]: DEBUG oslo_service.periodic_task [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61923) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 931.570814] env[61923]: DEBUG oslo_service.periodic_task [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61923) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 931.570814] env[61923]: DEBUG nova.compute.manager [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Starting heal instance info cache {{(pid=61923) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 931.605205] env[61923]: DEBUG nova.compute.manager [req-1fc49890-0896-4dd3-9dc2-dbbc847cb242 req-ef2ad025-a84e-46c7-a546-87884da219f0 service nova] [instance: 62c75dfd-6046-4de1-b1e3-13a307af1394] Received event network-vif-deleted-2d89de21-00bd-46da-b7c1-6ac2cbba1982 {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 931.605792] env[61923]: INFO nova.compute.manager [req-1fc49890-0896-4dd3-9dc2-dbbc847cb242 req-ef2ad025-a84e-46c7-a546-87884da219f0 service nova] [instance: 62c75dfd-6046-4de1-b1e3-13a307af1394] Neutron deleted interface 2d89de21-00bd-46da-b7c1-6ac2cbba1982; detaching it from the instance and deleting it from the info cache [ 931.605792] env[61923]: DEBUG nova.network.neutron [req-1fc49890-0896-4dd3-9dc2-dbbc847cb242 req-ef2ad025-a84e-46c7-a546-87884da219f0 service nova] [instance: 62c75dfd-6046-4de1-b1e3-13a307af1394] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 931.650859] env[61923]: DEBUG oslo_vmware.api [None req-29a01eb2-e144-458f-9b46-c49e64b03e88 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Task: {'id': task-1377825, 'name': PowerOffVM_Task, 'duration_secs': 0.196582} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 931.651152] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-29a01eb2-e144-458f-9b46-c49e64b03e88 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] [instance: a511392e-9ab6-42fb-b07c-f90e9786dcc1] Powered off the VM {{(pid=61923) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 931.651952] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a09345e-4dfd-4beb-aea1-c018075a0ecd {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.672246] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce973444-d5aa-4caf-860b-230656a66e2d {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.719535] env[61923]: DEBUG oslo_concurrency.lockutils [None req-3856851e-8c34-4281-b1ef-81601f186c03 tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.682s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 931.724338] env[61923]: DEBUG oslo_concurrency.lockutils [None req-f3aa73ee-e2c2-44a8-b10f-5d20f9062f30 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.302s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 931.725038] env[61923]: DEBUG nova.objects.instance [None req-f3aa73ee-e2c2-44a8-b10f-5d20f9062f30 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Lazy-loading 'resources' on Instance uuid c98c96a5-d18e-4519-996a-949ded16225f {{(pid=61923) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 931.750074] env[61923]: INFO nova.scheduler.client.report [None req-3856851e-8c34-4281-b1ef-81601f186c03 tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Deleted allocations for instance 2719569b-8572-4199-8158-7bb367d17dc5 [ 932.068957] env[61923]: DEBUG oslo_vmware.api [None req-e18f4a96-a9c1-4347-ad97-ec2c6f662dce tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Task: {'id': task-1377826, 'name': PowerOnVM_Task} progress is 88%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 932.072537] env[61923]: DEBUG nova.compute.manager [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Didn't find any instances for network info cache update. {{(pid=61923) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10014}} [ 932.072744] env[61923]: DEBUG oslo_service.periodic_task [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61923) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 932.072902] env[61923]: DEBUG oslo_service.periodic_task [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61923) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 932.073957] env[61923]: DEBUG oslo_service.periodic_task [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61923) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 932.073957] env[61923]: DEBUG oslo_service.periodic_task [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61923) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 932.073957] env[61923]: DEBUG oslo_service.periodic_task [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61923) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 932.073957] env[61923]: DEBUG oslo_service.periodic_task [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61923) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 932.073957] env[61923]: DEBUG nova.compute.manager [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61923) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 932.073957] env[61923]: DEBUG oslo_service.periodic_task [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Running periodic task ComputeManager.update_available_resource {{(pid=61923) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 932.077748] env[61923]: DEBUG nova.network.neutron [-] [instance: 62c75dfd-6046-4de1-b1e3-13a307af1394] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 932.108543] env[61923]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-db06eee3-ee82-4746-92b3-a7f76f5b2237 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.118144] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f4bb370-c3f9-41f9-b6f5-fc3c3cf17559 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.143638] env[61923]: DEBUG nova.compute.manager [req-1fc49890-0896-4dd3-9dc2-dbbc847cb242 req-ef2ad025-a84e-46c7-a546-87884da219f0 service nova] [instance: 62c75dfd-6046-4de1-b1e3-13a307af1394] Detach interface failed, port_id=2d89de21-00bd-46da-b7c1-6ac2cbba1982, reason: Instance 62c75dfd-6046-4de1-b1e3-13a307af1394 could not be found. {{(pid=61923) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 932.184107] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-29a01eb2-e144-458f-9b46-c49e64b03e88 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] [instance: a511392e-9ab6-42fb-b07c-f90e9786dcc1] Creating Snapshot of the VM instance {{(pid=61923) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 932.184481] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-db9c0c39-744d-48a6-b7f4-c0e2e9c7db87 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.192156] env[61923]: DEBUG oslo_vmware.api [None req-29a01eb2-e144-458f-9b46-c49e64b03e88 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Waiting for the task: (returnval){ [ 932.192156] env[61923]: value = "task-1377827" [ 932.192156] env[61923]: _type = "Task" [ 932.192156] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 932.200800] env[61923]: DEBUG oslo_vmware.api [None req-29a01eb2-e144-458f-9b46-c49e64b03e88 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Task: {'id': task-1377827, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 932.261836] env[61923]: DEBUG oslo_concurrency.lockutils [None req-3856851e-8c34-4281-b1ef-81601f186c03 tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Lock "2719569b-8572-4199-8158-7bb367d17dc5" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 8.693s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 932.334978] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42b3416d-f85a-4ca4-9915-363d7c61c1f7 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.342582] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2a2e474-4ef6-4076-a0d4-b2a344435800 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.373878] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a51a478c-3fb4-4a63-8f6d-4a0d75b94c4e {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.381403] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2238a7d-df59-450e-bfae-2c1f13d8afca {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.396185] env[61923]: DEBUG nova.compute.provider_tree [None req-f3aa73ee-e2c2-44a8-b10f-5d20f9062f30 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 932.571023] env[61923]: DEBUG oslo_vmware.api [None req-e18f4a96-a9c1-4347-ad97-ec2c6f662dce tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Task: {'id': task-1377826, 'name': PowerOnVM_Task, 'duration_secs': 0.595652} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 932.571023] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-e18f4a96-a9c1-4347-ad97-ec2c6f662dce tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] [instance: b779d183-89ae-4e4d-ae99-e514e145ed43] Powered on the VM {{(pid=61923) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 932.571023] env[61923]: DEBUG nova.compute.manager [None req-e18f4a96-a9c1-4347-ad97-ec2c6f662dce tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] [instance: b779d183-89ae-4e4d-ae99-e514e145ed43] Checking state {{(pid=61923) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 932.571023] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c280e45-f5b7-4cd6-aa47-b7390d6ae3c9 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.576974] env[61923]: DEBUG oslo_concurrency.lockutils [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 932.579190] env[61923]: INFO nova.compute.manager [-] [instance: 62c75dfd-6046-4de1-b1e3-13a307af1394] Took 1.19 seconds to deallocate network for instance. [ 932.702115] env[61923]: DEBUG oslo_vmware.api [None req-29a01eb2-e144-458f-9b46-c49e64b03e88 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Task: {'id': task-1377827, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 932.788721] env[61923]: DEBUG oslo_concurrency.lockutils [None req-b9e3b0d3-c70e-4ebe-8e35-3d6fba68a21e tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Acquiring lock "87ae37dc-d0d0-4b76-afdd-0ba3e8f6ce0b" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 932.789231] env[61923]: DEBUG oslo_concurrency.lockutils [None req-b9e3b0d3-c70e-4ebe-8e35-3d6fba68a21e tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Lock "87ae37dc-d0d0-4b76-afdd-0ba3e8f6ce0b" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 932.789566] env[61923]: DEBUG oslo_concurrency.lockutils [None req-b9e3b0d3-c70e-4ebe-8e35-3d6fba68a21e tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Acquiring lock "87ae37dc-d0d0-4b76-afdd-0ba3e8f6ce0b-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 932.789889] env[61923]: DEBUG oslo_concurrency.lockutils [None req-b9e3b0d3-c70e-4ebe-8e35-3d6fba68a21e tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Lock "87ae37dc-d0d0-4b76-afdd-0ba3e8f6ce0b-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 932.790863] env[61923]: DEBUG oslo_concurrency.lockutils [None req-b9e3b0d3-c70e-4ebe-8e35-3d6fba68a21e tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Lock "87ae37dc-d0d0-4b76-afdd-0ba3e8f6ce0b-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 932.793751] env[61923]: INFO nova.compute.manager [None req-b9e3b0d3-c70e-4ebe-8e35-3d6fba68a21e tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] [instance: 87ae37dc-d0d0-4b76-afdd-0ba3e8f6ce0b] Terminating instance [ 932.796233] env[61923]: DEBUG nova.compute.manager [None req-b9e3b0d3-c70e-4ebe-8e35-3d6fba68a21e tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] [instance: 87ae37dc-d0d0-4b76-afdd-0ba3e8f6ce0b] Start destroying the instance on the hypervisor. {{(pid=61923) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 932.796993] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-b9e3b0d3-c70e-4ebe-8e35-3d6fba68a21e tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] [instance: 87ae37dc-d0d0-4b76-afdd-0ba3e8f6ce0b] Destroying instance {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 932.798227] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dbc616e2-b098-46b5-815d-ac3da82db25e {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.806206] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-b9e3b0d3-c70e-4ebe-8e35-3d6fba68a21e tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] [instance: 87ae37dc-d0d0-4b76-afdd-0ba3e8f6ce0b] Powering off the VM {{(pid=61923) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 932.806463] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-aa9b04a1-9f73-40cf-b469-7482aa37ce8e {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.812548] env[61923]: DEBUG oslo_vmware.api [None req-b9e3b0d3-c70e-4ebe-8e35-3d6fba68a21e tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Waiting for the task: (returnval){ [ 932.812548] env[61923]: value = "task-1377828" [ 932.812548] env[61923]: _type = "Task" [ 932.812548] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 932.822851] env[61923]: DEBUG oslo_vmware.api [None req-b9e3b0d3-c70e-4ebe-8e35-3d6fba68a21e tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Task: {'id': task-1377828, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 932.899672] env[61923]: DEBUG nova.scheduler.client.report [None req-f3aa73ee-e2c2-44a8-b10f-5d20f9062f30 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 933.086019] env[61923]: DEBUG oslo_concurrency.lockutils [None req-cc72dab8-aeca-42e3-aa89-07d278f05c01 tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 933.204868] env[61923]: DEBUG oslo_vmware.api [None req-29a01eb2-e144-458f-9b46-c49e64b03e88 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Task: {'id': task-1377827, 'name': CreateSnapshot_Task, 'duration_secs': 0.597799} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 933.205197] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-29a01eb2-e144-458f-9b46-c49e64b03e88 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] [instance: a511392e-9ab6-42fb-b07c-f90e9786dcc1] Created Snapshot of the VM instance {{(pid=61923) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 933.205953] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5ead248-2511-4fae-ac8a-95f0793fffb5 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.324157] env[61923]: DEBUG oslo_vmware.api [None req-b9e3b0d3-c70e-4ebe-8e35-3d6fba68a21e tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Task: {'id': task-1377828, 'name': PowerOffVM_Task, 'duration_secs': 0.190593} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 933.324520] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-b9e3b0d3-c70e-4ebe-8e35-3d6fba68a21e tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] [instance: 87ae37dc-d0d0-4b76-afdd-0ba3e8f6ce0b] Powered off the VM {{(pid=61923) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 933.324702] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-b9e3b0d3-c70e-4ebe-8e35-3d6fba68a21e tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] [instance: 87ae37dc-d0d0-4b76-afdd-0ba3e8f6ce0b] Unregistering the VM {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 933.324955] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-d660ac4c-2d0f-4a5f-a824-a2715eeb472a {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.387121] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-b9e3b0d3-c70e-4ebe-8e35-3d6fba68a21e tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] [instance: 87ae37dc-d0d0-4b76-afdd-0ba3e8f6ce0b] Unregistered the VM {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 933.387453] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-b9e3b0d3-c70e-4ebe-8e35-3d6fba68a21e tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] [instance: 87ae37dc-d0d0-4b76-afdd-0ba3e8f6ce0b] Deleting contents of the VM from datastore datastore1 {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 933.387670] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-b9e3b0d3-c70e-4ebe-8e35-3d6fba68a21e tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Deleting the datastore file [datastore1] 87ae37dc-d0d0-4b76-afdd-0ba3e8f6ce0b {{(pid=61923) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 933.387957] env[61923]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c9dbdd3c-823d-4ede-9daf-e1717407efbc {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.393924] env[61923]: DEBUG oslo_vmware.api [None req-b9e3b0d3-c70e-4ebe-8e35-3d6fba68a21e tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Waiting for the task: (returnval){ [ 933.393924] env[61923]: value = "task-1377830" [ 933.393924] env[61923]: _type = "Task" [ 933.393924] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 933.405430] env[61923]: DEBUG oslo_concurrency.lockutils [None req-f3aa73ee-e2c2-44a8-b10f-5d20f9062f30 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.681s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 933.407271] env[61923]: DEBUG oslo_vmware.api [None req-b9e3b0d3-c70e-4ebe-8e35-3d6fba68a21e tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Task: {'id': task-1377830, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.408025] env[61923]: DEBUG oslo_concurrency.lockutils [None req-a9734b29-9738-4ea3-b52c-d0194f5d6fd9 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.793s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 933.409189] env[61923]: INFO nova.compute.claims [None req-a9734b29-9738-4ea3-b52c-d0194f5d6fd9 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: 02a8f197-80bb-4cee-bdd6-b07705759986] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 933.429899] env[61923]: INFO nova.scheduler.client.report [None req-f3aa73ee-e2c2-44a8-b10f-5d20f9062f30 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Deleted allocations for instance c98c96a5-d18e-4519-996a-949ded16225f [ 933.724238] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-29a01eb2-e144-458f-9b46-c49e64b03e88 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] [instance: a511392e-9ab6-42fb-b07c-f90e9786dcc1] Creating linked-clone VM from snapshot {{(pid=61923) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 933.724591] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-2f49b9cb-3984-4543-a71e-6d07f60d8f47 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.733356] env[61923]: DEBUG oslo_vmware.api [None req-29a01eb2-e144-458f-9b46-c49e64b03e88 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Waiting for the task: (returnval){ [ 933.733356] env[61923]: value = "task-1377831" [ 933.733356] env[61923]: _type = "Task" [ 933.733356] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 933.741212] env[61923]: DEBUG oslo_vmware.api [None req-29a01eb2-e144-458f-9b46-c49e64b03e88 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Task: {'id': task-1377831, 'name': CloneVM_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.906456] env[61923]: DEBUG oslo_vmware.api [None req-b9e3b0d3-c70e-4ebe-8e35-3d6fba68a21e tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Task: {'id': task-1377830, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.446054} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 933.906726] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-b9e3b0d3-c70e-4ebe-8e35-3d6fba68a21e tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Deleted the datastore file {{(pid=61923) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 933.906916] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-b9e3b0d3-c70e-4ebe-8e35-3d6fba68a21e tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] [instance: 87ae37dc-d0d0-4b76-afdd-0ba3e8f6ce0b] Deleted contents of the VM from datastore datastore1 {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 933.907113] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-b9e3b0d3-c70e-4ebe-8e35-3d6fba68a21e tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] [instance: 87ae37dc-d0d0-4b76-afdd-0ba3e8f6ce0b] Instance destroyed {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 933.907358] env[61923]: INFO nova.compute.manager [None req-b9e3b0d3-c70e-4ebe-8e35-3d6fba68a21e tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] [instance: 87ae37dc-d0d0-4b76-afdd-0ba3e8f6ce0b] Took 1.11 seconds to destroy the instance on the hypervisor. [ 933.907629] env[61923]: DEBUG oslo.service.loopingcall [None req-b9e3b0d3-c70e-4ebe-8e35-3d6fba68a21e tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61923) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 933.907838] env[61923]: DEBUG nova.compute.manager [-] [instance: 87ae37dc-d0d0-4b76-afdd-0ba3e8f6ce0b] Deallocating network for instance {{(pid=61923) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 933.907937] env[61923]: DEBUG nova.network.neutron [-] [instance: 87ae37dc-d0d0-4b76-afdd-0ba3e8f6ce0b] deallocate_for_instance() {{(pid=61923) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 933.937834] env[61923]: DEBUG oslo_concurrency.lockutils [None req-f3aa73ee-e2c2-44a8-b10f-5d20f9062f30 tempest-ServerShowV247Test-1523384759 tempest-ServerShowV247Test-1523384759-project-member] Lock "c98c96a5-d18e-4519-996a-949ded16225f" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 8.392s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 934.249852] env[61923]: DEBUG oslo_vmware.api [None req-29a01eb2-e144-458f-9b46-c49e64b03e88 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Task: {'id': task-1377831, 'name': CloneVM_Task} progress is 94%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.304711] env[61923]: DEBUG nova.compute.manager [req-3a25c583-91d7-4155-87e0-3cb7bbfd2ff3 req-c13d9e3c-8777-40f0-a0e7-7f299dd45fb1 service nova] [instance: 87ae37dc-d0d0-4b76-afdd-0ba3e8f6ce0b] Received event network-vif-deleted-b0e2ee14-b051-47bc-9064-50f09916c0cb {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 934.304950] env[61923]: INFO nova.compute.manager [req-3a25c583-91d7-4155-87e0-3cb7bbfd2ff3 req-c13d9e3c-8777-40f0-a0e7-7f299dd45fb1 service nova] [instance: 87ae37dc-d0d0-4b76-afdd-0ba3e8f6ce0b] Neutron deleted interface b0e2ee14-b051-47bc-9064-50f09916c0cb; detaching it from the instance and deleting it from the info cache [ 934.307708] env[61923]: DEBUG nova.network.neutron [req-3a25c583-91d7-4155-87e0-3cb7bbfd2ff3 req-c13d9e3c-8777-40f0-a0e7-7f299dd45fb1 service nova] [instance: 87ae37dc-d0d0-4b76-afdd-0ba3e8f6ce0b] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 934.517255] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb960b33-33b5-4040-88d0-d676319599ec {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.525684] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e871167b-6eba-425b-876b-b8340f9a36df {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.561530] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-162dd174-c7e6-43ec-a1a4-146f546af08c {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.569992] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9bf9482-4940-4ccf-90c0-a0232a4e0ca7 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.584235] env[61923]: DEBUG nova.compute.provider_tree [None req-a9734b29-9738-4ea3-b52c-d0194f5d6fd9 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 934.745797] env[61923]: DEBUG oslo_vmware.api [None req-29a01eb2-e144-458f-9b46-c49e64b03e88 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Task: {'id': task-1377831, 'name': CloneVM_Task} progress is 95%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.777360] env[61923]: DEBUG nova.network.neutron [-] [instance: 87ae37dc-d0d0-4b76-afdd-0ba3e8f6ce0b] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 934.808711] env[61923]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-219bfcd6-5a1a-4031-8a63-4c173a859ee0 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.819368] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ddf0202-5143-4fd6-a8c6-cf094a29e1d5 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.845763] env[61923]: DEBUG nova.compute.manager [req-3a25c583-91d7-4155-87e0-3cb7bbfd2ff3 req-c13d9e3c-8777-40f0-a0e7-7f299dd45fb1 service nova] [instance: 87ae37dc-d0d0-4b76-afdd-0ba3e8f6ce0b] Detach interface failed, port_id=b0e2ee14-b051-47bc-9064-50f09916c0cb, reason: Instance 87ae37dc-d0d0-4b76-afdd-0ba3e8f6ce0b could not be found. {{(pid=61923) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 935.088033] env[61923]: DEBUG nova.scheduler.client.report [None req-a9734b29-9738-4ea3-b52c-d0194f5d6fd9 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 935.246612] env[61923]: DEBUG oslo_vmware.api [None req-29a01eb2-e144-458f-9b46-c49e64b03e88 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Task: {'id': task-1377831, 'name': CloneVM_Task, 'duration_secs': 1.306401} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 935.248059] env[61923]: INFO nova.virt.vmwareapi.vmops [None req-29a01eb2-e144-458f-9b46-c49e64b03e88 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] [instance: a511392e-9ab6-42fb-b07c-f90e9786dcc1] Created linked-clone VM from snapshot [ 935.248059] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-971bc259-3d8b-4fde-907d-3ec9fd626d68 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.255637] env[61923]: DEBUG nova.virt.vmwareapi.images [None req-29a01eb2-e144-458f-9b46-c49e64b03e88 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] [instance: a511392e-9ab6-42fb-b07c-f90e9786dcc1] Uploading image 5f919f0b-0e82-4adc-b49b-362960821be5 {{(pid=61923) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 935.280533] env[61923]: INFO nova.compute.manager [-] [instance: 87ae37dc-d0d0-4b76-afdd-0ba3e8f6ce0b] Took 1.37 seconds to deallocate network for instance. [ 935.282656] env[61923]: DEBUG oslo_vmware.rw_handles [None req-29a01eb2-e144-458f-9b46-c49e64b03e88 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 935.282656] env[61923]: value = "vm-292723" [ 935.282656] env[61923]: _type = "VirtualMachine" [ 935.282656] env[61923]: }. {{(pid=61923) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 935.284404] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-c498a010-5942-438a-b355-fb34fd025169 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.293972] env[61923]: DEBUG oslo_vmware.rw_handles [None req-29a01eb2-e144-458f-9b46-c49e64b03e88 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Lease: (returnval){ [ 935.293972] env[61923]: value = "session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]521482a2-ce63-ac10-4e32-61b0bc4c5b1e" [ 935.293972] env[61923]: _type = "HttpNfcLease" [ 935.293972] env[61923]: } obtained for exporting VM: (result){ [ 935.293972] env[61923]: value = "vm-292723" [ 935.293972] env[61923]: _type = "VirtualMachine" [ 935.293972] env[61923]: }. {{(pid=61923) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 935.294290] env[61923]: DEBUG oslo_vmware.api [None req-29a01eb2-e144-458f-9b46-c49e64b03e88 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Waiting for the lease: (returnval){ [ 935.294290] env[61923]: value = "session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]521482a2-ce63-ac10-4e32-61b0bc4c5b1e" [ 935.294290] env[61923]: _type = "HttpNfcLease" [ 935.294290] env[61923]: } to be ready. {{(pid=61923) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 935.300652] env[61923]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 935.300652] env[61923]: value = "session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]521482a2-ce63-ac10-4e32-61b0bc4c5b1e" [ 935.300652] env[61923]: _type = "HttpNfcLease" [ 935.300652] env[61923]: } is initializing. {{(pid=61923) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 935.593194] env[61923]: DEBUG oslo_concurrency.lockutils [None req-a9734b29-9738-4ea3-b52c-d0194f5d6fd9 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.185s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 935.595286] env[61923]: DEBUG nova.compute.manager [None req-a9734b29-9738-4ea3-b52c-d0194f5d6fd9 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: 02a8f197-80bb-4cee-bdd6-b07705759986] Start building networks asynchronously for instance. {{(pid=61923) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 935.597845] env[61923]: DEBUG oslo_concurrency.lockutils [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 3.021s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 935.597845] env[61923]: DEBUG oslo_concurrency.lockutils [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 935.597845] env[61923]: DEBUG nova.compute.resource_tracker [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61923) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 935.598125] env[61923]: DEBUG oslo_concurrency.lockutils [None req-cc72dab8-aeca-42e3-aa89-07d278f05c01 tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.512s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 935.602018] env[61923]: DEBUG nova.objects.instance [None req-cc72dab8-aeca-42e3-aa89-07d278f05c01 tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Lazy-loading 'resources' on Instance uuid 62c75dfd-6046-4de1-b1e3-13a307af1394 {{(pid=61923) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 935.602018] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6aaa4dc-344c-4595-91aa-392fd413a905 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.613719] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-afbc39eb-1a6d-4758-a7d8-031ad598dbe1 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.629251] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23f8a4ee-d68c-4cb2-a512-33de579a22b8 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.636242] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91d36e23-3a5d-4037-804e-d74cbca2d0fb {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.666711] env[61923]: DEBUG nova.compute.resource_tracker [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=179704MB free_disk=178GB free_vcpus=48 pci_devices=None {{(pid=61923) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 935.666871] env[61923]: DEBUG oslo_concurrency.lockutils [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 935.793861] env[61923]: DEBUG oslo_concurrency.lockutils [None req-b9e3b0d3-c70e-4ebe-8e35-3d6fba68a21e tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 935.804420] env[61923]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 935.804420] env[61923]: value = "session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]521482a2-ce63-ac10-4e32-61b0bc4c5b1e" [ 935.804420] env[61923]: _type = "HttpNfcLease" [ 935.804420] env[61923]: } is ready. {{(pid=61923) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 935.804420] env[61923]: DEBUG oslo_vmware.rw_handles [None req-29a01eb2-e144-458f-9b46-c49e64b03e88 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 935.804420] env[61923]: value = "session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]521482a2-ce63-ac10-4e32-61b0bc4c5b1e" [ 935.804420] env[61923]: _type = "HttpNfcLease" [ 935.804420] env[61923]: }. {{(pid=61923) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 935.804420] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-509ba494-4d4a-416e-b96e-a25890ece8b0 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.813087] env[61923]: DEBUG oslo_vmware.rw_handles [None req-29a01eb2-e144-458f-9b46-c49e64b03e88 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52dc3e96-f966-08a5-32b5-47b9bb2d06c7/disk-0.vmdk from lease info. {{(pid=61923) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 935.813780] env[61923]: DEBUG oslo_vmware.rw_handles [None req-29a01eb2-e144-458f-9b46-c49e64b03e88 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52dc3e96-f966-08a5-32b5-47b9bb2d06c7/disk-0.vmdk for reading. {{(pid=61923) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 935.923444] env[61923]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-f36fdcca-74d9-4503-8120-3311aa38ad4d {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.103256] env[61923]: DEBUG nova.compute.utils [None req-a9734b29-9738-4ea3-b52c-d0194f5d6fd9 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Using /dev/sd instead of None {{(pid=61923) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 936.104834] env[61923]: DEBUG nova.compute.manager [None req-a9734b29-9738-4ea3-b52c-d0194f5d6fd9 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: 02a8f197-80bb-4cee-bdd6-b07705759986] Allocating IP information in the background. {{(pid=61923) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 936.104931] env[61923]: DEBUG nova.network.neutron [None req-a9734b29-9738-4ea3-b52c-d0194f5d6fd9 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: 02a8f197-80bb-4cee-bdd6-b07705759986] allocate_for_instance() {{(pid=61923) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 936.161079] env[61923]: DEBUG nova.policy [None req-a9734b29-9738-4ea3-b52c-d0194f5d6fd9 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5b8a639e99914e75857b4571f0d58a3c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '69cc941eb6dd4780ac12aa29656c37f7', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61923) authorize /opt/stack/nova/nova/policy.py:201}} [ 936.215746] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09ed49bf-1f4c-48a8-801c-55549d3eb943 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.225103] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff6de4e7-bbc7-4231-9404-eeb69fbbc3c0 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.258661] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7858e287-5be4-443e-960d-10419f14e8c5 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.268313] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e31accd8-71f3-48d9-8c96-b93d4cf092ab {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.283025] env[61923]: DEBUG nova.compute.provider_tree [None req-cc72dab8-aeca-42e3-aa89-07d278f05c01 tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 936.368775] env[61923]: DEBUG oslo_concurrency.lockutils [None req-102bab57-0500-4a8d-94a7-052c22c7c916 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Acquiring lock "7867c6cf-2ad6-414f-ab52-581827860836" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 936.369227] env[61923]: DEBUG oslo_concurrency.lockutils [None req-102bab57-0500-4a8d-94a7-052c22c7c916 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Lock "7867c6cf-2ad6-414f-ab52-581827860836" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 936.443436] env[61923]: DEBUG nova.network.neutron [None req-a9734b29-9738-4ea3-b52c-d0194f5d6fd9 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: 02a8f197-80bb-4cee-bdd6-b07705759986] Successfully created port: 0b0df455-855c-4e3f-832c-4b1c688aa252 {{(pid=61923) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 936.609847] env[61923]: DEBUG nova.compute.manager [None req-a9734b29-9738-4ea3-b52c-d0194f5d6fd9 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: 02a8f197-80bb-4cee-bdd6-b07705759986] Start building block device mappings for instance. {{(pid=61923) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 936.787105] env[61923]: DEBUG nova.scheduler.client.report [None req-cc72dab8-aeca-42e3-aa89-07d278f05c01 tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 936.873018] env[61923]: DEBUG nova.compute.manager [None req-102bab57-0500-4a8d-94a7-052c22c7c916 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: 7867c6cf-2ad6-414f-ab52-581827860836] Starting instance... {{(pid=61923) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 937.294886] env[61923]: DEBUG oslo_concurrency.lockutils [None req-cc72dab8-aeca-42e3-aa89-07d278f05c01 tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.697s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 937.297433] env[61923]: DEBUG oslo_concurrency.lockutils [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 1.630s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 937.320828] env[61923]: INFO nova.scheduler.client.report [None req-cc72dab8-aeca-42e3-aa89-07d278f05c01 tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Deleted allocations for instance 62c75dfd-6046-4de1-b1e3-13a307af1394 [ 937.394723] env[61923]: DEBUG oslo_concurrency.lockutils [None req-102bab57-0500-4a8d-94a7-052c22c7c916 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 937.620829] env[61923]: DEBUG nova.compute.manager [None req-a9734b29-9738-4ea3-b52c-d0194f5d6fd9 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: 02a8f197-80bb-4cee-bdd6-b07705759986] Start spawning the instance on the hypervisor. {{(pid=61923) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 937.645295] env[61923]: DEBUG nova.virt.hardware [None req-a9734b29-9738-4ea3-b52c-d0194f5d6fd9 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-29T20:07:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-29T20:07:35Z,direct_url=,disk_format='vmdk',id=0f153f63-ae0a-45b1-b7f5-7f9b673c947f,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='4e7b5e3b3c3443c8bd5c70f8a15739f5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-29T20:07:36Z,virtual_size=,visibility=), allow threads: False {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 937.645659] env[61923]: DEBUG nova.virt.hardware [None req-a9734b29-9738-4ea3-b52c-d0194f5d6fd9 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Flavor limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 937.645905] env[61923]: DEBUG nova.virt.hardware [None req-a9734b29-9738-4ea3-b52c-d0194f5d6fd9 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Image limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 937.646258] env[61923]: DEBUG nova.virt.hardware [None req-a9734b29-9738-4ea3-b52c-d0194f5d6fd9 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Flavor pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 937.646469] env[61923]: DEBUG nova.virt.hardware [None req-a9734b29-9738-4ea3-b52c-d0194f5d6fd9 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Image pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 937.646648] env[61923]: DEBUG nova.virt.hardware [None req-a9734b29-9738-4ea3-b52c-d0194f5d6fd9 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 937.646875] env[61923]: DEBUG nova.virt.hardware [None req-a9734b29-9738-4ea3-b52c-d0194f5d6fd9 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 937.647153] env[61923]: DEBUG nova.virt.hardware [None req-a9734b29-9738-4ea3-b52c-d0194f5d6fd9 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 937.647422] env[61923]: DEBUG nova.virt.hardware [None req-a9734b29-9738-4ea3-b52c-d0194f5d6fd9 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Got 1 possible topologies {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 937.647599] env[61923]: DEBUG nova.virt.hardware [None req-a9734b29-9738-4ea3-b52c-d0194f5d6fd9 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 937.647774] env[61923]: DEBUG nova.virt.hardware [None req-a9734b29-9738-4ea3-b52c-d0194f5d6fd9 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 937.648681] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f98f79a-bf2a-4a72-aa2a-1dc70e119b20 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.657065] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14fb0628-dd3b-457d-9766-829e8585bf16 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.828216] env[61923]: DEBUG oslo_concurrency.lockutils [None req-cc72dab8-aeca-42e3-aa89-07d278f05c01 tempest-ImagesTestJSON-1951383190 tempest-ImagesTestJSON-1951383190-project-member] Lock "62c75dfd-6046-4de1-b1e3-13a307af1394" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.566s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 938.325634] env[61923]: WARNING nova.compute.resource_tracker [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Instance 87ae37dc-d0d0-4b76-afdd-0ba3e8f6ce0b is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 938.325906] env[61923]: DEBUG nova.compute.resource_tracker [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Instance b779d183-89ae-4e4d-ae99-e514e145ed43 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61923) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 938.326086] env[61923]: DEBUG nova.compute.resource_tracker [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Instance a511392e-9ab6-42fb-b07c-f90e9786dcc1 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61923) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 938.326260] env[61923]: DEBUG nova.compute.resource_tracker [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Instance 02a8f197-80bb-4cee-bdd6-b07705759986 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61923) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 938.364297] env[61923]: DEBUG nova.compute.manager [req-8c17e9ee-9a32-4c45-9903-392004947712 req-a1fb9340-e535-421c-981e-fec2180c0bba service nova] [instance: 02a8f197-80bb-4cee-bdd6-b07705759986] Received event network-vif-plugged-0b0df455-855c-4e3f-832c-4b1c688aa252 {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 938.364538] env[61923]: DEBUG oslo_concurrency.lockutils [req-8c17e9ee-9a32-4c45-9903-392004947712 req-a1fb9340-e535-421c-981e-fec2180c0bba service nova] Acquiring lock "02a8f197-80bb-4cee-bdd6-b07705759986-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 938.364751] env[61923]: DEBUG oslo_concurrency.lockutils [req-8c17e9ee-9a32-4c45-9903-392004947712 req-a1fb9340-e535-421c-981e-fec2180c0bba service nova] Lock "02a8f197-80bb-4cee-bdd6-b07705759986-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 938.364924] env[61923]: DEBUG oslo_concurrency.lockutils [req-8c17e9ee-9a32-4c45-9903-392004947712 req-a1fb9340-e535-421c-981e-fec2180c0bba service nova] Lock "02a8f197-80bb-4cee-bdd6-b07705759986-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 938.365106] env[61923]: DEBUG nova.compute.manager [req-8c17e9ee-9a32-4c45-9903-392004947712 req-a1fb9340-e535-421c-981e-fec2180c0bba service nova] [instance: 02a8f197-80bb-4cee-bdd6-b07705759986] No waiting events found dispatching network-vif-plugged-0b0df455-855c-4e3f-832c-4b1c688aa252 {{(pid=61923) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 938.365324] env[61923]: WARNING nova.compute.manager [req-8c17e9ee-9a32-4c45-9903-392004947712 req-a1fb9340-e535-421c-981e-fec2180c0bba service nova] [instance: 02a8f197-80bb-4cee-bdd6-b07705759986] Received unexpected event network-vif-plugged-0b0df455-855c-4e3f-832c-4b1c688aa252 for instance with vm_state building and task_state spawning. [ 938.554155] env[61923]: DEBUG nova.network.neutron [None req-a9734b29-9738-4ea3-b52c-d0194f5d6fd9 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: 02a8f197-80bb-4cee-bdd6-b07705759986] Successfully updated port: 0b0df455-855c-4e3f-832c-4b1c688aa252 {{(pid=61923) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 938.829244] env[61923]: DEBUG nova.compute.resource_tracker [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Instance 7867c6cf-2ad6-414f-ab52-581827860836 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61923) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 938.829553] env[61923]: DEBUG nova.compute.resource_tracker [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Total usable vcpus: 48, total allocated vcpus: 3 {{(pid=61923) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 938.829769] env[61923]: DEBUG nova.compute.resource_tracker [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1088MB phys_disk=200GB used_disk=3GB total_vcpus=48 used_vcpus=3 pci_stats=[] {{(pid=61923) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 938.917055] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4caf12c-94b7-46d5-a07c-cd1206b6abfb {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.926719] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87d89ebb-989f-427b-9de8-aaf44edf8abf {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.963683] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f828e567-2562-4eab-b0c7-d1637ab5779b {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.972424] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab8e2f30-50f9-4d18-9e06-29db03b11b5b {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.986724] env[61923]: DEBUG nova.compute.provider_tree [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 939.056757] env[61923]: DEBUG oslo_concurrency.lockutils [None req-a9734b29-9738-4ea3-b52c-d0194f5d6fd9 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Acquiring lock "refresh_cache-02a8f197-80bb-4cee-bdd6-b07705759986" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 939.056922] env[61923]: DEBUG oslo_concurrency.lockutils [None req-a9734b29-9738-4ea3-b52c-d0194f5d6fd9 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Acquired lock "refresh_cache-02a8f197-80bb-4cee-bdd6-b07705759986" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 939.057115] env[61923]: DEBUG nova.network.neutron [None req-a9734b29-9738-4ea3-b52c-d0194f5d6fd9 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: 02a8f197-80bb-4cee-bdd6-b07705759986] Building network info cache for instance {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 939.492604] env[61923]: DEBUG nova.scheduler.client.report [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 939.601113] env[61923]: DEBUG nova.network.neutron [None req-a9734b29-9738-4ea3-b52c-d0194f5d6fd9 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: 02a8f197-80bb-4cee-bdd6-b07705759986] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 939.830714] env[61923]: DEBUG nova.network.neutron [None req-a9734b29-9738-4ea3-b52c-d0194f5d6fd9 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: 02a8f197-80bb-4cee-bdd6-b07705759986] Updating instance_info_cache with network_info: [{"id": "0b0df455-855c-4e3f-832c-4b1c688aa252", "address": "fa:16:3e:1a:e5:62", "network": {"id": "d1c38272-0461-43e5-93ae-3964346f77d0", "bridge": "br-int", "label": "tempest-ServersTestJSON-151839165-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "69cc941eb6dd4780ac12aa29656c37f7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8ba3bd22-c936-470e-89bd-b3a5587e87a0", "external-id": "nsx-vlan-transportzone-605", "segmentation_id": 605, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0b0df455-85", "ovs_interfaceid": "0b0df455-855c-4e3f-832c-4b1c688aa252", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 940.000778] env[61923]: DEBUG nova.compute.resource_tracker [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61923) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 940.000957] env[61923]: DEBUG oslo_concurrency.lockutils [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.704s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 940.001252] env[61923]: DEBUG oslo_concurrency.lockutils [None req-b9e3b0d3-c70e-4ebe-8e35-3d6fba68a21e tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 4.208s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 940.001913] env[61923]: DEBUG oslo_concurrency.lockutils [None req-b9e3b0d3-c70e-4ebe-8e35-3d6fba68a21e tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 940.004010] env[61923]: DEBUG oslo_concurrency.lockutils [None req-102bab57-0500-4a8d-94a7-052c22c7c916 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.609s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 940.005500] env[61923]: INFO nova.compute.claims [None req-102bab57-0500-4a8d-94a7-052c22c7c916 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: 7867c6cf-2ad6-414f-ab52-581827860836] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 940.029985] env[61923]: INFO nova.scheduler.client.report [None req-b9e3b0d3-c70e-4ebe-8e35-3d6fba68a21e tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Deleted allocations for instance 87ae37dc-d0d0-4b76-afdd-0ba3e8f6ce0b [ 940.333480] env[61923]: DEBUG oslo_concurrency.lockutils [None req-a9734b29-9738-4ea3-b52c-d0194f5d6fd9 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Releasing lock "refresh_cache-02a8f197-80bb-4cee-bdd6-b07705759986" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 940.334510] env[61923]: DEBUG nova.compute.manager [None req-a9734b29-9738-4ea3-b52c-d0194f5d6fd9 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: 02a8f197-80bb-4cee-bdd6-b07705759986] Instance network_info: |[{"id": "0b0df455-855c-4e3f-832c-4b1c688aa252", "address": "fa:16:3e:1a:e5:62", "network": {"id": "d1c38272-0461-43e5-93ae-3964346f77d0", "bridge": "br-int", "label": "tempest-ServersTestJSON-151839165-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "69cc941eb6dd4780ac12aa29656c37f7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8ba3bd22-c936-470e-89bd-b3a5587e87a0", "external-id": "nsx-vlan-transportzone-605", "segmentation_id": 605, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0b0df455-85", "ovs_interfaceid": "0b0df455-855c-4e3f-832c-4b1c688aa252", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61923) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 940.334510] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-a9734b29-9738-4ea3-b52c-d0194f5d6fd9 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: 02a8f197-80bb-4cee-bdd6-b07705759986] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:1a:e5:62', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '8ba3bd22-c936-470e-89bd-b3a5587e87a0', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '0b0df455-855c-4e3f-832c-4b1c688aa252', 'vif_model': 'vmxnet3'}] {{(pid=61923) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 940.342779] env[61923]: DEBUG oslo.service.loopingcall [None req-a9734b29-9738-4ea3-b52c-d0194f5d6fd9 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61923) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 940.343246] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 02a8f197-80bb-4cee-bdd6-b07705759986] Creating VM on the ESX host {{(pid=61923) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 940.343384] env[61923]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-097ecc36-1964-4156-92c5-0fc80357daa4 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.364477] env[61923]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 940.364477] env[61923]: value = "task-1377833" [ 940.364477] env[61923]: _type = "Task" [ 940.364477] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 940.373651] env[61923]: DEBUG oslo_vmware.api [-] Task: {'id': task-1377833, 'name': CreateVM_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.395555] env[61923]: DEBUG nova.compute.manager [req-b8629f23-64f5-4bce-9e5b-6de1be11c291 req-8e0145f1-d136-4f8c-b37a-53515d1f218d service nova] [instance: 02a8f197-80bb-4cee-bdd6-b07705759986] Received event network-changed-0b0df455-855c-4e3f-832c-4b1c688aa252 {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 940.395684] env[61923]: DEBUG nova.compute.manager [req-b8629f23-64f5-4bce-9e5b-6de1be11c291 req-8e0145f1-d136-4f8c-b37a-53515d1f218d service nova] [instance: 02a8f197-80bb-4cee-bdd6-b07705759986] Refreshing instance network info cache due to event network-changed-0b0df455-855c-4e3f-832c-4b1c688aa252. {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 940.395914] env[61923]: DEBUG oslo_concurrency.lockutils [req-b8629f23-64f5-4bce-9e5b-6de1be11c291 req-8e0145f1-d136-4f8c-b37a-53515d1f218d service nova] Acquiring lock "refresh_cache-02a8f197-80bb-4cee-bdd6-b07705759986" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 940.396076] env[61923]: DEBUG oslo_concurrency.lockutils [req-b8629f23-64f5-4bce-9e5b-6de1be11c291 req-8e0145f1-d136-4f8c-b37a-53515d1f218d service nova] Acquired lock "refresh_cache-02a8f197-80bb-4cee-bdd6-b07705759986" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 940.396242] env[61923]: DEBUG nova.network.neutron [req-b8629f23-64f5-4bce-9e5b-6de1be11c291 req-8e0145f1-d136-4f8c-b37a-53515d1f218d service nova] [instance: 02a8f197-80bb-4cee-bdd6-b07705759986] Refreshing network info cache for port 0b0df455-855c-4e3f-832c-4b1c688aa252 {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 940.540528] env[61923]: DEBUG oslo_concurrency.lockutils [None req-b9e3b0d3-c70e-4ebe-8e35-3d6fba68a21e tempest-ListImageFiltersTestJSON-192383749 tempest-ListImageFiltersTestJSON-192383749-project-member] Lock "87ae37dc-d0d0-4b76-afdd-0ba3e8f6ce0b" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.751s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 940.874576] env[61923]: DEBUG oslo_vmware.api [-] Task: {'id': task-1377833, 'name': CreateVM_Task, 'duration_secs': 0.336027} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 940.874771] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 02a8f197-80bb-4cee-bdd6-b07705759986] Created VM on the ESX host {{(pid=61923) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 940.875492] env[61923]: DEBUG oslo_concurrency.lockutils [None req-a9734b29-9738-4ea3-b52c-d0194f5d6fd9 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 940.875668] env[61923]: DEBUG oslo_concurrency.lockutils [None req-a9734b29-9738-4ea3-b52c-d0194f5d6fd9 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 940.876116] env[61923]: DEBUG oslo_concurrency.lockutils [None req-a9734b29-9738-4ea3-b52c-d0194f5d6fd9 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 940.876380] env[61923]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b9c03e4a-a669-46c7-967a-433b2962d744 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.881193] env[61923]: DEBUG oslo_vmware.api [None req-a9734b29-9738-4ea3-b52c-d0194f5d6fd9 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Waiting for the task: (returnval){ [ 940.881193] env[61923]: value = "session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52936194-d91a-5ca1-b796-9c752c5d3850" [ 940.881193] env[61923]: _type = "Task" [ 940.881193] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 940.890379] env[61923]: DEBUG oslo_vmware.api [None req-a9734b29-9738-4ea3-b52c-d0194f5d6fd9 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52936194-d91a-5ca1-b796-9c752c5d3850, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.099692] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47fbb803-082c-434b-9997-d937039c0a0f {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.108624] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad70b71e-8343-49f8-befe-d2496a08c382 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.139207] env[61923]: DEBUG nova.network.neutron [req-b8629f23-64f5-4bce-9e5b-6de1be11c291 req-8e0145f1-d136-4f8c-b37a-53515d1f218d service nova] [instance: 02a8f197-80bb-4cee-bdd6-b07705759986] Updated VIF entry in instance network info cache for port 0b0df455-855c-4e3f-832c-4b1c688aa252. {{(pid=61923) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 941.139597] env[61923]: DEBUG nova.network.neutron [req-b8629f23-64f5-4bce-9e5b-6de1be11c291 req-8e0145f1-d136-4f8c-b37a-53515d1f218d service nova] [instance: 02a8f197-80bb-4cee-bdd6-b07705759986] Updating instance_info_cache with network_info: [{"id": "0b0df455-855c-4e3f-832c-4b1c688aa252", "address": "fa:16:3e:1a:e5:62", "network": {"id": "d1c38272-0461-43e5-93ae-3964346f77d0", "bridge": "br-int", "label": "tempest-ServersTestJSON-151839165-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "69cc941eb6dd4780ac12aa29656c37f7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8ba3bd22-c936-470e-89bd-b3a5587e87a0", "external-id": "nsx-vlan-transportzone-605", "segmentation_id": 605, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0b0df455-85", "ovs_interfaceid": "0b0df455-855c-4e3f-832c-4b1c688aa252", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 941.141365] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-338e54d0-e971-464b-8aa8-f93aa34788dc {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.149480] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa1f943a-5570-44bf-bc59-f40cddc3498d {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.164662] env[61923]: DEBUG nova.compute.provider_tree [None req-102bab57-0500-4a8d-94a7-052c22c7c916 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 941.391658] env[61923]: DEBUG oslo_vmware.api [None req-a9734b29-9738-4ea3-b52c-d0194f5d6fd9 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52936194-d91a-5ca1-b796-9c752c5d3850, 'name': SearchDatastore_Task, 'duration_secs': 0.017835} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 941.391975] env[61923]: DEBUG oslo_concurrency.lockutils [None req-a9734b29-9738-4ea3-b52c-d0194f5d6fd9 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 941.393087] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-a9734b29-9738-4ea3-b52c-d0194f5d6fd9 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: 02a8f197-80bb-4cee-bdd6-b07705759986] Processing image 0f153f63-ae0a-45b1-b7f5-7f9b673c947f {{(pid=61923) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 941.393087] env[61923]: DEBUG oslo_concurrency.lockutils [None req-a9734b29-9738-4ea3-b52c-d0194f5d6fd9 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 941.393087] env[61923]: DEBUG oslo_concurrency.lockutils [None req-a9734b29-9738-4ea3-b52c-d0194f5d6fd9 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 941.393087] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-a9734b29-9738-4ea3-b52c-d0194f5d6fd9 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61923) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 941.393325] env[61923]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-9a3e049f-6fe4-414d-ade0-9c01e17db584 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.404259] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-a9734b29-9738-4ea3-b52c-d0194f5d6fd9 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61923) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 941.404259] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-a9734b29-9738-4ea3-b52c-d0194f5d6fd9 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61923) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 941.404259] env[61923]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a7e58437-7a0c-4891-a7a9-c986a3438079 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.408955] env[61923]: DEBUG oslo_vmware.api [None req-a9734b29-9738-4ea3-b52c-d0194f5d6fd9 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Waiting for the task: (returnval){ [ 941.408955] env[61923]: value = "session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52a48fb3-f767-e004-8fef-8dc4fd44fcb5" [ 941.408955] env[61923]: _type = "Task" [ 941.408955] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 941.418281] env[61923]: DEBUG oslo_vmware.api [None req-a9734b29-9738-4ea3-b52c-d0194f5d6fd9 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52a48fb3-f767-e004-8fef-8dc4fd44fcb5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.645660] env[61923]: DEBUG oslo_concurrency.lockutils [req-b8629f23-64f5-4bce-9e5b-6de1be11c291 req-8e0145f1-d136-4f8c-b37a-53515d1f218d service nova] Releasing lock "refresh_cache-02a8f197-80bb-4cee-bdd6-b07705759986" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 941.669304] env[61923]: DEBUG nova.scheduler.client.report [None req-102bab57-0500-4a8d-94a7-052c22c7c916 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 941.921267] env[61923]: DEBUG oslo_vmware.api [None req-a9734b29-9738-4ea3-b52c-d0194f5d6fd9 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52a48fb3-f767-e004-8fef-8dc4fd44fcb5, 'name': SearchDatastore_Task, 'duration_secs': 0.021499} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 941.922103] env[61923]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e7ed3e9e-35c2-46de-bb42-bdaf6105b7f6 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.927758] env[61923]: DEBUG oslo_vmware.api [None req-a9734b29-9738-4ea3-b52c-d0194f5d6fd9 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Waiting for the task: (returnval){ [ 941.927758] env[61923]: value = "session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52b94dec-6678-9e7c-762d-91fa69327df0" [ 941.927758] env[61923]: _type = "Task" [ 941.927758] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 941.936353] env[61923]: DEBUG oslo_vmware.api [None req-a9734b29-9738-4ea3-b52c-d0194f5d6fd9 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52b94dec-6678-9e7c-762d-91fa69327df0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.969818] env[61923]: DEBUG oslo_concurrency.lockutils [None req-69e9bd4b-8d1d-4678-80b0-4de2eeffdae8 tempest-ServersAaction247Test-1300676990 tempest-ServersAaction247Test-1300676990-project-member] Acquiring lock "44196b4c-1401-40fa-bd14-04a49947ab15" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 941.970100] env[61923]: DEBUG oslo_concurrency.lockutils [None req-69e9bd4b-8d1d-4678-80b0-4de2eeffdae8 tempest-ServersAaction247Test-1300676990 tempest-ServersAaction247Test-1300676990-project-member] Lock "44196b4c-1401-40fa-bd14-04a49947ab15" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 942.174570] env[61923]: DEBUG oslo_concurrency.lockutils [None req-102bab57-0500-4a8d-94a7-052c22c7c916 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.170s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 942.175208] env[61923]: DEBUG nova.compute.manager [None req-102bab57-0500-4a8d-94a7-052c22c7c916 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: 7867c6cf-2ad6-414f-ab52-581827860836] Start building networks asynchronously for instance. {{(pid=61923) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 942.445305] env[61923]: DEBUG oslo_vmware.api [None req-a9734b29-9738-4ea3-b52c-d0194f5d6fd9 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52b94dec-6678-9e7c-762d-91fa69327df0, 'name': SearchDatastore_Task, 'duration_secs': 0.010613} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 942.445775] env[61923]: DEBUG oslo_concurrency.lockutils [None req-a9734b29-9738-4ea3-b52c-d0194f5d6fd9 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 942.445944] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-a9734b29-9738-4ea3-b52c-d0194f5d6fd9 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk to [datastore1] 02a8f197-80bb-4cee-bdd6-b07705759986/02a8f197-80bb-4cee-bdd6-b07705759986.vmdk {{(pid=61923) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 942.446255] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-56d3f5b1-3a20-45ac-8e29-f3acd6f16699 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.453798] env[61923]: DEBUG oslo_vmware.api [None req-a9734b29-9738-4ea3-b52c-d0194f5d6fd9 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Waiting for the task: (returnval){ [ 942.453798] env[61923]: value = "task-1377834" [ 942.453798] env[61923]: _type = "Task" [ 942.453798] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 942.462855] env[61923]: DEBUG oslo_vmware.api [None req-a9734b29-9738-4ea3-b52c-d0194f5d6fd9 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Task: {'id': task-1377834, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.472538] env[61923]: DEBUG nova.compute.manager [None req-69e9bd4b-8d1d-4678-80b0-4de2eeffdae8 tempest-ServersAaction247Test-1300676990 tempest-ServersAaction247Test-1300676990-project-member] [instance: 44196b4c-1401-40fa-bd14-04a49947ab15] Starting instance... {{(pid=61923) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 942.683026] env[61923]: DEBUG nova.compute.utils [None req-102bab57-0500-4a8d-94a7-052c22c7c916 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Using /dev/sd instead of None {{(pid=61923) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 942.683518] env[61923]: DEBUG nova.compute.manager [None req-102bab57-0500-4a8d-94a7-052c22c7c916 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: 7867c6cf-2ad6-414f-ab52-581827860836] Allocating IP information in the background. {{(pid=61923) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 942.683737] env[61923]: DEBUG nova.network.neutron [None req-102bab57-0500-4a8d-94a7-052c22c7c916 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: 7867c6cf-2ad6-414f-ab52-581827860836] allocate_for_instance() {{(pid=61923) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 942.787985] env[61923]: DEBUG nova.policy [None req-102bab57-0500-4a8d-94a7-052c22c7c916 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '69be89006ff141b686261d643b3b9a73', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '780d3541d9604417b977bb62390c4299', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61923) authorize /opt/stack/nova/nova/policy.py:201}} [ 942.963827] env[61923]: DEBUG oslo_vmware.api [None req-a9734b29-9738-4ea3-b52c-d0194f5d6fd9 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Task: {'id': task-1377834, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.48813} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 942.964341] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-a9734b29-9738-4ea3-b52c-d0194f5d6fd9 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk to [datastore1] 02a8f197-80bb-4cee-bdd6-b07705759986/02a8f197-80bb-4cee-bdd6-b07705759986.vmdk {{(pid=61923) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 942.964341] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-a9734b29-9738-4ea3-b52c-d0194f5d6fd9 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: 02a8f197-80bb-4cee-bdd6-b07705759986] Extending root virtual disk to 1048576 {{(pid=61923) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 942.966306] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-dc97a38d-6e70-42ff-9692-28077fd44120 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.972226] env[61923]: DEBUG oslo_vmware.api [None req-a9734b29-9738-4ea3-b52c-d0194f5d6fd9 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Waiting for the task: (returnval){ [ 942.972226] env[61923]: value = "task-1377835" [ 942.972226] env[61923]: _type = "Task" [ 942.972226] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 942.977874] env[61923]: DEBUG oslo_vmware.rw_handles [None req-29a01eb2-e144-458f-9b46-c49e64b03e88 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52dc3e96-f966-08a5-32b5-47b9bb2d06c7/disk-0.vmdk. {{(pid=61923) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 942.978744] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15b3e941-8804-45c1-a632-404716ca1ec5 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.987564] env[61923]: DEBUG oslo_vmware.api [None req-a9734b29-9738-4ea3-b52c-d0194f5d6fd9 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Task: {'id': task-1377835, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.991472] env[61923]: DEBUG oslo_vmware.rw_handles [None req-29a01eb2-e144-458f-9b46-c49e64b03e88 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52dc3e96-f966-08a5-32b5-47b9bb2d06c7/disk-0.vmdk is in state: ready. {{(pid=61923) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 942.991644] env[61923]: ERROR oslo_vmware.rw_handles [None req-29a01eb2-e144-458f-9b46-c49e64b03e88 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52dc3e96-f966-08a5-32b5-47b9bb2d06c7/disk-0.vmdk due to incomplete transfer. [ 942.991866] env[61923]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-718b26cc-5d19-4927-8498-991c166b95aa {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.997770] env[61923]: DEBUG oslo_vmware.rw_handles [None req-29a01eb2-e144-458f-9b46-c49e64b03e88 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52dc3e96-f966-08a5-32b5-47b9bb2d06c7/disk-0.vmdk. {{(pid=61923) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 942.997980] env[61923]: DEBUG nova.virt.vmwareapi.images [None req-29a01eb2-e144-458f-9b46-c49e64b03e88 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] [instance: a511392e-9ab6-42fb-b07c-f90e9786dcc1] Uploaded image 5f919f0b-0e82-4adc-b49b-362960821be5 to the Glance image server {{(pid=61923) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 943.000237] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-29a01eb2-e144-458f-9b46-c49e64b03e88 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] [instance: a511392e-9ab6-42fb-b07c-f90e9786dcc1] Destroying the VM {{(pid=61923) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 943.000802] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-21fb26f9-4e90-4dcc-b909-3c93a574179e {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.004902] env[61923]: DEBUG oslo_concurrency.lockutils [None req-69e9bd4b-8d1d-4678-80b0-4de2eeffdae8 tempest-ServersAaction247Test-1300676990 tempest-ServersAaction247Test-1300676990-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 943.005188] env[61923]: DEBUG oslo_concurrency.lockutils [None req-69e9bd4b-8d1d-4678-80b0-4de2eeffdae8 tempest-ServersAaction247Test-1300676990 tempest-ServersAaction247Test-1300676990-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 943.007130] env[61923]: INFO nova.compute.claims [None req-69e9bd4b-8d1d-4678-80b0-4de2eeffdae8 tempest-ServersAaction247Test-1300676990 tempest-ServersAaction247Test-1300676990-project-member] [instance: 44196b4c-1401-40fa-bd14-04a49947ab15] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 943.010720] env[61923]: DEBUG oslo_vmware.api [None req-29a01eb2-e144-458f-9b46-c49e64b03e88 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Waiting for the task: (returnval){ [ 943.010720] env[61923]: value = "task-1377836" [ 943.010720] env[61923]: _type = "Task" [ 943.010720] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 943.018334] env[61923]: DEBUG oslo_vmware.api [None req-29a01eb2-e144-458f-9b46-c49e64b03e88 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Task: {'id': task-1377836, 'name': Destroy_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.093123] env[61923]: DEBUG nova.network.neutron [None req-102bab57-0500-4a8d-94a7-052c22c7c916 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: 7867c6cf-2ad6-414f-ab52-581827860836] Successfully created port: 972754b8-9333-4b89-8435-ca8a98aa7f39 {{(pid=61923) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 943.184707] env[61923]: DEBUG nova.compute.manager [None req-102bab57-0500-4a8d-94a7-052c22c7c916 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: 7867c6cf-2ad6-414f-ab52-581827860836] Start building block device mappings for instance. {{(pid=61923) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 943.482475] env[61923]: DEBUG oslo_vmware.api [None req-a9734b29-9738-4ea3-b52c-d0194f5d6fd9 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Task: {'id': task-1377835, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.080028} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 943.482781] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-a9734b29-9738-4ea3-b52c-d0194f5d6fd9 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: 02a8f197-80bb-4cee-bdd6-b07705759986] Extended root virtual disk {{(pid=61923) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 943.483566] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f661a216-aee1-4872-a246-f7ed635d0537 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.510479] env[61923]: DEBUG nova.virt.vmwareapi.volumeops [None req-a9734b29-9738-4ea3-b52c-d0194f5d6fd9 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: 02a8f197-80bb-4cee-bdd6-b07705759986] Reconfiguring VM instance instance-00000058 to attach disk [datastore1] 02a8f197-80bb-4cee-bdd6-b07705759986/02a8f197-80bb-4cee-bdd6-b07705759986.vmdk or device None with type sparse {{(pid=61923) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 943.510815] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9a6465d4-3414-4fef-9ffc-ccd23616bb48 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.539598] env[61923]: DEBUG oslo_vmware.api [None req-29a01eb2-e144-458f-9b46-c49e64b03e88 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Task: {'id': task-1377836, 'name': Destroy_Task, 'duration_secs': 0.308143} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 943.541076] env[61923]: INFO nova.virt.vmwareapi.vm_util [None req-29a01eb2-e144-458f-9b46-c49e64b03e88 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] [instance: a511392e-9ab6-42fb-b07c-f90e9786dcc1] Destroyed the VM [ 943.541424] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-29a01eb2-e144-458f-9b46-c49e64b03e88 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] [instance: a511392e-9ab6-42fb-b07c-f90e9786dcc1] Deleting Snapshot of the VM instance {{(pid=61923) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 943.541808] env[61923]: DEBUG oslo_vmware.api [None req-a9734b29-9738-4ea3-b52c-d0194f5d6fd9 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Waiting for the task: (returnval){ [ 943.541808] env[61923]: value = "task-1377837" [ 943.541808] env[61923]: _type = "Task" [ 943.541808] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 943.542086] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-7c7624a7-71ec-4323-b4a4-ae1fd0c50610 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.552786] env[61923]: DEBUG oslo_vmware.api [None req-29a01eb2-e144-458f-9b46-c49e64b03e88 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Waiting for the task: (returnval){ [ 943.552786] env[61923]: value = "task-1377838" [ 943.552786] env[61923]: _type = "Task" [ 943.552786] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 943.556360] env[61923]: DEBUG oslo_vmware.api [None req-a9734b29-9738-4ea3-b52c-d0194f5d6fd9 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Task: {'id': task-1377837, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.567033] env[61923]: DEBUG oslo_vmware.api [None req-29a01eb2-e144-458f-9b46-c49e64b03e88 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Task: {'id': task-1377838, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.056971] env[61923]: DEBUG oslo_vmware.api [None req-a9734b29-9738-4ea3-b52c-d0194f5d6fd9 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Task: {'id': task-1377837, 'name': ReconfigVM_Task, 'duration_secs': 0.318857} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 944.060073] env[61923]: DEBUG nova.virt.vmwareapi.volumeops [None req-a9734b29-9738-4ea3-b52c-d0194f5d6fd9 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: 02a8f197-80bb-4cee-bdd6-b07705759986] Reconfigured VM instance instance-00000058 to attach disk [datastore1] 02a8f197-80bb-4cee-bdd6-b07705759986/02a8f197-80bb-4cee-bdd6-b07705759986.vmdk or device None with type sparse {{(pid=61923) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 944.060750] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f1eebbe8-8554-4ffc-8a7d-934c5406063c {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.068772] env[61923]: DEBUG oslo_vmware.api [None req-29a01eb2-e144-458f-9b46-c49e64b03e88 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Task: {'id': task-1377838, 'name': RemoveSnapshot_Task, 'duration_secs': 0.377661} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 944.069945] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-29a01eb2-e144-458f-9b46-c49e64b03e88 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] [instance: a511392e-9ab6-42fb-b07c-f90e9786dcc1] Deleted Snapshot of the VM instance {{(pid=61923) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 944.070262] env[61923]: DEBUG nova.compute.manager [None req-29a01eb2-e144-458f-9b46-c49e64b03e88 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] [instance: a511392e-9ab6-42fb-b07c-f90e9786dcc1] Checking state {{(pid=61923) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 944.070609] env[61923]: DEBUG oslo_vmware.api [None req-a9734b29-9738-4ea3-b52c-d0194f5d6fd9 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Waiting for the task: (returnval){ [ 944.070609] env[61923]: value = "task-1377839" [ 944.070609] env[61923]: _type = "Task" [ 944.070609] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 944.071305] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d3706c6-1fa6-419d-a370-5ca333a5be1f {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.087921] env[61923]: DEBUG oslo_vmware.api [None req-a9734b29-9738-4ea3-b52c-d0194f5d6fd9 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Task: {'id': task-1377839, 'name': Rename_Task} progress is 10%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.124388] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b2f96fd-8db2-42a4-a792-bcd0fe2cb413 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.132842] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a32312c-f13d-4041-9c62-b3455641bc66 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.163917] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97c248e7-df5c-4aad-9747-c8caba5eec16 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.171758] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-273e0e11-b35c-420b-ace5-90c3b5547216 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.185370] env[61923]: DEBUG nova.compute.provider_tree [None req-69e9bd4b-8d1d-4678-80b0-4de2eeffdae8 tempest-ServersAaction247Test-1300676990 tempest-ServersAaction247Test-1300676990-project-member] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 944.194182] env[61923]: DEBUG nova.compute.manager [None req-102bab57-0500-4a8d-94a7-052c22c7c916 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: 7867c6cf-2ad6-414f-ab52-581827860836] Start spawning the instance on the hypervisor. {{(pid=61923) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 944.220489] env[61923]: DEBUG nova.virt.hardware [None req-102bab57-0500-4a8d-94a7-052c22c7c916 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-29T20:07:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-29T20:07:35Z,direct_url=,disk_format='vmdk',id=0f153f63-ae0a-45b1-b7f5-7f9b673c947f,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='4e7b5e3b3c3443c8bd5c70f8a15739f5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-29T20:07:36Z,virtual_size=,visibility=), allow threads: False {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 944.220789] env[61923]: DEBUG nova.virt.hardware [None req-102bab57-0500-4a8d-94a7-052c22c7c916 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Flavor limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 944.220960] env[61923]: DEBUG nova.virt.hardware [None req-102bab57-0500-4a8d-94a7-052c22c7c916 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Image limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 944.221161] env[61923]: DEBUG nova.virt.hardware [None req-102bab57-0500-4a8d-94a7-052c22c7c916 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Flavor pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 944.221310] env[61923]: DEBUG nova.virt.hardware [None req-102bab57-0500-4a8d-94a7-052c22c7c916 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Image pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 944.221462] env[61923]: DEBUG nova.virt.hardware [None req-102bab57-0500-4a8d-94a7-052c22c7c916 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 944.221746] env[61923]: DEBUG nova.virt.hardware [None req-102bab57-0500-4a8d-94a7-052c22c7c916 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 944.221916] env[61923]: DEBUG nova.virt.hardware [None req-102bab57-0500-4a8d-94a7-052c22c7c916 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 944.222128] env[61923]: DEBUG nova.virt.hardware [None req-102bab57-0500-4a8d-94a7-052c22c7c916 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Got 1 possible topologies {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 944.222346] env[61923]: DEBUG nova.virt.hardware [None req-102bab57-0500-4a8d-94a7-052c22c7c916 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 944.222528] env[61923]: DEBUG nova.virt.hardware [None req-102bab57-0500-4a8d-94a7-052c22c7c916 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 944.223401] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd9e13a7-3cf1-4a90-9c3a-589993bca1ca {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.232081] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c649549-5b6a-4101-9ab7-1757e0db4fe2 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.329691] env[61923]: DEBUG oslo_concurrency.lockutils [None req-704a26eb-9a13-4b75-8212-d4cbe21aa5e4 tempest-ServersV294TestFqdnHostnames-833840938 tempest-ServersV294TestFqdnHostnames-833840938-project-member] Acquiring lock "aba844c1-d488-4e95-a550-bd1cb7331fb2" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 944.329954] env[61923]: DEBUG oslo_concurrency.lockutils [None req-704a26eb-9a13-4b75-8212-d4cbe21aa5e4 tempest-ServersV294TestFqdnHostnames-833840938 tempest-ServersV294TestFqdnHostnames-833840938-project-member] Lock "aba844c1-d488-4e95-a550-bd1cb7331fb2" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 944.541604] env[61923]: DEBUG nova.compute.manager [req-e9caaaa3-40a6-44d4-b747-72c4d75897ea req-7f443671-0902-4dde-ac9b-01db834c9d91 service nova] [instance: 7867c6cf-2ad6-414f-ab52-581827860836] Received event network-vif-plugged-972754b8-9333-4b89-8435-ca8a98aa7f39 {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 944.541949] env[61923]: DEBUG oslo_concurrency.lockutils [req-e9caaaa3-40a6-44d4-b747-72c4d75897ea req-7f443671-0902-4dde-ac9b-01db834c9d91 service nova] Acquiring lock "7867c6cf-2ad6-414f-ab52-581827860836-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 944.542303] env[61923]: DEBUG oslo_concurrency.lockutils [req-e9caaaa3-40a6-44d4-b747-72c4d75897ea req-7f443671-0902-4dde-ac9b-01db834c9d91 service nova] Lock "7867c6cf-2ad6-414f-ab52-581827860836-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 944.542567] env[61923]: DEBUG oslo_concurrency.lockutils [req-e9caaaa3-40a6-44d4-b747-72c4d75897ea req-7f443671-0902-4dde-ac9b-01db834c9d91 service nova] Lock "7867c6cf-2ad6-414f-ab52-581827860836-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 944.542832] env[61923]: DEBUG nova.compute.manager [req-e9caaaa3-40a6-44d4-b747-72c4d75897ea req-7f443671-0902-4dde-ac9b-01db834c9d91 service nova] [instance: 7867c6cf-2ad6-414f-ab52-581827860836] No waiting events found dispatching network-vif-plugged-972754b8-9333-4b89-8435-ca8a98aa7f39 {{(pid=61923) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 944.543113] env[61923]: WARNING nova.compute.manager [req-e9caaaa3-40a6-44d4-b747-72c4d75897ea req-7f443671-0902-4dde-ac9b-01db834c9d91 service nova] [instance: 7867c6cf-2ad6-414f-ab52-581827860836] Received unexpected event network-vif-plugged-972754b8-9333-4b89-8435-ca8a98aa7f39 for instance with vm_state building and task_state spawning. [ 944.586347] env[61923]: DEBUG oslo_vmware.api [None req-a9734b29-9738-4ea3-b52c-d0194f5d6fd9 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Task: {'id': task-1377839, 'name': Rename_Task, 'duration_secs': 0.132094} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 944.586779] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-a9734b29-9738-4ea3-b52c-d0194f5d6fd9 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: 02a8f197-80bb-4cee-bdd6-b07705759986] Powering on the VM {{(pid=61923) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 944.587145] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d1a5721a-0fd6-4269-92a8-32863c03c2da {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.594552] env[61923]: INFO nova.compute.manager [None req-29a01eb2-e144-458f-9b46-c49e64b03e88 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] [instance: a511392e-9ab6-42fb-b07c-f90e9786dcc1] Shelve offloading [ 944.598672] env[61923]: DEBUG oslo_vmware.api [None req-a9734b29-9738-4ea3-b52c-d0194f5d6fd9 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Waiting for the task: (returnval){ [ 944.598672] env[61923]: value = "task-1377840" [ 944.598672] env[61923]: _type = "Task" [ 944.598672] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 944.598672] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-29a01eb2-e144-458f-9b46-c49e64b03e88 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] [instance: a511392e-9ab6-42fb-b07c-f90e9786dcc1] Powering off the VM {{(pid=61923) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 944.599317] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-70e410a1-7447-4aae-9931-be3f6a24a88d {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.612319] env[61923]: DEBUG oslo_vmware.api [None req-a9734b29-9738-4ea3-b52c-d0194f5d6fd9 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Task: {'id': task-1377840, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.614132] env[61923]: DEBUG oslo_vmware.api [None req-29a01eb2-e144-458f-9b46-c49e64b03e88 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Waiting for the task: (returnval){ [ 944.614132] env[61923]: value = "task-1377841" [ 944.614132] env[61923]: _type = "Task" [ 944.614132] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 944.624559] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-29a01eb2-e144-458f-9b46-c49e64b03e88 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] [instance: a511392e-9ab6-42fb-b07c-f90e9786dcc1] VM already powered off {{(pid=61923) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 944.624752] env[61923]: DEBUG nova.compute.manager [None req-29a01eb2-e144-458f-9b46-c49e64b03e88 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] [instance: a511392e-9ab6-42fb-b07c-f90e9786dcc1] Checking state {{(pid=61923) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 944.625508] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0e588a8-dea8-4062-a2b3-d263bf299842 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.630873] env[61923]: DEBUG oslo_concurrency.lockutils [None req-29a01eb2-e144-458f-9b46-c49e64b03e88 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Acquiring lock "refresh_cache-a511392e-9ab6-42fb-b07c-f90e9786dcc1" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 944.631051] env[61923]: DEBUG oslo_concurrency.lockutils [None req-29a01eb2-e144-458f-9b46-c49e64b03e88 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Acquired lock "refresh_cache-a511392e-9ab6-42fb-b07c-f90e9786dcc1" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 944.631229] env[61923]: DEBUG nova.network.neutron [None req-29a01eb2-e144-458f-9b46-c49e64b03e88 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] [instance: a511392e-9ab6-42fb-b07c-f90e9786dcc1] Building network info cache for instance {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 944.676687] env[61923]: DEBUG nova.network.neutron [None req-102bab57-0500-4a8d-94a7-052c22c7c916 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: 7867c6cf-2ad6-414f-ab52-581827860836] Successfully updated port: 972754b8-9333-4b89-8435-ca8a98aa7f39 {{(pid=61923) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 944.688861] env[61923]: DEBUG nova.scheduler.client.report [None req-69e9bd4b-8d1d-4678-80b0-4de2eeffdae8 tempest-ServersAaction247Test-1300676990 tempest-ServersAaction247Test-1300676990-project-member] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 944.832155] env[61923]: DEBUG nova.compute.manager [None req-704a26eb-9a13-4b75-8212-d4cbe21aa5e4 tempest-ServersV294TestFqdnHostnames-833840938 tempest-ServersV294TestFqdnHostnames-833840938-project-member] [instance: aba844c1-d488-4e95-a550-bd1cb7331fb2] Starting instance... {{(pid=61923) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 945.109290] env[61923]: DEBUG oslo_vmware.api [None req-a9734b29-9738-4ea3-b52c-d0194f5d6fd9 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Task: {'id': task-1377840, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.183432] env[61923]: DEBUG oslo_concurrency.lockutils [None req-102bab57-0500-4a8d-94a7-052c22c7c916 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Acquiring lock "refresh_cache-7867c6cf-2ad6-414f-ab52-581827860836" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 945.183661] env[61923]: DEBUG oslo_concurrency.lockutils [None req-102bab57-0500-4a8d-94a7-052c22c7c916 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Acquired lock "refresh_cache-7867c6cf-2ad6-414f-ab52-581827860836" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 945.183753] env[61923]: DEBUG nova.network.neutron [None req-102bab57-0500-4a8d-94a7-052c22c7c916 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: 7867c6cf-2ad6-414f-ab52-581827860836] Building network info cache for instance {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 945.194039] env[61923]: DEBUG oslo_concurrency.lockutils [None req-69e9bd4b-8d1d-4678-80b0-4de2eeffdae8 tempest-ServersAaction247Test-1300676990 tempest-ServersAaction247Test-1300676990-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.188s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 945.194359] env[61923]: DEBUG nova.compute.manager [None req-69e9bd4b-8d1d-4678-80b0-4de2eeffdae8 tempest-ServersAaction247Test-1300676990 tempest-ServersAaction247Test-1300676990-project-member] [instance: 44196b4c-1401-40fa-bd14-04a49947ab15] Start building networks asynchronously for instance. {{(pid=61923) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 945.341349] env[61923]: DEBUG nova.network.neutron [None req-29a01eb2-e144-458f-9b46-c49e64b03e88 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] [instance: a511392e-9ab6-42fb-b07c-f90e9786dcc1] Updating instance_info_cache with network_info: [{"id": "0325cd38-bd49-4132-85e3-c66fc9efe49f", "address": "fa:16:3e:7f:89:31", "network": {"id": "07fc148e-c3ba-4e53-bec2-36500ec38add", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1374533437-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0d24f4a47d1f4b79a7d999a4dacb99ba", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b2907cce-d529-4809-af05-d29397bed211", "external-id": "nsx-vlan-transportzone-427", "segmentation_id": 427, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0325cd38-bd", "ovs_interfaceid": "0325cd38-bd49-4132-85e3-c66fc9efe49f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 945.356124] env[61923]: DEBUG oslo_concurrency.lockutils [None req-704a26eb-9a13-4b75-8212-d4cbe21aa5e4 tempest-ServersV294TestFqdnHostnames-833840938 tempest-ServersV294TestFqdnHostnames-833840938-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 945.356391] env[61923]: DEBUG oslo_concurrency.lockutils [None req-704a26eb-9a13-4b75-8212-d4cbe21aa5e4 tempest-ServersV294TestFqdnHostnames-833840938 tempest-ServersV294TestFqdnHostnames-833840938-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 945.357942] env[61923]: INFO nova.compute.claims [None req-704a26eb-9a13-4b75-8212-d4cbe21aa5e4 tempest-ServersV294TestFqdnHostnames-833840938 tempest-ServersV294TestFqdnHostnames-833840938-project-member] [instance: aba844c1-d488-4e95-a550-bd1cb7331fb2] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 945.608879] env[61923]: DEBUG oslo_vmware.api [None req-a9734b29-9738-4ea3-b52c-d0194f5d6fd9 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Task: {'id': task-1377840, 'name': PowerOnVM_Task, 'duration_secs': 0.542078} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 945.609127] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-a9734b29-9738-4ea3-b52c-d0194f5d6fd9 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: 02a8f197-80bb-4cee-bdd6-b07705759986] Powered on the VM {{(pid=61923) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 945.609334] env[61923]: INFO nova.compute.manager [None req-a9734b29-9738-4ea3-b52c-d0194f5d6fd9 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: 02a8f197-80bb-4cee-bdd6-b07705759986] Took 7.99 seconds to spawn the instance on the hypervisor. [ 945.609576] env[61923]: DEBUG nova.compute.manager [None req-a9734b29-9738-4ea3-b52c-d0194f5d6fd9 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: 02a8f197-80bb-4cee-bdd6-b07705759986] Checking state {{(pid=61923) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 945.610296] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4cdb71c-4766-4a38-95fb-846324f61d14 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.698546] env[61923]: DEBUG nova.compute.utils [None req-69e9bd4b-8d1d-4678-80b0-4de2eeffdae8 tempest-ServersAaction247Test-1300676990 tempest-ServersAaction247Test-1300676990-project-member] Using /dev/sd instead of None {{(pid=61923) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 945.699907] env[61923]: DEBUG nova.compute.manager [None req-69e9bd4b-8d1d-4678-80b0-4de2eeffdae8 tempest-ServersAaction247Test-1300676990 tempest-ServersAaction247Test-1300676990-project-member] [instance: 44196b4c-1401-40fa-bd14-04a49947ab15] Not allocating networking since 'none' was specified. {{(pid=61923) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 945.716049] env[61923]: DEBUG nova.network.neutron [None req-102bab57-0500-4a8d-94a7-052c22c7c916 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: 7867c6cf-2ad6-414f-ab52-581827860836] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 945.846675] env[61923]: DEBUG oslo_concurrency.lockutils [None req-29a01eb2-e144-458f-9b46-c49e64b03e88 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Releasing lock "refresh_cache-a511392e-9ab6-42fb-b07c-f90e9786dcc1" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 945.971351] env[61923]: DEBUG nova.network.neutron [None req-102bab57-0500-4a8d-94a7-052c22c7c916 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: 7867c6cf-2ad6-414f-ab52-581827860836] Updating instance_info_cache with network_info: [{"id": "972754b8-9333-4b89-8435-ca8a98aa7f39", "address": "fa:16:3e:de:09:04", "network": {"id": "9533e79d-fde2-4c10-86a0-86a36845a8cf", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1736743137-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "780d3541d9604417b977bb62390c4299", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6d650b26-c3e7-4de7-98db-5e4b816d123a", "external-id": "nsx-vlan-transportzone-757", "segmentation_id": 757, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap972754b8-93", "ovs_interfaceid": "972754b8-9333-4b89-8435-ca8a98aa7f39", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 946.127952] env[61923]: INFO nova.compute.manager [None req-a9734b29-9738-4ea3-b52c-d0194f5d6fd9 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: 02a8f197-80bb-4cee-bdd6-b07705759986] Took 15.53 seconds to build instance. [ 946.201257] env[61923]: DEBUG nova.compute.manager [None req-69e9bd4b-8d1d-4678-80b0-4de2eeffdae8 tempest-ServersAaction247Test-1300676990 tempest-ServersAaction247Test-1300676990-project-member] [instance: 44196b4c-1401-40fa-bd14-04a49947ab15] Start building block device mappings for instance. {{(pid=61923) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 946.272032] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-29a01eb2-e144-458f-9b46-c49e64b03e88 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] [instance: a511392e-9ab6-42fb-b07c-f90e9786dcc1] Destroying instance {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 946.272032] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a6bb5d2-fb0c-42ae-a8f5-d326ba83ee57 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.278605] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-29a01eb2-e144-458f-9b46-c49e64b03e88 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] [instance: a511392e-9ab6-42fb-b07c-f90e9786dcc1] Unregistering the VM {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 946.278843] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-18c7b293-9c03-4dff-ab0f-6363fba0b087 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.347923] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-29a01eb2-e144-458f-9b46-c49e64b03e88 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] [instance: a511392e-9ab6-42fb-b07c-f90e9786dcc1] Unregistered the VM {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 946.348064] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-29a01eb2-e144-458f-9b46-c49e64b03e88 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] [instance: a511392e-9ab6-42fb-b07c-f90e9786dcc1] Deleting contents of the VM from datastore datastore2 {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 946.348265] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-29a01eb2-e144-458f-9b46-c49e64b03e88 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Deleting the datastore file [datastore2] a511392e-9ab6-42fb-b07c-f90e9786dcc1 {{(pid=61923) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 946.348548] env[61923]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-0b6477f7-2bce-4dac-9d6c-a61d063e08a3 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.354939] env[61923]: DEBUG oslo_vmware.api [None req-29a01eb2-e144-458f-9b46-c49e64b03e88 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Waiting for the task: (returnval){ [ 946.354939] env[61923]: value = "task-1377843" [ 946.354939] env[61923]: _type = "Task" [ 946.354939] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 946.362290] env[61923]: DEBUG oslo_vmware.api [None req-29a01eb2-e144-458f-9b46-c49e64b03e88 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Task: {'id': task-1377843, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.457022] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0372ce8-39ee-4a1f-bd96-7558fed2f531 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.464072] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-652dab65-a033-49a0-9927-735c91a82a01 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.493197] env[61923]: DEBUG oslo_concurrency.lockutils [None req-102bab57-0500-4a8d-94a7-052c22c7c916 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Releasing lock "refresh_cache-7867c6cf-2ad6-414f-ab52-581827860836" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 946.493478] env[61923]: DEBUG nova.compute.manager [None req-102bab57-0500-4a8d-94a7-052c22c7c916 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: 7867c6cf-2ad6-414f-ab52-581827860836] Instance network_info: |[{"id": "972754b8-9333-4b89-8435-ca8a98aa7f39", "address": "fa:16:3e:de:09:04", "network": {"id": "9533e79d-fde2-4c10-86a0-86a36845a8cf", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1736743137-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "780d3541d9604417b977bb62390c4299", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6d650b26-c3e7-4de7-98db-5e4b816d123a", "external-id": "nsx-vlan-transportzone-757", "segmentation_id": 757, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap972754b8-93", "ovs_interfaceid": "972754b8-9333-4b89-8435-ca8a98aa7f39", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61923) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 946.494433] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-102bab57-0500-4a8d-94a7-052c22c7c916 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: 7867c6cf-2ad6-414f-ab52-581827860836] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:de:09:04', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '6d650b26-c3e7-4de7-98db-5e4b816d123a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '972754b8-9333-4b89-8435-ca8a98aa7f39', 'vif_model': 'vmxnet3'}] {{(pid=61923) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 946.501656] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-102bab57-0500-4a8d-94a7-052c22c7c916 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Creating folder: Project (780d3541d9604417b977bb62390c4299). Parent ref: group-v292629. {{(pid=61923) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 946.502374] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-487f2bf5-bf91-4ac0-93a5-5956671e1529 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.505103] env[61923]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-078dc0f6-1eec-4ca5-b253-c08e2380a1cb {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.511676] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2417aea9-d8a2-4c2d-923f-9c14968b30df {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.517022] env[61923]: INFO nova.virt.vmwareapi.vm_util [None req-102bab57-0500-4a8d-94a7-052c22c7c916 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Created folder: Project (780d3541d9604417b977bb62390c4299) in parent group-v292629. [ 946.517022] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-102bab57-0500-4a8d-94a7-052c22c7c916 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Creating folder: Instances. Parent ref: group-v292725. {{(pid=61923) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 946.517022] env[61923]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-77908679-72ee-4110-868e-909f3db9aa1e {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.525765] env[61923]: DEBUG nova.compute.provider_tree [None req-704a26eb-9a13-4b75-8212-d4cbe21aa5e4 tempest-ServersV294TestFqdnHostnames-833840938 tempest-ServersV294TestFqdnHostnames-833840938-project-member] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 946.527847] env[61923]: INFO nova.virt.vmwareapi.vm_util [None req-102bab57-0500-4a8d-94a7-052c22c7c916 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Created folder: Instances in parent group-v292725. [ 946.528088] env[61923]: DEBUG oslo.service.loopingcall [None req-102bab57-0500-4a8d-94a7-052c22c7c916 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61923) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 946.528477] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7867c6cf-2ad6-414f-ab52-581827860836] Creating VM on the ESX host {{(pid=61923) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 946.528674] env[61923]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-441bd609-fd87-41ef-9ee4-3619094061b2 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.548973] env[61923]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 946.548973] env[61923]: value = "task-1377846" [ 946.548973] env[61923]: _type = "Task" [ 946.548973] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 946.556292] env[61923]: DEBUG oslo_vmware.api [-] Task: {'id': task-1377846, 'name': CreateVM_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.570222] env[61923]: DEBUG nova.compute.manager [req-047f323b-e90e-4198-8cca-69523538266b req-39a6f9be-532c-46e8-9edb-7a83d60bb578 service nova] [instance: 7867c6cf-2ad6-414f-ab52-581827860836] Received event network-changed-972754b8-9333-4b89-8435-ca8a98aa7f39 {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 946.570411] env[61923]: DEBUG nova.compute.manager [req-047f323b-e90e-4198-8cca-69523538266b req-39a6f9be-532c-46e8-9edb-7a83d60bb578 service nova] [instance: 7867c6cf-2ad6-414f-ab52-581827860836] Refreshing instance network info cache due to event network-changed-972754b8-9333-4b89-8435-ca8a98aa7f39. {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 946.570608] env[61923]: DEBUG oslo_concurrency.lockutils [req-047f323b-e90e-4198-8cca-69523538266b req-39a6f9be-532c-46e8-9edb-7a83d60bb578 service nova] Acquiring lock "refresh_cache-7867c6cf-2ad6-414f-ab52-581827860836" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 946.570756] env[61923]: DEBUG oslo_concurrency.lockutils [req-047f323b-e90e-4198-8cca-69523538266b req-39a6f9be-532c-46e8-9edb-7a83d60bb578 service nova] Acquired lock "refresh_cache-7867c6cf-2ad6-414f-ab52-581827860836" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 946.570914] env[61923]: DEBUG nova.network.neutron [req-047f323b-e90e-4198-8cca-69523538266b req-39a6f9be-532c-46e8-9edb-7a83d60bb578 service nova] [instance: 7867c6cf-2ad6-414f-ab52-581827860836] Refreshing network info cache for port 972754b8-9333-4b89-8435-ca8a98aa7f39 {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 946.630215] env[61923]: DEBUG oslo_concurrency.lockutils [None req-a9734b29-9738-4ea3-b52c-d0194f5d6fd9 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Lock "02a8f197-80bb-4cee-bdd6-b07705759986" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 17.041s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 946.864517] env[61923]: DEBUG oslo_vmware.api [None req-29a01eb2-e144-458f-9b46-c49e64b03e88 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Task: {'id': task-1377843, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.135256} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 946.864813] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-29a01eb2-e144-458f-9b46-c49e64b03e88 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Deleted the datastore file {{(pid=61923) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 946.865016] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-29a01eb2-e144-458f-9b46-c49e64b03e88 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] [instance: a511392e-9ab6-42fb-b07c-f90e9786dcc1] Deleted contents of the VM from datastore datastore2 {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 946.865199] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-29a01eb2-e144-458f-9b46-c49e64b03e88 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] [instance: a511392e-9ab6-42fb-b07c-f90e9786dcc1] Instance destroyed {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 946.884156] env[61923]: INFO nova.scheduler.client.report [None req-29a01eb2-e144-458f-9b46-c49e64b03e88 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Deleted allocations for instance a511392e-9ab6-42fb-b07c-f90e9786dcc1 [ 946.919334] env[61923]: DEBUG oslo_concurrency.lockutils [None req-af395613-edaf-421a-b79b-9579e1c9e371 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Acquiring lock "02a8f197-80bb-4cee-bdd6-b07705759986" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 946.919662] env[61923]: DEBUG oslo_concurrency.lockutils [None req-af395613-edaf-421a-b79b-9579e1c9e371 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Lock "02a8f197-80bb-4cee-bdd6-b07705759986" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 946.919880] env[61923]: DEBUG oslo_concurrency.lockutils [None req-af395613-edaf-421a-b79b-9579e1c9e371 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Acquiring lock "02a8f197-80bb-4cee-bdd6-b07705759986-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 946.920138] env[61923]: DEBUG oslo_concurrency.lockutils [None req-af395613-edaf-421a-b79b-9579e1c9e371 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Lock "02a8f197-80bb-4cee-bdd6-b07705759986-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 946.920325] env[61923]: DEBUG oslo_concurrency.lockutils [None req-af395613-edaf-421a-b79b-9579e1c9e371 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Lock "02a8f197-80bb-4cee-bdd6-b07705759986-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 946.925951] env[61923]: INFO nova.compute.manager [None req-af395613-edaf-421a-b79b-9579e1c9e371 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: 02a8f197-80bb-4cee-bdd6-b07705759986] Terminating instance [ 946.928080] env[61923]: DEBUG nova.compute.manager [None req-af395613-edaf-421a-b79b-9579e1c9e371 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: 02a8f197-80bb-4cee-bdd6-b07705759986] Start destroying the instance on the hypervisor. {{(pid=61923) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 946.928285] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-af395613-edaf-421a-b79b-9579e1c9e371 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: 02a8f197-80bb-4cee-bdd6-b07705759986] Destroying instance {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 946.929122] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58666bf6-8b34-4781-9100-405ae66357b6 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.936675] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-af395613-edaf-421a-b79b-9579e1c9e371 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: 02a8f197-80bb-4cee-bdd6-b07705759986] Powering off the VM {{(pid=61923) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 946.936910] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-979f985d-73ff-41da-a697-ece1b2389f43 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.942752] env[61923]: DEBUG oslo_vmware.api [None req-af395613-edaf-421a-b79b-9579e1c9e371 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Waiting for the task: (returnval){ [ 946.942752] env[61923]: value = "task-1377847" [ 946.942752] env[61923]: _type = "Task" [ 946.942752] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 946.950636] env[61923]: DEBUG oslo_vmware.api [None req-af395613-edaf-421a-b79b-9579e1c9e371 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Task: {'id': task-1377847, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.029929] env[61923]: DEBUG nova.scheduler.client.report [None req-704a26eb-9a13-4b75-8212-d4cbe21aa5e4 tempest-ServersV294TestFqdnHostnames-833840938 tempest-ServersV294TestFqdnHostnames-833840938-project-member] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 947.060141] env[61923]: DEBUG oslo_vmware.api [-] Task: {'id': task-1377846, 'name': CreateVM_Task, 'duration_secs': 0.3119} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 947.060312] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 7867c6cf-2ad6-414f-ab52-581827860836] Created VM on the ESX host {{(pid=61923) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 947.061051] env[61923]: DEBUG oslo_concurrency.lockutils [None req-102bab57-0500-4a8d-94a7-052c22c7c916 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 947.061242] env[61923]: DEBUG oslo_concurrency.lockutils [None req-102bab57-0500-4a8d-94a7-052c22c7c916 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 947.061623] env[61923]: DEBUG oslo_concurrency.lockutils [None req-102bab57-0500-4a8d-94a7-052c22c7c916 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 947.061996] env[61923]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-df740b31-a401-40ff-a517-b01d4516c947 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.067034] env[61923]: DEBUG oslo_vmware.api [None req-102bab57-0500-4a8d-94a7-052c22c7c916 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Waiting for the task: (returnval){ [ 947.067034] env[61923]: value = "session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]5231d4bb-d542-c809-e57c-11e2d2796e69" [ 947.067034] env[61923]: _type = "Task" [ 947.067034] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 947.078552] env[61923]: DEBUG oslo_vmware.api [None req-102bab57-0500-4a8d-94a7-052c22c7c916 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]5231d4bb-d542-c809-e57c-11e2d2796e69, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.211188] env[61923]: DEBUG nova.compute.manager [None req-69e9bd4b-8d1d-4678-80b0-4de2eeffdae8 tempest-ServersAaction247Test-1300676990 tempest-ServersAaction247Test-1300676990-project-member] [instance: 44196b4c-1401-40fa-bd14-04a49947ab15] Start spawning the instance on the hypervisor. {{(pid=61923) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 947.235237] env[61923]: DEBUG nova.virt.hardware [None req-69e9bd4b-8d1d-4678-80b0-4de2eeffdae8 tempest-ServersAaction247Test-1300676990 tempest-ServersAaction247Test-1300676990-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-29T20:07:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-29T20:07:35Z,direct_url=,disk_format='vmdk',id=0f153f63-ae0a-45b1-b7f5-7f9b673c947f,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='4e7b5e3b3c3443c8bd5c70f8a15739f5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-29T20:07:36Z,virtual_size=,visibility=), allow threads: False {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 947.235502] env[61923]: DEBUG nova.virt.hardware [None req-69e9bd4b-8d1d-4678-80b0-4de2eeffdae8 tempest-ServersAaction247Test-1300676990 tempest-ServersAaction247Test-1300676990-project-member] Flavor limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 947.235665] env[61923]: DEBUG nova.virt.hardware [None req-69e9bd4b-8d1d-4678-80b0-4de2eeffdae8 tempest-ServersAaction247Test-1300676990 tempest-ServersAaction247Test-1300676990-project-member] Image limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 947.235847] env[61923]: DEBUG nova.virt.hardware [None req-69e9bd4b-8d1d-4678-80b0-4de2eeffdae8 tempest-ServersAaction247Test-1300676990 tempest-ServersAaction247Test-1300676990-project-member] Flavor pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 947.235995] env[61923]: DEBUG nova.virt.hardware [None req-69e9bd4b-8d1d-4678-80b0-4de2eeffdae8 tempest-ServersAaction247Test-1300676990 tempest-ServersAaction247Test-1300676990-project-member] Image pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 947.236160] env[61923]: DEBUG nova.virt.hardware [None req-69e9bd4b-8d1d-4678-80b0-4de2eeffdae8 tempest-ServersAaction247Test-1300676990 tempest-ServersAaction247Test-1300676990-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 947.236377] env[61923]: DEBUG nova.virt.hardware [None req-69e9bd4b-8d1d-4678-80b0-4de2eeffdae8 tempest-ServersAaction247Test-1300676990 tempest-ServersAaction247Test-1300676990-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 947.236537] env[61923]: DEBUG nova.virt.hardware [None req-69e9bd4b-8d1d-4678-80b0-4de2eeffdae8 tempest-ServersAaction247Test-1300676990 tempest-ServersAaction247Test-1300676990-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 947.236710] env[61923]: DEBUG nova.virt.hardware [None req-69e9bd4b-8d1d-4678-80b0-4de2eeffdae8 tempest-ServersAaction247Test-1300676990 tempest-ServersAaction247Test-1300676990-project-member] Got 1 possible topologies {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 947.236873] env[61923]: DEBUG nova.virt.hardware [None req-69e9bd4b-8d1d-4678-80b0-4de2eeffdae8 tempest-ServersAaction247Test-1300676990 tempest-ServersAaction247Test-1300676990-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 947.237056] env[61923]: DEBUG nova.virt.hardware [None req-69e9bd4b-8d1d-4678-80b0-4de2eeffdae8 tempest-ServersAaction247Test-1300676990 tempest-ServersAaction247Test-1300676990-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 947.237935] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3b56ac8-746d-423a-b3af-55c0d8aa20d1 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.245386] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8380ac95-83f4-411c-96ee-14e55d8c61ad {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.260100] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-69e9bd4b-8d1d-4678-80b0-4de2eeffdae8 tempest-ServersAaction247Test-1300676990 tempest-ServersAaction247Test-1300676990-project-member] [instance: 44196b4c-1401-40fa-bd14-04a49947ab15] Instance VIF info [] {{(pid=61923) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 947.265510] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-69e9bd4b-8d1d-4678-80b0-4de2eeffdae8 tempest-ServersAaction247Test-1300676990 tempest-ServersAaction247Test-1300676990-project-member] Creating folder: Project (0bd3b1570ba245b6928e03356350e857). Parent ref: group-v292629. {{(pid=61923) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 947.265781] env[61923]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-77a06c4e-7351-4f57-b754-ce19c8b8dfee {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.275818] env[61923]: INFO nova.virt.vmwareapi.vm_util [None req-69e9bd4b-8d1d-4678-80b0-4de2eeffdae8 tempest-ServersAaction247Test-1300676990 tempest-ServersAaction247Test-1300676990-project-member] Created folder: Project (0bd3b1570ba245b6928e03356350e857) in parent group-v292629. [ 947.275818] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-69e9bd4b-8d1d-4678-80b0-4de2eeffdae8 tempest-ServersAaction247Test-1300676990 tempest-ServersAaction247Test-1300676990-project-member] Creating folder: Instances. Parent ref: group-v292728. {{(pid=61923) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 947.275818] env[61923]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-881e44f6-873b-4b94-9843-b4b99dfef2d6 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.283035] env[61923]: DEBUG nova.network.neutron [req-047f323b-e90e-4198-8cca-69523538266b req-39a6f9be-532c-46e8-9edb-7a83d60bb578 service nova] [instance: 7867c6cf-2ad6-414f-ab52-581827860836] Updated VIF entry in instance network info cache for port 972754b8-9333-4b89-8435-ca8a98aa7f39. {{(pid=61923) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 947.283374] env[61923]: DEBUG nova.network.neutron [req-047f323b-e90e-4198-8cca-69523538266b req-39a6f9be-532c-46e8-9edb-7a83d60bb578 service nova] [instance: 7867c6cf-2ad6-414f-ab52-581827860836] Updating instance_info_cache with network_info: [{"id": "972754b8-9333-4b89-8435-ca8a98aa7f39", "address": "fa:16:3e:de:09:04", "network": {"id": "9533e79d-fde2-4c10-86a0-86a36845a8cf", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1736743137-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "780d3541d9604417b977bb62390c4299", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6d650b26-c3e7-4de7-98db-5e4b816d123a", "external-id": "nsx-vlan-transportzone-757", "segmentation_id": 757, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap972754b8-93", "ovs_interfaceid": "972754b8-9333-4b89-8435-ca8a98aa7f39", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 947.285751] env[61923]: INFO nova.virt.vmwareapi.vm_util [None req-69e9bd4b-8d1d-4678-80b0-4de2eeffdae8 tempest-ServersAaction247Test-1300676990 tempest-ServersAaction247Test-1300676990-project-member] Created folder: Instances in parent group-v292728. [ 947.286309] env[61923]: DEBUG oslo.service.loopingcall [None req-69e9bd4b-8d1d-4678-80b0-4de2eeffdae8 tempest-ServersAaction247Test-1300676990 tempest-ServersAaction247Test-1300676990-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61923) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 947.286427] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 44196b4c-1401-40fa-bd14-04a49947ab15] Creating VM on the ESX host {{(pid=61923) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 947.286576] env[61923]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-04df4e01-27b0-40cd-8c86-6c7506962098 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.305167] env[61923]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 947.305167] env[61923]: value = "task-1377850" [ 947.305167] env[61923]: _type = "Task" [ 947.305167] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 947.312381] env[61923]: DEBUG oslo_vmware.api [-] Task: {'id': task-1377850, 'name': CreateVM_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.389525] env[61923]: DEBUG oslo_concurrency.lockutils [None req-29a01eb2-e144-458f-9b46-c49e64b03e88 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 947.451843] env[61923]: DEBUG oslo_vmware.api [None req-af395613-edaf-421a-b79b-9579e1c9e371 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Task: {'id': task-1377847, 'name': PowerOffVM_Task, 'duration_secs': 0.209644} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 947.452143] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-af395613-edaf-421a-b79b-9579e1c9e371 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: 02a8f197-80bb-4cee-bdd6-b07705759986] Powered off the VM {{(pid=61923) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 947.452321] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-af395613-edaf-421a-b79b-9579e1c9e371 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: 02a8f197-80bb-4cee-bdd6-b07705759986] Unregistering the VM {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 947.452582] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-6c525c45-a822-4d50-8f95-944afc6347ef {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.526232] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-af395613-edaf-421a-b79b-9579e1c9e371 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: 02a8f197-80bb-4cee-bdd6-b07705759986] Unregistered the VM {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 947.526528] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-af395613-edaf-421a-b79b-9579e1c9e371 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: 02a8f197-80bb-4cee-bdd6-b07705759986] Deleting contents of the VM from datastore datastore1 {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 947.526748] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-af395613-edaf-421a-b79b-9579e1c9e371 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Deleting the datastore file [datastore1] 02a8f197-80bb-4cee-bdd6-b07705759986 {{(pid=61923) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 947.527137] env[61923]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-55ffc22b-a163-42dd-934d-a51a69e179fa {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.534738] env[61923]: DEBUG oslo_concurrency.lockutils [None req-704a26eb-9a13-4b75-8212-d4cbe21aa5e4 tempest-ServersV294TestFqdnHostnames-833840938 tempest-ServersV294TestFqdnHostnames-833840938-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.178s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 947.535313] env[61923]: DEBUG nova.compute.manager [None req-704a26eb-9a13-4b75-8212-d4cbe21aa5e4 tempest-ServersV294TestFqdnHostnames-833840938 tempest-ServersV294TestFqdnHostnames-833840938-project-member] [instance: aba844c1-d488-4e95-a550-bd1cb7331fb2] Start building networks asynchronously for instance. {{(pid=61923) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 947.539515] env[61923]: DEBUG oslo_concurrency.lockutils [None req-29a01eb2-e144-458f-9b46-c49e64b03e88 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.151s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 947.539728] env[61923]: DEBUG nova.objects.instance [None req-29a01eb2-e144-458f-9b46-c49e64b03e88 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Lazy-loading 'resources' on Instance uuid a511392e-9ab6-42fb-b07c-f90e9786dcc1 {{(pid=61923) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 947.540757] env[61923]: DEBUG oslo_vmware.api [None req-af395613-edaf-421a-b79b-9579e1c9e371 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Waiting for the task: (returnval){ [ 947.540757] env[61923]: value = "task-1377852" [ 947.540757] env[61923]: _type = "Task" [ 947.540757] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 947.551029] env[61923]: DEBUG oslo_vmware.api [None req-af395613-edaf-421a-b79b-9579e1c9e371 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Task: {'id': task-1377852, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.577597] env[61923]: DEBUG oslo_vmware.api [None req-102bab57-0500-4a8d-94a7-052c22c7c916 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]5231d4bb-d542-c809-e57c-11e2d2796e69, 'name': SearchDatastore_Task, 'duration_secs': 0.009483} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 947.577911] env[61923]: DEBUG oslo_concurrency.lockutils [None req-102bab57-0500-4a8d-94a7-052c22c7c916 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 947.578162] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-102bab57-0500-4a8d-94a7-052c22c7c916 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: 7867c6cf-2ad6-414f-ab52-581827860836] Processing image 0f153f63-ae0a-45b1-b7f5-7f9b673c947f {{(pid=61923) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 947.578654] env[61923]: DEBUG oslo_concurrency.lockutils [None req-102bab57-0500-4a8d-94a7-052c22c7c916 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 947.578654] env[61923]: DEBUG oslo_concurrency.lockutils [None req-102bab57-0500-4a8d-94a7-052c22c7c916 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 947.578771] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-102bab57-0500-4a8d-94a7-052c22c7c916 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61923) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 947.579018] env[61923]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-716fd7ae-2318-4f61-922f-9e2782b24434 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.722698] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-102bab57-0500-4a8d-94a7-052c22c7c916 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61923) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 947.722908] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-102bab57-0500-4a8d-94a7-052c22c7c916 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61923) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 947.723691] env[61923]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-90026d6f-6cec-44b6-b1a8-80db4c79a4a9 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.728965] env[61923]: DEBUG oslo_vmware.api [None req-102bab57-0500-4a8d-94a7-052c22c7c916 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Waiting for the task: (returnval){ [ 947.728965] env[61923]: value = "session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52344400-bd6b-8472-3748-e096b6066d0d" [ 947.728965] env[61923]: _type = "Task" [ 947.728965] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 947.736903] env[61923]: DEBUG oslo_vmware.api [None req-102bab57-0500-4a8d-94a7-052c22c7c916 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52344400-bd6b-8472-3748-e096b6066d0d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.787761] env[61923]: DEBUG oslo_concurrency.lockutils [req-047f323b-e90e-4198-8cca-69523538266b req-39a6f9be-532c-46e8-9edb-7a83d60bb578 service nova] Releasing lock "refresh_cache-7867c6cf-2ad6-414f-ab52-581827860836" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 947.788062] env[61923]: DEBUG nova.compute.manager [req-047f323b-e90e-4198-8cca-69523538266b req-39a6f9be-532c-46e8-9edb-7a83d60bb578 service nova] [instance: a511392e-9ab6-42fb-b07c-f90e9786dcc1] Received event network-vif-unplugged-0325cd38-bd49-4132-85e3-c66fc9efe49f {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 947.788287] env[61923]: DEBUG oslo_concurrency.lockutils [req-047f323b-e90e-4198-8cca-69523538266b req-39a6f9be-532c-46e8-9edb-7a83d60bb578 service nova] Acquiring lock "a511392e-9ab6-42fb-b07c-f90e9786dcc1-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 947.788502] env[61923]: DEBUG oslo_concurrency.lockutils [req-047f323b-e90e-4198-8cca-69523538266b req-39a6f9be-532c-46e8-9edb-7a83d60bb578 service nova] Lock "a511392e-9ab6-42fb-b07c-f90e9786dcc1-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 947.788670] env[61923]: DEBUG oslo_concurrency.lockutils [req-047f323b-e90e-4198-8cca-69523538266b req-39a6f9be-532c-46e8-9edb-7a83d60bb578 service nova] Lock "a511392e-9ab6-42fb-b07c-f90e9786dcc1-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 947.788843] env[61923]: DEBUG nova.compute.manager [req-047f323b-e90e-4198-8cca-69523538266b req-39a6f9be-532c-46e8-9edb-7a83d60bb578 service nova] [instance: a511392e-9ab6-42fb-b07c-f90e9786dcc1] No waiting events found dispatching network-vif-unplugged-0325cd38-bd49-4132-85e3-c66fc9efe49f {{(pid=61923) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 947.789024] env[61923]: WARNING nova.compute.manager [req-047f323b-e90e-4198-8cca-69523538266b req-39a6f9be-532c-46e8-9edb-7a83d60bb578 service nova] [instance: a511392e-9ab6-42fb-b07c-f90e9786dcc1] Received unexpected event network-vif-unplugged-0325cd38-bd49-4132-85e3-c66fc9efe49f for instance with vm_state shelved and task_state shelving_offloading. [ 947.789199] env[61923]: DEBUG nova.compute.manager [req-047f323b-e90e-4198-8cca-69523538266b req-39a6f9be-532c-46e8-9edb-7a83d60bb578 service nova] [instance: a511392e-9ab6-42fb-b07c-f90e9786dcc1] Received event network-changed-0325cd38-bd49-4132-85e3-c66fc9efe49f {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 947.789361] env[61923]: DEBUG nova.compute.manager [req-047f323b-e90e-4198-8cca-69523538266b req-39a6f9be-532c-46e8-9edb-7a83d60bb578 service nova] [instance: a511392e-9ab6-42fb-b07c-f90e9786dcc1] Refreshing instance network info cache due to event network-changed-0325cd38-bd49-4132-85e3-c66fc9efe49f. {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 947.789608] env[61923]: DEBUG oslo_concurrency.lockutils [req-047f323b-e90e-4198-8cca-69523538266b req-39a6f9be-532c-46e8-9edb-7a83d60bb578 service nova] Acquiring lock "refresh_cache-a511392e-9ab6-42fb-b07c-f90e9786dcc1" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 947.789758] env[61923]: DEBUG oslo_concurrency.lockutils [req-047f323b-e90e-4198-8cca-69523538266b req-39a6f9be-532c-46e8-9edb-7a83d60bb578 service nova] Acquired lock "refresh_cache-a511392e-9ab6-42fb-b07c-f90e9786dcc1" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 947.789921] env[61923]: DEBUG nova.network.neutron [req-047f323b-e90e-4198-8cca-69523538266b req-39a6f9be-532c-46e8-9edb-7a83d60bb578 service nova] [instance: a511392e-9ab6-42fb-b07c-f90e9786dcc1] Refreshing network info cache for port 0325cd38-bd49-4132-85e3-c66fc9efe49f {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 947.815283] env[61923]: DEBUG oslo_vmware.api [-] Task: {'id': task-1377850, 'name': CreateVM_Task, 'duration_secs': 0.269235} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 947.815450] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 44196b4c-1401-40fa-bd14-04a49947ab15] Created VM on the ESX host {{(pid=61923) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 947.815866] env[61923]: DEBUG oslo_concurrency.lockutils [None req-69e9bd4b-8d1d-4678-80b0-4de2eeffdae8 tempest-ServersAaction247Test-1300676990 tempest-ServersAaction247Test-1300676990-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 947.816040] env[61923]: DEBUG oslo_concurrency.lockutils [None req-69e9bd4b-8d1d-4678-80b0-4de2eeffdae8 tempest-ServersAaction247Test-1300676990 tempest-ServersAaction247Test-1300676990-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 947.816369] env[61923]: DEBUG oslo_concurrency.lockutils [None req-69e9bd4b-8d1d-4678-80b0-4de2eeffdae8 tempest-ServersAaction247Test-1300676990 tempest-ServersAaction247Test-1300676990-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 947.816616] env[61923]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9f1e97b1-d659-4e06-87ab-82541d8cb3a9 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.821297] env[61923]: DEBUG oslo_vmware.api [None req-69e9bd4b-8d1d-4678-80b0-4de2eeffdae8 tempest-ServersAaction247Test-1300676990 tempest-ServersAaction247Test-1300676990-project-member] Waiting for the task: (returnval){ [ 947.821297] env[61923]: value = "session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52d47cac-f735-2d8b-01cc-80bb097f491f" [ 947.821297] env[61923]: _type = "Task" [ 947.821297] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 947.831341] env[61923]: DEBUG oslo_vmware.api [None req-69e9bd4b-8d1d-4678-80b0-4de2eeffdae8 tempest-ServersAaction247Test-1300676990 tempest-ServersAaction247Test-1300676990-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52d47cac-f735-2d8b-01cc-80bb097f491f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.043513] env[61923]: DEBUG nova.compute.utils [None req-704a26eb-9a13-4b75-8212-d4cbe21aa5e4 tempest-ServersV294TestFqdnHostnames-833840938 tempest-ServersV294TestFqdnHostnames-833840938-project-member] Using /dev/sd instead of None {{(pid=61923) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 948.045868] env[61923]: DEBUG nova.objects.instance [None req-29a01eb2-e144-458f-9b46-c49e64b03e88 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Lazy-loading 'numa_topology' on Instance uuid a511392e-9ab6-42fb-b07c-f90e9786dcc1 {{(pid=61923) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 948.046567] env[61923]: DEBUG nova.compute.manager [None req-704a26eb-9a13-4b75-8212-d4cbe21aa5e4 tempest-ServersV294TestFqdnHostnames-833840938 tempest-ServersV294TestFqdnHostnames-833840938-project-member] [instance: aba844c1-d488-4e95-a550-bd1cb7331fb2] Allocating IP information in the background. {{(pid=61923) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 948.046728] env[61923]: DEBUG nova.network.neutron [None req-704a26eb-9a13-4b75-8212-d4cbe21aa5e4 tempest-ServersV294TestFqdnHostnames-833840938 tempest-ServersV294TestFqdnHostnames-833840938-project-member] [instance: aba844c1-d488-4e95-a550-bd1cb7331fb2] allocate_for_instance() {{(pid=61923) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 948.059054] env[61923]: DEBUG oslo_vmware.api [None req-af395613-edaf-421a-b79b-9579e1c9e371 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Task: {'id': task-1377852, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.101764] env[61923]: DEBUG nova.policy [None req-704a26eb-9a13-4b75-8212-d4cbe21aa5e4 tempest-ServersV294TestFqdnHostnames-833840938 tempest-ServersV294TestFqdnHostnames-833840938-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a7da54525b5848f795fac485217b09cb', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e9e5485c43514e55b88bdd7d2dc2aee7', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61923) authorize /opt/stack/nova/nova/policy.py:201}} [ 948.238888] env[61923]: DEBUG oslo_vmware.api [None req-102bab57-0500-4a8d-94a7-052c22c7c916 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52344400-bd6b-8472-3748-e096b6066d0d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.331154] env[61923]: DEBUG oslo_vmware.api [None req-69e9bd4b-8d1d-4678-80b0-4de2eeffdae8 tempest-ServersAaction247Test-1300676990 tempest-ServersAaction247Test-1300676990-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52d47cac-f735-2d8b-01cc-80bb097f491f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.443600] env[61923]: DEBUG nova.network.neutron [None req-704a26eb-9a13-4b75-8212-d4cbe21aa5e4 tempest-ServersV294TestFqdnHostnames-833840938 tempest-ServersV294TestFqdnHostnames-833840938-project-member] [instance: aba844c1-d488-4e95-a550-bd1cb7331fb2] Successfully created port: 6326df12-72bc-44d6-b660-4abdb5ca5cfb {{(pid=61923) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 948.546991] env[61923]: DEBUG nova.compute.manager [None req-704a26eb-9a13-4b75-8212-d4cbe21aa5e4 tempest-ServersV294TestFqdnHostnames-833840938 tempest-ServersV294TestFqdnHostnames-833840938-project-member] [instance: aba844c1-d488-4e95-a550-bd1cb7331fb2] Start building block device mappings for instance. {{(pid=61923) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 948.553246] env[61923]: DEBUG nova.objects.base [None req-29a01eb2-e144-458f-9b46-c49e64b03e88 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Object Instance lazy-loaded attributes: resources,numa_topology {{(pid=61923) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 948.561091] env[61923]: DEBUG oslo_vmware.api [None req-af395613-edaf-421a-b79b-9579e1c9e371 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Task: {'id': task-1377852, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.628121] env[61923]: DEBUG nova.network.neutron [req-047f323b-e90e-4198-8cca-69523538266b req-39a6f9be-532c-46e8-9edb-7a83d60bb578 service nova] [instance: a511392e-9ab6-42fb-b07c-f90e9786dcc1] Updated VIF entry in instance network info cache for port 0325cd38-bd49-4132-85e3-c66fc9efe49f. {{(pid=61923) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 948.628525] env[61923]: DEBUG nova.network.neutron [req-047f323b-e90e-4198-8cca-69523538266b req-39a6f9be-532c-46e8-9edb-7a83d60bb578 service nova] [instance: a511392e-9ab6-42fb-b07c-f90e9786dcc1] Updating instance_info_cache with network_info: [{"id": "0325cd38-bd49-4132-85e3-c66fc9efe49f", "address": "fa:16:3e:7f:89:31", "network": {"id": "07fc148e-c3ba-4e53-bec2-36500ec38add", "bridge": null, "label": "tempest-ServersNegativeTestJSON-1374533437-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0d24f4a47d1f4b79a7d999a4dacb99ba", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tap0325cd38-bd", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 948.646557] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9845baaa-bc20-422c-a987-ed6cb8be463b {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.653735] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2eb28bc1-66f1-4007-96a5-770156500b7e {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.683040] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9519d5f-5b87-446b-8e04-c3fb20479843 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.690129] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19fe4c04-1006-4e60-b16b-f5d1fc651f72 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.703044] env[61923]: DEBUG nova.compute.provider_tree [None req-29a01eb2-e144-458f-9b46-c49e64b03e88 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 948.738596] env[61923]: DEBUG oslo_vmware.api [None req-102bab57-0500-4a8d-94a7-052c22c7c916 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52344400-bd6b-8472-3748-e096b6066d0d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.834125] env[61923]: DEBUG oslo_vmware.api [None req-69e9bd4b-8d1d-4678-80b0-4de2eeffdae8 tempest-ServersAaction247Test-1300676990 tempest-ServersAaction247Test-1300676990-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52d47cac-f735-2d8b-01cc-80bb097f491f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.856100] env[61923]: DEBUG oslo_concurrency.lockutils [None req-eea05bb2-c691-4804-8c8d-019f8feaa521 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Acquiring lock "a511392e-9ab6-42fb-b07c-f90e9786dcc1" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 949.060520] env[61923]: DEBUG oslo_vmware.api [None req-af395613-edaf-421a-b79b-9579e1c9e371 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Task: {'id': task-1377852, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.131237] env[61923]: DEBUG oslo_concurrency.lockutils [req-047f323b-e90e-4198-8cca-69523538266b req-39a6f9be-532c-46e8-9edb-7a83d60bb578 service nova] Releasing lock "refresh_cache-a511392e-9ab6-42fb-b07c-f90e9786dcc1" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 949.206896] env[61923]: DEBUG nova.scheduler.client.report [None req-29a01eb2-e144-458f-9b46-c49e64b03e88 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 949.241465] env[61923]: DEBUG oslo_vmware.api [None req-102bab57-0500-4a8d-94a7-052c22c7c916 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52344400-bd6b-8472-3748-e096b6066d0d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.332308] env[61923]: DEBUG oslo_vmware.api [None req-69e9bd4b-8d1d-4678-80b0-4de2eeffdae8 tempest-ServersAaction247Test-1300676990 tempest-ServersAaction247Test-1300676990-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52d47cac-f735-2d8b-01cc-80bb097f491f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.557062] env[61923]: DEBUG oslo_vmware.api [None req-af395613-edaf-421a-b79b-9579e1c9e371 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Task: {'id': task-1377852, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.563926] env[61923]: DEBUG nova.compute.manager [None req-704a26eb-9a13-4b75-8212-d4cbe21aa5e4 tempest-ServersV294TestFqdnHostnames-833840938 tempest-ServersV294TestFqdnHostnames-833840938-project-member] [instance: aba844c1-d488-4e95-a550-bd1cb7331fb2] Start spawning the instance on the hypervisor. {{(pid=61923) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 949.589836] env[61923]: DEBUG nova.virt.hardware [None req-704a26eb-9a13-4b75-8212-d4cbe21aa5e4 tempest-ServersV294TestFqdnHostnames-833840938 tempest-ServersV294TestFqdnHostnames-833840938-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-29T20:07:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-29T20:07:35Z,direct_url=,disk_format='vmdk',id=0f153f63-ae0a-45b1-b7f5-7f9b673c947f,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='4e7b5e3b3c3443c8bd5c70f8a15739f5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-29T20:07:36Z,virtual_size=,visibility=), allow threads: False {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 949.590137] env[61923]: DEBUG nova.virt.hardware [None req-704a26eb-9a13-4b75-8212-d4cbe21aa5e4 tempest-ServersV294TestFqdnHostnames-833840938 tempest-ServersV294TestFqdnHostnames-833840938-project-member] Flavor limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 949.590303] env[61923]: DEBUG nova.virt.hardware [None req-704a26eb-9a13-4b75-8212-d4cbe21aa5e4 tempest-ServersV294TestFqdnHostnames-833840938 tempest-ServersV294TestFqdnHostnames-833840938-project-member] Image limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 949.590534] env[61923]: DEBUG nova.virt.hardware [None req-704a26eb-9a13-4b75-8212-d4cbe21aa5e4 tempest-ServersV294TestFqdnHostnames-833840938 tempest-ServersV294TestFqdnHostnames-833840938-project-member] Flavor pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 949.590708] env[61923]: DEBUG nova.virt.hardware [None req-704a26eb-9a13-4b75-8212-d4cbe21aa5e4 tempest-ServersV294TestFqdnHostnames-833840938 tempest-ServersV294TestFqdnHostnames-833840938-project-member] Image pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 949.590862] env[61923]: DEBUG nova.virt.hardware [None req-704a26eb-9a13-4b75-8212-d4cbe21aa5e4 tempest-ServersV294TestFqdnHostnames-833840938 tempest-ServersV294TestFqdnHostnames-833840938-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 949.591096] env[61923]: DEBUG nova.virt.hardware [None req-704a26eb-9a13-4b75-8212-d4cbe21aa5e4 tempest-ServersV294TestFqdnHostnames-833840938 tempest-ServersV294TestFqdnHostnames-833840938-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 949.591258] env[61923]: DEBUG nova.virt.hardware [None req-704a26eb-9a13-4b75-8212-d4cbe21aa5e4 tempest-ServersV294TestFqdnHostnames-833840938 tempest-ServersV294TestFqdnHostnames-833840938-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 949.591436] env[61923]: DEBUG nova.virt.hardware [None req-704a26eb-9a13-4b75-8212-d4cbe21aa5e4 tempest-ServersV294TestFqdnHostnames-833840938 tempest-ServersV294TestFqdnHostnames-833840938-project-member] Got 1 possible topologies {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 949.591605] env[61923]: DEBUG nova.virt.hardware [None req-704a26eb-9a13-4b75-8212-d4cbe21aa5e4 tempest-ServersV294TestFqdnHostnames-833840938 tempest-ServersV294TestFqdnHostnames-833840938-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 949.591795] env[61923]: DEBUG nova.virt.hardware [None req-704a26eb-9a13-4b75-8212-d4cbe21aa5e4 tempest-ServersV294TestFqdnHostnames-833840938 tempest-ServersV294TestFqdnHostnames-833840938-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 949.592701] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12a8fe77-240f-44b4-b62f-94b84daf11a1 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.600288] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48c9d707-f26f-4360-8a61-2944d27298be {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.712172] env[61923]: DEBUG oslo_concurrency.lockutils [None req-29a01eb2-e144-458f-9b46-c49e64b03e88 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.173s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 949.741555] env[61923]: DEBUG oslo_vmware.api [None req-102bab57-0500-4a8d-94a7-052c22c7c916 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52344400-bd6b-8472-3748-e096b6066d0d, 'name': SearchDatastore_Task, 'duration_secs': 1.848052} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 949.742180] env[61923]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-da585ecf-44f0-41d1-9c02-8018d02ceb31 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.747801] env[61923]: DEBUG oslo_vmware.api [None req-102bab57-0500-4a8d-94a7-052c22c7c916 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Waiting for the task: (returnval){ [ 949.747801] env[61923]: value = "session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]5251125a-a481-791b-025b-6c510dd95741" [ 949.747801] env[61923]: _type = "Task" [ 949.747801] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 949.755967] env[61923]: DEBUG oslo_vmware.api [None req-102bab57-0500-4a8d-94a7-052c22c7c916 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]5251125a-a481-791b-025b-6c510dd95741, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.833581] env[61923]: DEBUG oslo_vmware.api [None req-69e9bd4b-8d1d-4678-80b0-4de2eeffdae8 tempest-ServersAaction247Test-1300676990 tempest-ServersAaction247Test-1300676990-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52d47cac-f735-2d8b-01cc-80bb097f491f, 'name': SearchDatastore_Task, 'duration_secs': 1.75506} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 949.833893] env[61923]: DEBUG oslo_concurrency.lockutils [None req-69e9bd4b-8d1d-4678-80b0-4de2eeffdae8 tempest-ServersAaction247Test-1300676990 tempest-ServersAaction247Test-1300676990-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 949.834206] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-69e9bd4b-8d1d-4678-80b0-4de2eeffdae8 tempest-ServersAaction247Test-1300676990 tempest-ServersAaction247Test-1300676990-project-member] [instance: 44196b4c-1401-40fa-bd14-04a49947ab15] Processing image 0f153f63-ae0a-45b1-b7f5-7f9b673c947f {{(pid=61923) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 949.834427] env[61923]: DEBUG oslo_concurrency.lockutils [None req-69e9bd4b-8d1d-4678-80b0-4de2eeffdae8 tempest-ServersAaction247Test-1300676990 tempest-ServersAaction247Test-1300676990-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 950.058240] env[61923]: DEBUG oslo_vmware.api [None req-af395613-edaf-421a-b79b-9579e1c9e371 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Task: {'id': task-1377852, 'name': DeleteDatastoreFile_Task, 'duration_secs': 2.133465} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 950.058376] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-af395613-edaf-421a-b79b-9579e1c9e371 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Deleted the datastore file {{(pid=61923) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 950.058545] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-af395613-edaf-421a-b79b-9579e1c9e371 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: 02a8f197-80bb-4cee-bdd6-b07705759986] Deleted contents of the VM from datastore datastore1 {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 950.058733] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-af395613-edaf-421a-b79b-9579e1c9e371 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: 02a8f197-80bb-4cee-bdd6-b07705759986] Instance destroyed {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 950.058910] env[61923]: INFO nova.compute.manager [None req-af395613-edaf-421a-b79b-9579e1c9e371 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: 02a8f197-80bb-4cee-bdd6-b07705759986] Took 3.13 seconds to destroy the instance on the hypervisor. [ 950.059176] env[61923]: DEBUG oslo.service.loopingcall [None req-af395613-edaf-421a-b79b-9579e1c9e371 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61923) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 950.059374] env[61923]: DEBUG nova.compute.manager [-] [instance: 02a8f197-80bb-4cee-bdd6-b07705759986] Deallocating network for instance {{(pid=61923) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 950.059624] env[61923]: DEBUG nova.network.neutron [-] [instance: 02a8f197-80bb-4cee-bdd6-b07705759986] deallocate_for_instance() {{(pid=61923) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 950.220163] env[61923]: DEBUG oslo_concurrency.lockutils [None req-29a01eb2-e144-458f-9b46-c49e64b03e88 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Lock "a511392e-9ab6-42fb-b07c-f90e9786dcc1" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 19.595s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 950.221112] env[61923]: DEBUG oslo_concurrency.lockutils [None req-eea05bb2-c691-4804-8c8d-019f8feaa521 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Lock "a511392e-9ab6-42fb-b07c-f90e9786dcc1" acquired by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: waited 1.365s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 950.221299] env[61923]: INFO nova.compute.manager [None req-eea05bb2-c691-4804-8c8d-019f8feaa521 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] [instance: a511392e-9ab6-42fb-b07c-f90e9786dcc1] Unshelving [ 950.258829] env[61923]: DEBUG oslo_vmware.api [None req-102bab57-0500-4a8d-94a7-052c22c7c916 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]5251125a-a481-791b-025b-6c510dd95741, 'name': SearchDatastore_Task, 'duration_secs': 0.008784} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 950.259144] env[61923]: DEBUG oslo_concurrency.lockutils [None req-102bab57-0500-4a8d-94a7-052c22c7c916 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 950.259377] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-102bab57-0500-4a8d-94a7-052c22c7c916 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk to [datastore1] 7867c6cf-2ad6-414f-ab52-581827860836/7867c6cf-2ad6-414f-ab52-581827860836.vmdk {{(pid=61923) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 950.259681] env[61923]: DEBUG oslo_concurrency.lockutils [None req-69e9bd4b-8d1d-4678-80b0-4de2eeffdae8 tempest-ServersAaction247Test-1300676990 tempest-ServersAaction247Test-1300676990-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 950.259876] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-69e9bd4b-8d1d-4678-80b0-4de2eeffdae8 tempest-ServersAaction247Test-1300676990 tempest-ServersAaction247Test-1300676990-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61923) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 950.260111] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e6cb3e8b-7401-45fc-87b8-e359677c9394 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.263664] env[61923]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-8b78a221-55bd-4747-8282-775c0ea4abb4 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.270222] env[61923]: DEBUG oslo_vmware.api [None req-102bab57-0500-4a8d-94a7-052c22c7c916 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Waiting for the task: (returnval){ [ 950.270222] env[61923]: value = "task-1377853" [ 950.270222] env[61923]: _type = "Task" [ 950.270222] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 950.274330] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-69e9bd4b-8d1d-4678-80b0-4de2eeffdae8 tempest-ServersAaction247Test-1300676990 tempest-ServersAaction247Test-1300676990-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61923) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 950.274511] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-69e9bd4b-8d1d-4678-80b0-4de2eeffdae8 tempest-ServersAaction247Test-1300676990 tempest-ServersAaction247Test-1300676990-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61923) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 950.275643] env[61923]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-39e28f59-0eb4-4681-946a-f719cdec5e93 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.280235] env[61923]: DEBUG oslo_vmware.api [None req-102bab57-0500-4a8d-94a7-052c22c7c916 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Task: {'id': task-1377853, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.283276] env[61923]: DEBUG oslo_vmware.api [None req-69e9bd4b-8d1d-4678-80b0-4de2eeffdae8 tempest-ServersAaction247Test-1300676990 tempest-ServersAaction247Test-1300676990-project-member] Waiting for the task: (returnval){ [ 950.283276] env[61923]: value = "session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]5295971f-80c6-e99a-bcad-21f6b0a98911" [ 950.283276] env[61923]: _type = "Task" [ 950.283276] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 950.290944] env[61923]: DEBUG oslo_vmware.api [None req-69e9bd4b-8d1d-4678-80b0-4de2eeffdae8 tempest-ServersAaction247Test-1300676990 tempest-ServersAaction247Test-1300676990-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]5295971f-80c6-e99a-bcad-21f6b0a98911, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.539356] env[61923]: DEBUG nova.compute.manager [req-5e9a0e09-24eb-43a7-8810-597c9673727d req-99e644f2-33ae-4058-935f-1cdd6c64d236 service nova] [instance: aba844c1-d488-4e95-a550-bd1cb7331fb2] Received event network-vif-plugged-6326df12-72bc-44d6-b660-4abdb5ca5cfb {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 950.539814] env[61923]: DEBUG oslo_concurrency.lockutils [req-5e9a0e09-24eb-43a7-8810-597c9673727d req-99e644f2-33ae-4058-935f-1cdd6c64d236 service nova] Acquiring lock "aba844c1-d488-4e95-a550-bd1cb7331fb2-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 950.540134] env[61923]: DEBUG oslo_concurrency.lockutils [req-5e9a0e09-24eb-43a7-8810-597c9673727d req-99e644f2-33ae-4058-935f-1cdd6c64d236 service nova] Lock "aba844c1-d488-4e95-a550-bd1cb7331fb2-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 950.540298] env[61923]: DEBUG oslo_concurrency.lockutils [req-5e9a0e09-24eb-43a7-8810-597c9673727d req-99e644f2-33ae-4058-935f-1cdd6c64d236 service nova] Lock "aba844c1-d488-4e95-a550-bd1cb7331fb2-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 950.540480] env[61923]: DEBUG nova.compute.manager [req-5e9a0e09-24eb-43a7-8810-597c9673727d req-99e644f2-33ae-4058-935f-1cdd6c64d236 service nova] [instance: aba844c1-d488-4e95-a550-bd1cb7331fb2] No waiting events found dispatching network-vif-plugged-6326df12-72bc-44d6-b660-4abdb5ca5cfb {{(pid=61923) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 950.540700] env[61923]: WARNING nova.compute.manager [req-5e9a0e09-24eb-43a7-8810-597c9673727d req-99e644f2-33ae-4058-935f-1cdd6c64d236 service nova] [instance: aba844c1-d488-4e95-a550-bd1cb7331fb2] Received unexpected event network-vif-plugged-6326df12-72bc-44d6-b660-4abdb5ca5cfb for instance with vm_state building and task_state spawning. [ 950.636079] env[61923]: DEBUG nova.compute.manager [req-6b1a8421-ac7d-42c1-ad5e-338054e0f4a0 req-8b9ce3b6-4028-4301-b4eb-181f88852b48 service nova] [instance: 02a8f197-80bb-4cee-bdd6-b07705759986] Received event network-vif-deleted-0b0df455-855c-4e3f-832c-4b1c688aa252 {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 950.636211] env[61923]: INFO nova.compute.manager [req-6b1a8421-ac7d-42c1-ad5e-338054e0f4a0 req-8b9ce3b6-4028-4301-b4eb-181f88852b48 service nova] [instance: 02a8f197-80bb-4cee-bdd6-b07705759986] Neutron deleted interface 0b0df455-855c-4e3f-832c-4b1c688aa252; detaching it from the instance and deleting it from the info cache [ 950.636395] env[61923]: DEBUG nova.network.neutron [req-6b1a8421-ac7d-42c1-ad5e-338054e0f4a0 req-8b9ce3b6-4028-4301-b4eb-181f88852b48 service nova] [instance: 02a8f197-80bb-4cee-bdd6-b07705759986] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 950.781306] env[61923]: DEBUG oslo_vmware.api [None req-102bab57-0500-4a8d-94a7-052c22c7c916 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Task: {'id': task-1377853, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.794901] env[61923]: DEBUG oslo_vmware.api [None req-69e9bd4b-8d1d-4678-80b0-4de2eeffdae8 tempest-ServersAaction247Test-1300676990 tempest-ServersAaction247Test-1300676990-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]5295971f-80c6-e99a-bcad-21f6b0a98911, 'name': SearchDatastore_Task, 'duration_secs': 0.007767} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 950.796730] env[61923]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e8137fcf-ef4c-40cb-9b5c-48c4f2a2f0e3 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.803438] env[61923]: DEBUG oslo_vmware.api [None req-69e9bd4b-8d1d-4678-80b0-4de2eeffdae8 tempest-ServersAaction247Test-1300676990 tempest-ServersAaction247Test-1300676990-project-member] Waiting for the task: (returnval){ [ 950.803438] env[61923]: value = "session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52bcc3df-2a22-6a88-78e1-2b4b79e940a5" [ 950.803438] env[61923]: _type = "Task" [ 950.803438] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 950.811924] env[61923]: DEBUG oslo_vmware.api [None req-69e9bd4b-8d1d-4678-80b0-4de2eeffdae8 tempest-ServersAaction247Test-1300676990 tempest-ServersAaction247Test-1300676990-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52bcc3df-2a22-6a88-78e1-2b4b79e940a5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.103379] env[61923]: DEBUG nova.network.neutron [None req-704a26eb-9a13-4b75-8212-d4cbe21aa5e4 tempest-ServersV294TestFqdnHostnames-833840938 tempest-ServersV294TestFqdnHostnames-833840938-project-member] [instance: aba844c1-d488-4e95-a550-bd1cb7331fb2] Successfully updated port: 6326df12-72bc-44d6-b660-4abdb5ca5cfb {{(pid=61923) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 951.110951] env[61923]: DEBUG nova.network.neutron [-] [instance: 02a8f197-80bb-4cee-bdd6-b07705759986] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 951.138770] env[61923]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-6c4c3aa0-1b3c-44f6-bd3f-dc32f8689087 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.147998] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-afebcee7-02d7-4931-b709-f7e0937fd821 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.173326] env[61923]: DEBUG nova.compute.manager [req-6b1a8421-ac7d-42c1-ad5e-338054e0f4a0 req-8b9ce3b6-4028-4301-b4eb-181f88852b48 service nova] [instance: 02a8f197-80bb-4cee-bdd6-b07705759986] Detach interface failed, port_id=0b0df455-855c-4e3f-832c-4b1c688aa252, reason: Instance 02a8f197-80bb-4cee-bdd6-b07705759986 could not be found. {{(pid=61923) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 951.252554] env[61923]: DEBUG oslo_concurrency.lockutils [None req-eea05bb2-c691-4804-8c8d-019f8feaa521 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 951.252924] env[61923]: DEBUG oslo_concurrency.lockutils [None req-eea05bb2-c691-4804-8c8d-019f8feaa521 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 951.253197] env[61923]: DEBUG nova.objects.instance [None req-eea05bb2-c691-4804-8c8d-019f8feaa521 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Lazy-loading 'pci_requests' on Instance uuid a511392e-9ab6-42fb-b07c-f90e9786dcc1 {{(pid=61923) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 951.281671] env[61923]: DEBUG oslo_vmware.api [None req-102bab57-0500-4a8d-94a7-052c22c7c916 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Task: {'id': task-1377853, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.547747} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 951.281671] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-102bab57-0500-4a8d-94a7-052c22c7c916 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk to [datastore1] 7867c6cf-2ad6-414f-ab52-581827860836/7867c6cf-2ad6-414f-ab52-581827860836.vmdk {{(pid=61923) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 951.281671] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-102bab57-0500-4a8d-94a7-052c22c7c916 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: 7867c6cf-2ad6-414f-ab52-581827860836] Extending root virtual disk to 1048576 {{(pid=61923) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 951.282119] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-2ff53431-38da-4ed0-92c1-1f8f42d13b82 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.287943] env[61923]: DEBUG oslo_vmware.api [None req-102bab57-0500-4a8d-94a7-052c22c7c916 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Waiting for the task: (returnval){ [ 951.287943] env[61923]: value = "task-1377854" [ 951.287943] env[61923]: _type = "Task" [ 951.287943] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 951.295351] env[61923]: DEBUG oslo_vmware.api [None req-102bab57-0500-4a8d-94a7-052c22c7c916 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Task: {'id': task-1377854, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.312125] env[61923]: DEBUG oslo_vmware.api [None req-69e9bd4b-8d1d-4678-80b0-4de2eeffdae8 tempest-ServersAaction247Test-1300676990 tempest-ServersAaction247Test-1300676990-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52bcc3df-2a22-6a88-78e1-2b4b79e940a5, 'name': SearchDatastore_Task, 'duration_secs': 0.011867} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 951.312377] env[61923]: DEBUG oslo_concurrency.lockutils [None req-69e9bd4b-8d1d-4678-80b0-4de2eeffdae8 tempest-ServersAaction247Test-1300676990 tempest-ServersAaction247Test-1300676990-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 951.312636] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-69e9bd4b-8d1d-4678-80b0-4de2eeffdae8 tempest-ServersAaction247Test-1300676990 tempest-ServersAaction247Test-1300676990-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk to [datastore1] 44196b4c-1401-40fa-bd14-04a49947ab15/44196b4c-1401-40fa-bd14-04a49947ab15.vmdk {{(pid=61923) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 951.312884] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a99f4cc3-4df6-4ead-9aba-b13420447a5c {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.318524] env[61923]: DEBUG oslo_vmware.api [None req-69e9bd4b-8d1d-4678-80b0-4de2eeffdae8 tempest-ServersAaction247Test-1300676990 tempest-ServersAaction247Test-1300676990-project-member] Waiting for the task: (returnval){ [ 951.318524] env[61923]: value = "task-1377855" [ 951.318524] env[61923]: _type = "Task" [ 951.318524] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 951.325887] env[61923]: DEBUG oslo_vmware.api [None req-69e9bd4b-8d1d-4678-80b0-4de2eeffdae8 tempest-ServersAaction247Test-1300676990 tempest-ServersAaction247Test-1300676990-project-member] Task: {'id': task-1377855, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.608272] env[61923]: DEBUG oslo_concurrency.lockutils [None req-704a26eb-9a13-4b75-8212-d4cbe21aa5e4 tempest-ServersV294TestFqdnHostnames-833840938 tempest-ServersV294TestFqdnHostnames-833840938-project-member] Acquiring lock "refresh_cache-aba844c1-d488-4e95-a550-bd1cb7331fb2" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 951.608818] env[61923]: DEBUG oslo_concurrency.lockutils [None req-704a26eb-9a13-4b75-8212-d4cbe21aa5e4 tempest-ServersV294TestFqdnHostnames-833840938 tempest-ServersV294TestFqdnHostnames-833840938-project-member] Acquired lock "refresh_cache-aba844c1-d488-4e95-a550-bd1cb7331fb2" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 951.608818] env[61923]: DEBUG nova.network.neutron [None req-704a26eb-9a13-4b75-8212-d4cbe21aa5e4 tempest-ServersV294TestFqdnHostnames-833840938 tempest-ServersV294TestFqdnHostnames-833840938-project-member] [instance: aba844c1-d488-4e95-a550-bd1cb7331fb2] Building network info cache for instance {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 951.612706] env[61923]: INFO nova.compute.manager [-] [instance: 02a8f197-80bb-4cee-bdd6-b07705759986] Took 1.55 seconds to deallocate network for instance. [ 951.757326] env[61923]: DEBUG nova.objects.instance [None req-eea05bb2-c691-4804-8c8d-019f8feaa521 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Lazy-loading 'numa_topology' on Instance uuid a511392e-9ab6-42fb-b07c-f90e9786dcc1 {{(pid=61923) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 951.797792] env[61923]: DEBUG oslo_vmware.api [None req-102bab57-0500-4a8d-94a7-052c22c7c916 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Task: {'id': task-1377854, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.06435} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 951.798101] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-102bab57-0500-4a8d-94a7-052c22c7c916 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: 7867c6cf-2ad6-414f-ab52-581827860836] Extended root virtual disk {{(pid=61923) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 951.798973] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23cdbeda-7bd0-439b-ae46-1f49a123b9d8 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.821749] env[61923]: DEBUG nova.virt.vmwareapi.volumeops [None req-102bab57-0500-4a8d-94a7-052c22c7c916 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: 7867c6cf-2ad6-414f-ab52-581827860836] Reconfiguring VM instance instance-00000059 to attach disk [datastore1] 7867c6cf-2ad6-414f-ab52-581827860836/7867c6cf-2ad6-414f-ab52-581827860836.vmdk or device None with type sparse {{(pid=61923) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 951.822156] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1d9c0eab-d7f8-40dd-8c19-8d209d1bd5df {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.845810] env[61923]: DEBUG oslo_vmware.api [None req-69e9bd4b-8d1d-4678-80b0-4de2eeffdae8 tempest-ServersAaction247Test-1300676990 tempest-ServersAaction247Test-1300676990-project-member] Task: {'id': task-1377855, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.847330] env[61923]: DEBUG oslo_vmware.api [None req-102bab57-0500-4a8d-94a7-052c22c7c916 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Waiting for the task: (returnval){ [ 951.847330] env[61923]: value = "task-1377856" [ 951.847330] env[61923]: _type = "Task" [ 951.847330] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 951.855464] env[61923]: DEBUG oslo_vmware.api [None req-102bab57-0500-4a8d-94a7-052c22c7c916 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Task: {'id': task-1377856, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.119305] env[61923]: DEBUG oslo_concurrency.lockutils [None req-af395613-edaf-421a-b79b-9579e1c9e371 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 952.142421] env[61923]: DEBUG nova.network.neutron [None req-704a26eb-9a13-4b75-8212-d4cbe21aa5e4 tempest-ServersV294TestFqdnHostnames-833840938 tempest-ServersV294TestFqdnHostnames-833840938-project-member] [instance: aba844c1-d488-4e95-a550-bd1cb7331fb2] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 952.260199] env[61923]: INFO nova.compute.claims [None req-eea05bb2-c691-4804-8c8d-019f8feaa521 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] [instance: a511392e-9ab6-42fb-b07c-f90e9786dcc1] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 952.280329] env[61923]: DEBUG nova.network.neutron [None req-704a26eb-9a13-4b75-8212-d4cbe21aa5e4 tempest-ServersV294TestFqdnHostnames-833840938 tempest-ServersV294TestFqdnHostnames-833840938-project-member] [instance: aba844c1-d488-4e95-a550-bd1cb7331fb2] Updating instance_info_cache with network_info: [{"id": "6326df12-72bc-44d6-b660-4abdb5ca5cfb", "address": "fa:16:3e:dd:2b:a6", "network": {"id": "49b0b81e-2f1d-42ae-bc27-de297b6bef9e", "bridge": "br-int", "label": "tempest-ServersV294TestFqdnHostnames-151433926-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e9e5485c43514e55b88bdd7d2dc2aee7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d275d7c6-2a7b-4ee8-b6f4-fabf1ba1905f", "external-id": "nsx-vlan-transportzone-513", "segmentation_id": 513, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6326df12-72", "ovs_interfaceid": "6326df12-72bc-44d6-b660-4abdb5ca5cfb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 952.332342] env[61923]: DEBUG oslo_vmware.api [None req-69e9bd4b-8d1d-4678-80b0-4de2eeffdae8 tempest-ServersAaction247Test-1300676990 tempest-ServersAaction247Test-1300676990-project-member] Task: {'id': task-1377855, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.679252} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 952.332652] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-69e9bd4b-8d1d-4678-80b0-4de2eeffdae8 tempest-ServersAaction247Test-1300676990 tempest-ServersAaction247Test-1300676990-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk to [datastore1] 44196b4c-1401-40fa-bd14-04a49947ab15/44196b4c-1401-40fa-bd14-04a49947ab15.vmdk {{(pid=61923) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 952.332815] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-69e9bd4b-8d1d-4678-80b0-4de2eeffdae8 tempest-ServersAaction247Test-1300676990 tempest-ServersAaction247Test-1300676990-project-member] [instance: 44196b4c-1401-40fa-bd14-04a49947ab15] Extending root virtual disk to 1048576 {{(pid=61923) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 952.333079] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-e6ebc1c6-d16a-45f6-a81c-01d63753c6f9 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.339407] env[61923]: DEBUG oslo_vmware.api [None req-69e9bd4b-8d1d-4678-80b0-4de2eeffdae8 tempest-ServersAaction247Test-1300676990 tempest-ServersAaction247Test-1300676990-project-member] Waiting for the task: (returnval){ [ 952.339407] env[61923]: value = "task-1377857" [ 952.339407] env[61923]: _type = "Task" [ 952.339407] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 952.349475] env[61923]: DEBUG oslo_vmware.api [None req-69e9bd4b-8d1d-4678-80b0-4de2eeffdae8 tempest-ServersAaction247Test-1300676990 tempest-ServersAaction247Test-1300676990-project-member] Task: {'id': task-1377857, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.357137] env[61923]: DEBUG oslo_vmware.api [None req-102bab57-0500-4a8d-94a7-052c22c7c916 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Task: {'id': task-1377856, 'name': ReconfigVM_Task, 'duration_secs': 0.391231} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 952.357806] env[61923]: DEBUG nova.virt.vmwareapi.volumeops [None req-102bab57-0500-4a8d-94a7-052c22c7c916 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: 7867c6cf-2ad6-414f-ab52-581827860836] Reconfigured VM instance instance-00000059 to attach disk [datastore1] 7867c6cf-2ad6-414f-ab52-581827860836/7867c6cf-2ad6-414f-ab52-581827860836.vmdk or device None with type sparse {{(pid=61923) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 952.358061] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-9ede4243-f7e4-4a01-9199-ff8a88654ebf {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.364316] env[61923]: DEBUG oslo_vmware.api [None req-102bab57-0500-4a8d-94a7-052c22c7c916 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Waiting for the task: (returnval){ [ 952.364316] env[61923]: value = "task-1377858" [ 952.364316] env[61923]: _type = "Task" [ 952.364316] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 952.372120] env[61923]: DEBUG oslo_vmware.api [None req-102bab57-0500-4a8d-94a7-052c22c7c916 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Task: {'id': task-1377858, 'name': Rename_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.661214] env[61923]: DEBUG nova.compute.manager [req-d39da0aa-386c-4c04-841e-a5f14bf019a4 req-14b8c8b9-af46-488b-a625-a5ad0dcccc15 service nova] [instance: aba844c1-d488-4e95-a550-bd1cb7331fb2] Received event network-changed-6326df12-72bc-44d6-b660-4abdb5ca5cfb {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 952.661516] env[61923]: DEBUG nova.compute.manager [req-d39da0aa-386c-4c04-841e-a5f14bf019a4 req-14b8c8b9-af46-488b-a625-a5ad0dcccc15 service nova] [instance: aba844c1-d488-4e95-a550-bd1cb7331fb2] Refreshing instance network info cache due to event network-changed-6326df12-72bc-44d6-b660-4abdb5ca5cfb. {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 952.661879] env[61923]: DEBUG oslo_concurrency.lockutils [req-d39da0aa-386c-4c04-841e-a5f14bf019a4 req-14b8c8b9-af46-488b-a625-a5ad0dcccc15 service nova] Acquiring lock "refresh_cache-aba844c1-d488-4e95-a550-bd1cb7331fb2" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 952.783306] env[61923]: DEBUG oslo_concurrency.lockutils [None req-704a26eb-9a13-4b75-8212-d4cbe21aa5e4 tempest-ServersV294TestFqdnHostnames-833840938 tempest-ServersV294TestFqdnHostnames-833840938-project-member] Releasing lock "refresh_cache-aba844c1-d488-4e95-a550-bd1cb7331fb2" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 952.783772] env[61923]: DEBUG nova.compute.manager [None req-704a26eb-9a13-4b75-8212-d4cbe21aa5e4 tempest-ServersV294TestFqdnHostnames-833840938 tempest-ServersV294TestFqdnHostnames-833840938-project-member] [instance: aba844c1-d488-4e95-a550-bd1cb7331fb2] Instance network_info: |[{"id": "6326df12-72bc-44d6-b660-4abdb5ca5cfb", "address": "fa:16:3e:dd:2b:a6", "network": {"id": "49b0b81e-2f1d-42ae-bc27-de297b6bef9e", "bridge": "br-int", "label": "tempest-ServersV294TestFqdnHostnames-151433926-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e9e5485c43514e55b88bdd7d2dc2aee7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d275d7c6-2a7b-4ee8-b6f4-fabf1ba1905f", "external-id": "nsx-vlan-transportzone-513", "segmentation_id": 513, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6326df12-72", "ovs_interfaceid": "6326df12-72bc-44d6-b660-4abdb5ca5cfb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61923) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 952.784135] env[61923]: DEBUG oslo_concurrency.lockutils [req-d39da0aa-386c-4c04-841e-a5f14bf019a4 req-14b8c8b9-af46-488b-a625-a5ad0dcccc15 service nova] Acquired lock "refresh_cache-aba844c1-d488-4e95-a550-bd1cb7331fb2" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 952.784340] env[61923]: DEBUG nova.network.neutron [req-d39da0aa-386c-4c04-841e-a5f14bf019a4 req-14b8c8b9-af46-488b-a625-a5ad0dcccc15 service nova] [instance: aba844c1-d488-4e95-a550-bd1cb7331fb2] Refreshing network info cache for port 6326df12-72bc-44d6-b660-4abdb5ca5cfb {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 952.785529] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-704a26eb-9a13-4b75-8212-d4cbe21aa5e4 tempest-ServersV294TestFqdnHostnames-833840938 tempest-ServersV294TestFqdnHostnames-833840938-project-member] [instance: aba844c1-d488-4e95-a550-bd1cb7331fb2] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:dd:2b:a6', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'd275d7c6-2a7b-4ee8-b6f4-fabf1ba1905f', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '6326df12-72bc-44d6-b660-4abdb5ca5cfb', 'vif_model': 'vmxnet3'}] {{(pid=61923) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 952.793487] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-704a26eb-9a13-4b75-8212-d4cbe21aa5e4 tempest-ServersV294TestFqdnHostnames-833840938 tempest-ServersV294TestFqdnHostnames-833840938-project-member] Creating folder: Project (e9e5485c43514e55b88bdd7d2dc2aee7). Parent ref: group-v292629. {{(pid=61923) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 952.794736] env[61923]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-458f2194-6321-4d6b-b91e-3577ee500f85 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.806636] env[61923]: INFO nova.virt.vmwareapi.vm_util [None req-704a26eb-9a13-4b75-8212-d4cbe21aa5e4 tempest-ServersV294TestFqdnHostnames-833840938 tempest-ServersV294TestFqdnHostnames-833840938-project-member] Created folder: Project (e9e5485c43514e55b88bdd7d2dc2aee7) in parent group-v292629. [ 952.806831] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-704a26eb-9a13-4b75-8212-d4cbe21aa5e4 tempest-ServersV294TestFqdnHostnames-833840938 tempest-ServersV294TestFqdnHostnames-833840938-project-member] Creating folder: Instances. Parent ref: group-v292731. {{(pid=61923) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 952.807076] env[61923]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-3f2de045-12f8-43cd-b5fd-192205a5ca9b {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.816106] env[61923]: INFO nova.virt.vmwareapi.vm_util [None req-704a26eb-9a13-4b75-8212-d4cbe21aa5e4 tempest-ServersV294TestFqdnHostnames-833840938 tempest-ServersV294TestFqdnHostnames-833840938-project-member] Created folder: Instances in parent group-v292731. [ 952.816335] env[61923]: DEBUG oslo.service.loopingcall [None req-704a26eb-9a13-4b75-8212-d4cbe21aa5e4 tempest-ServersV294TestFqdnHostnames-833840938 tempest-ServersV294TestFqdnHostnames-833840938-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61923) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 952.816519] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: aba844c1-d488-4e95-a550-bd1cb7331fb2] Creating VM on the ESX host {{(pid=61923) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 952.816713] env[61923]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-a5b620c5-a8d1-4cce-96ed-3a5877891802 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.834521] env[61923]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 952.834521] env[61923]: value = "task-1377861" [ 952.834521] env[61923]: _type = "Task" [ 952.834521] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 952.842351] env[61923]: DEBUG oslo_vmware.api [-] Task: {'id': task-1377861, 'name': CreateVM_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.849196] env[61923]: DEBUG oslo_vmware.api [None req-69e9bd4b-8d1d-4678-80b0-4de2eeffdae8 tempest-ServersAaction247Test-1300676990 tempest-ServersAaction247Test-1300676990-project-member] Task: {'id': task-1377857, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.154541} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 952.849469] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-69e9bd4b-8d1d-4678-80b0-4de2eeffdae8 tempest-ServersAaction247Test-1300676990 tempest-ServersAaction247Test-1300676990-project-member] [instance: 44196b4c-1401-40fa-bd14-04a49947ab15] Extended root virtual disk {{(pid=61923) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 952.850206] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ddedf4dd-6925-40b9-9fad-cf37e368a2ec {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.871640] env[61923]: DEBUG nova.virt.vmwareapi.volumeops [None req-69e9bd4b-8d1d-4678-80b0-4de2eeffdae8 tempest-ServersAaction247Test-1300676990 tempest-ServersAaction247Test-1300676990-project-member] [instance: 44196b4c-1401-40fa-bd14-04a49947ab15] Reconfiguring VM instance instance-0000005a to attach disk [datastore1] 44196b4c-1401-40fa-bd14-04a49947ab15/44196b4c-1401-40fa-bd14-04a49947ab15.vmdk or device None with type sparse {{(pid=61923) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 952.872953] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ca4aed2f-1f8a-44e4-bed6-dc2550e13080 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.897135] env[61923]: DEBUG oslo_vmware.api [None req-102bab57-0500-4a8d-94a7-052c22c7c916 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Task: {'id': task-1377858, 'name': Rename_Task, 'duration_secs': 0.192056} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 952.901019] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-102bab57-0500-4a8d-94a7-052c22c7c916 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: 7867c6cf-2ad6-414f-ab52-581827860836] Powering on the VM {{(pid=61923) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 952.901019] env[61923]: DEBUG oslo_vmware.api [None req-69e9bd4b-8d1d-4678-80b0-4de2eeffdae8 tempest-ServersAaction247Test-1300676990 tempest-ServersAaction247Test-1300676990-project-member] Waiting for the task: (returnval){ [ 952.901019] env[61923]: value = "task-1377862" [ 952.901019] env[61923]: _type = "Task" [ 952.901019] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 952.901019] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-54a74903-11a3-4481-9f71-a3cd0f7b9f17 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.908184] env[61923]: DEBUG oslo_vmware.api [None req-102bab57-0500-4a8d-94a7-052c22c7c916 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Waiting for the task: (returnval){ [ 952.908184] env[61923]: value = "task-1377863" [ 952.908184] env[61923]: _type = "Task" [ 952.908184] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 952.911857] env[61923]: DEBUG oslo_vmware.api [None req-69e9bd4b-8d1d-4678-80b0-4de2eeffdae8 tempest-ServersAaction247Test-1300676990 tempest-ServersAaction247Test-1300676990-project-member] Task: {'id': task-1377862, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.919153] env[61923]: DEBUG oslo_vmware.api [None req-102bab57-0500-4a8d-94a7-052c22c7c916 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Task: {'id': task-1377863, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.345345] env[61923]: DEBUG oslo_vmware.api [-] Task: {'id': task-1377861, 'name': CreateVM_Task, 'duration_secs': 0.358028} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 953.346224] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: aba844c1-d488-4e95-a550-bd1cb7331fb2] Created VM on the ESX host {{(pid=61923) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 953.346961] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-633bca7b-5e2a-4430-b39d-209583420c87 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.349738] env[61923]: DEBUG oslo_concurrency.lockutils [None req-704a26eb-9a13-4b75-8212-d4cbe21aa5e4 tempest-ServersV294TestFqdnHostnames-833840938 tempest-ServersV294TestFqdnHostnames-833840938-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 953.349904] env[61923]: DEBUG oslo_concurrency.lockutils [None req-704a26eb-9a13-4b75-8212-d4cbe21aa5e4 tempest-ServersV294TestFqdnHostnames-833840938 tempest-ServersV294TestFqdnHostnames-833840938-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 953.350232] env[61923]: DEBUG oslo_concurrency.lockutils [None req-704a26eb-9a13-4b75-8212-d4cbe21aa5e4 tempest-ServersV294TestFqdnHostnames-833840938 tempest-ServersV294TestFqdnHostnames-833840938-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 953.350461] env[61923]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-aa4b6ec0-af30-4181-856a-099e7079fe04 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.355320] env[61923]: DEBUG oslo_vmware.api [None req-704a26eb-9a13-4b75-8212-d4cbe21aa5e4 tempest-ServersV294TestFqdnHostnames-833840938 tempest-ServersV294TestFqdnHostnames-833840938-project-member] Waiting for the task: (returnval){ [ 953.355320] env[61923]: value = "session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52b81269-d145-aeb1-9042-d234beb577d7" [ 953.355320] env[61923]: _type = "Task" [ 953.355320] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 953.362178] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad4875d2-c7b0-492b-8fbf-cfcbb4fdb296 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.370997] env[61923]: DEBUG oslo_vmware.api [None req-704a26eb-9a13-4b75-8212-d4cbe21aa5e4 tempest-ServersV294TestFqdnHostnames-833840938 tempest-ServersV294TestFqdnHostnames-833840938-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52b81269-d145-aeb1-9042-d234beb577d7, 'name': SearchDatastore_Task, 'duration_secs': 0.010316} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 953.394213] env[61923]: DEBUG oslo_concurrency.lockutils [None req-704a26eb-9a13-4b75-8212-d4cbe21aa5e4 tempest-ServersV294TestFqdnHostnames-833840938 tempest-ServersV294TestFqdnHostnames-833840938-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 953.394469] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-704a26eb-9a13-4b75-8212-d4cbe21aa5e4 tempest-ServersV294TestFqdnHostnames-833840938 tempest-ServersV294TestFqdnHostnames-833840938-project-member] [instance: aba844c1-d488-4e95-a550-bd1cb7331fb2] Processing image 0f153f63-ae0a-45b1-b7f5-7f9b673c947f {{(pid=61923) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 953.394707] env[61923]: DEBUG oslo_concurrency.lockutils [None req-704a26eb-9a13-4b75-8212-d4cbe21aa5e4 tempest-ServersV294TestFqdnHostnames-833840938 tempest-ServersV294TestFqdnHostnames-833840938-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 953.394862] env[61923]: DEBUG oslo_concurrency.lockutils [None req-704a26eb-9a13-4b75-8212-d4cbe21aa5e4 tempest-ServersV294TestFqdnHostnames-833840938 tempest-ServersV294TestFqdnHostnames-833840938-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 953.395053] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-704a26eb-9a13-4b75-8212-d4cbe21aa5e4 tempest-ServersV294TestFqdnHostnames-833840938 tempest-ServersV294TestFqdnHostnames-833840938-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61923) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 953.397384] env[61923]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d2b25cf0-30f4-4739-86fc-91b1e1a6bc02 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.399486] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33e9c4a9-cc49-474e-8582-4013c32685b8 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.412268] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22c37dea-ed13-49a4-8766-61c787e9316d {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.416058] env[61923]: DEBUG oslo_vmware.api [None req-69e9bd4b-8d1d-4678-80b0-4de2eeffdae8 tempest-ServersAaction247Test-1300676990 tempest-ServersAaction247Test-1300676990-project-member] Task: {'id': task-1377862, 'name': ReconfigVM_Task, 'duration_secs': 0.358846} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 953.416179] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-704a26eb-9a13-4b75-8212-d4cbe21aa5e4 tempest-ServersV294TestFqdnHostnames-833840938 tempest-ServersV294TestFqdnHostnames-833840938-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61923) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 953.416293] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-704a26eb-9a13-4b75-8212-d4cbe21aa5e4 tempest-ServersV294TestFqdnHostnames-833840938 tempest-ServersV294TestFqdnHostnames-833840938-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61923) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 953.417051] env[61923]: DEBUG nova.virt.vmwareapi.volumeops [None req-69e9bd4b-8d1d-4678-80b0-4de2eeffdae8 tempest-ServersAaction247Test-1300676990 tempest-ServersAaction247Test-1300676990-project-member] [instance: 44196b4c-1401-40fa-bd14-04a49947ab15] Reconfigured VM instance instance-0000005a to attach disk [datastore1] 44196b4c-1401-40fa-bd14-04a49947ab15/44196b4c-1401-40fa-bd14-04a49947ab15.vmdk or device None with type sparse {{(pid=61923) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 953.417603] env[61923]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4cf9beea-0ebc-4b73-a9d0-232e9336c966 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.422113] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-0c000e56-16c6-404a-b384-2e94206cba37 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.431106] env[61923]: DEBUG nova.compute.provider_tree [None req-eea05bb2-c691-4804-8c8d-019f8feaa521 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 953.435546] env[61923]: DEBUG oslo_vmware.api [None req-704a26eb-9a13-4b75-8212-d4cbe21aa5e4 tempest-ServersV294TestFqdnHostnames-833840938 tempest-ServersV294TestFqdnHostnames-833840938-project-member] Waiting for the task: (returnval){ [ 953.435546] env[61923]: value = "session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52f7e994-c239-f701-2eb4-9212c4e4ab54" [ 953.435546] env[61923]: _type = "Task" [ 953.435546] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 953.443335] env[61923]: DEBUG oslo_vmware.api [None req-102bab57-0500-4a8d-94a7-052c22c7c916 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Task: {'id': task-1377863, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.443623] env[61923]: DEBUG oslo_vmware.api [None req-69e9bd4b-8d1d-4678-80b0-4de2eeffdae8 tempest-ServersAaction247Test-1300676990 tempest-ServersAaction247Test-1300676990-project-member] Waiting for the task: (returnval){ [ 953.443623] env[61923]: value = "task-1377864" [ 953.443623] env[61923]: _type = "Task" [ 953.443623] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 953.450681] env[61923]: DEBUG oslo_vmware.api [None req-704a26eb-9a13-4b75-8212-d4cbe21aa5e4 tempest-ServersV294TestFqdnHostnames-833840938 tempest-ServersV294TestFqdnHostnames-833840938-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52f7e994-c239-f701-2eb4-9212c4e4ab54, 'name': SearchDatastore_Task, 'duration_secs': 0.009903} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 953.451682] env[61923]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ff5b9086-fb08-4588-8992-0b714d1273e5 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.456503] env[61923]: DEBUG oslo_vmware.api [None req-69e9bd4b-8d1d-4678-80b0-4de2eeffdae8 tempest-ServersAaction247Test-1300676990 tempest-ServersAaction247Test-1300676990-project-member] Task: {'id': task-1377864, 'name': Rename_Task} progress is 10%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.459340] env[61923]: DEBUG oslo_vmware.api [None req-704a26eb-9a13-4b75-8212-d4cbe21aa5e4 tempest-ServersV294TestFqdnHostnames-833840938 tempest-ServersV294TestFqdnHostnames-833840938-project-member] Waiting for the task: (returnval){ [ 953.459340] env[61923]: value = "session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]5289cb6c-702b-7ecd-8052-c9591e74a1c9" [ 953.459340] env[61923]: _type = "Task" [ 953.459340] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 953.471481] env[61923]: DEBUG oslo_vmware.api [None req-704a26eb-9a13-4b75-8212-d4cbe21aa5e4 tempest-ServersV294TestFqdnHostnames-833840938 tempest-ServersV294TestFqdnHostnames-833840938-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]5289cb6c-702b-7ecd-8052-c9591e74a1c9, 'name': SearchDatastore_Task, 'duration_secs': 0.009339} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 953.471750] env[61923]: DEBUG oslo_concurrency.lockutils [None req-704a26eb-9a13-4b75-8212-d4cbe21aa5e4 tempest-ServersV294TestFqdnHostnames-833840938 tempest-ServersV294TestFqdnHostnames-833840938-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 953.472017] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-704a26eb-9a13-4b75-8212-d4cbe21aa5e4 tempest-ServersV294TestFqdnHostnames-833840938 tempest-ServersV294TestFqdnHostnames-833840938-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk to [datastore1] aba844c1-d488-4e95-a550-bd1cb7331fb2/aba844c1-d488-4e95-a550-bd1cb7331fb2.vmdk {{(pid=61923) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 953.472295] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-3646614f-ab17-482a-ae03-252164494436 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.478257] env[61923]: DEBUG oslo_vmware.api [None req-704a26eb-9a13-4b75-8212-d4cbe21aa5e4 tempest-ServersV294TestFqdnHostnames-833840938 tempest-ServersV294TestFqdnHostnames-833840938-project-member] Waiting for the task: (returnval){ [ 953.478257] env[61923]: value = "task-1377865" [ 953.478257] env[61923]: _type = "Task" [ 953.478257] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 953.485724] env[61923]: DEBUG oslo_vmware.api [None req-704a26eb-9a13-4b75-8212-d4cbe21aa5e4 tempest-ServersV294TestFqdnHostnames-833840938 tempest-ServersV294TestFqdnHostnames-833840938-project-member] Task: {'id': task-1377865, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.545547] env[61923]: DEBUG nova.network.neutron [req-d39da0aa-386c-4c04-841e-a5f14bf019a4 req-14b8c8b9-af46-488b-a625-a5ad0dcccc15 service nova] [instance: aba844c1-d488-4e95-a550-bd1cb7331fb2] Updated VIF entry in instance network info cache for port 6326df12-72bc-44d6-b660-4abdb5ca5cfb. {{(pid=61923) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 953.546016] env[61923]: DEBUG nova.network.neutron [req-d39da0aa-386c-4c04-841e-a5f14bf019a4 req-14b8c8b9-af46-488b-a625-a5ad0dcccc15 service nova] [instance: aba844c1-d488-4e95-a550-bd1cb7331fb2] Updating instance_info_cache with network_info: [{"id": "6326df12-72bc-44d6-b660-4abdb5ca5cfb", "address": "fa:16:3e:dd:2b:a6", "network": {"id": "49b0b81e-2f1d-42ae-bc27-de297b6bef9e", "bridge": "br-int", "label": "tempest-ServersV294TestFqdnHostnames-151433926-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e9e5485c43514e55b88bdd7d2dc2aee7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d275d7c6-2a7b-4ee8-b6f4-fabf1ba1905f", "external-id": "nsx-vlan-transportzone-513", "segmentation_id": 513, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6326df12-72", "ovs_interfaceid": "6326df12-72bc-44d6-b660-4abdb5ca5cfb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 953.927284] env[61923]: DEBUG oslo_vmware.api [None req-102bab57-0500-4a8d-94a7-052c22c7c916 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Task: {'id': task-1377863, 'name': PowerOnVM_Task} progress is 78%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.937601] env[61923]: DEBUG nova.scheduler.client.report [None req-eea05bb2-c691-4804-8c8d-019f8feaa521 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 953.954954] env[61923]: DEBUG oslo_vmware.api [None req-69e9bd4b-8d1d-4678-80b0-4de2eeffdae8 tempest-ServersAaction247Test-1300676990 tempest-ServersAaction247Test-1300676990-project-member] Task: {'id': task-1377864, 'name': Rename_Task, 'duration_secs': 0.126054} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 953.955284] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-69e9bd4b-8d1d-4678-80b0-4de2eeffdae8 tempest-ServersAaction247Test-1300676990 tempest-ServersAaction247Test-1300676990-project-member] [instance: 44196b4c-1401-40fa-bd14-04a49947ab15] Powering on the VM {{(pid=61923) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 953.955548] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1cdeed28-0810-4e78-8bc7-c9aa3d90ba48 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.963196] env[61923]: DEBUG oslo_vmware.api [None req-69e9bd4b-8d1d-4678-80b0-4de2eeffdae8 tempest-ServersAaction247Test-1300676990 tempest-ServersAaction247Test-1300676990-project-member] Waiting for the task: (returnval){ [ 953.963196] env[61923]: value = "task-1377866" [ 953.963196] env[61923]: _type = "Task" [ 953.963196] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 953.971449] env[61923]: DEBUG oslo_vmware.api [None req-69e9bd4b-8d1d-4678-80b0-4de2eeffdae8 tempest-ServersAaction247Test-1300676990 tempest-ServersAaction247Test-1300676990-project-member] Task: {'id': task-1377866, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.986870] env[61923]: DEBUG oslo_vmware.api [None req-704a26eb-9a13-4b75-8212-d4cbe21aa5e4 tempest-ServersV294TestFqdnHostnames-833840938 tempest-ServersV294TestFqdnHostnames-833840938-project-member] Task: {'id': task-1377865, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.50581} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 953.987107] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-704a26eb-9a13-4b75-8212-d4cbe21aa5e4 tempest-ServersV294TestFqdnHostnames-833840938 tempest-ServersV294TestFqdnHostnames-833840938-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk to [datastore1] aba844c1-d488-4e95-a550-bd1cb7331fb2/aba844c1-d488-4e95-a550-bd1cb7331fb2.vmdk {{(pid=61923) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 953.987321] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-704a26eb-9a13-4b75-8212-d4cbe21aa5e4 tempest-ServersV294TestFqdnHostnames-833840938 tempest-ServersV294TestFqdnHostnames-833840938-project-member] [instance: aba844c1-d488-4e95-a550-bd1cb7331fb2] Extending root virtual disk to 1048576 {{(pid=61923) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 953.987596] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-cb9e3863-1b9b-41ef-8860-06c627622ce5 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.993618] env[61923]: DEBUG oslo_vmware.api [None req-704a26eb-9a13-4b75-8212-d4cbe21aa5e4 tempest-ServersV294TestFqdnHostnames-833840938 tempest-ServersV294TestFqdnHostnames-833840938-project-member] Waiting for the task: (returnval){ [ 953.993618] env[61923]: value = "task-1377867" [ 953.993618] env[61923]: _type = "Task" [ 953.993618] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 954.000654] env[61923]: DEBUG oslo_vmware.api [None req-704a26eb-9a13-4b75-8212-d4cbe21aa5e4 tempest-ServersV294TestFqdnHostnames-833840938 tempest-ServersV294TestFqdnHostnames-833840938-project-member] Task: {'id': task-1377867, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.049135] env[61923]: DEBUG oslo_concurrency.lockutils [req-d39da0aa-386c-4c04-841e-a5f14bf019a4 req-14b8c8b9-af46-488b-a625-a5ad0dcccc15 service nova] Releasing lock "refresh_cache-aba844c1-d488-4e95-a550-bd1cb7331fb2" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 954.427028] env[61923]: DEBUG oslo_vmware.api [None req-102bab57-0500-4a8d-94a7-052c22c7c916 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Task: {'id': task-1377863, 'name': PowerOnVM_Task, 'duration_secs': 1.31716} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 954.427389] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-102bab57-0500-4a8d-94a7-052c22c7c916 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: 7867c6cf-2ad6-414f-ab52-581827860836] Powered on the VM {{(pid=61923) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 954.427507] env[61923]: INFO nova.compute.manager [None req-102bab57-0500-4a8d-94a7-052c22c7c916 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: 7867c6cf-2ad6-414f-ab52-581827860836] Took 10.23 seconds to spawn the instance on the hypervisor. [ 954.427687] env[61923]: DEBUG nova.compute.manager [None req-102bab57-0500-4a8d-94a7-052c22c7c916 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: 7867c6cf-2ad6-414f-ab52-581827860836] Checking state {{(pid=61923) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 954.428420] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00602a3b-cd55-417e-b6c2-af8ebc9b9f8f {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.443165] env[61923]: DEBUG oslo_concurrency.lockutils [None req-eea05bb2-c691-4804-8c8d-019f8feaa521 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.190s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 954.445168] env[61923]: DEBUG oslo_concurrency.lockutils [None req-af395613-edaf-421a-b79b-9579e1c9e371 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.326s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 954.445396] env[61923]: DEBUG nova.objects.instance [None req-af395613-edaf-421a-b79b-9579e1c9e371 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Lazy-loading 'resources' on Instance uuid 02a8f197-80bb-4cee-bdd6-b07705759986 {{(pid=61923) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 954.474195] env[61923]: DEBUG oslo_vmware.api [None req-69e9bd4b-8d1d-4678-80b0-4de2eeffdae8 tempest-ServersAaction247Test-1300676990 tempest-ServersAaction247Test-1300676990-project-member] Task: {'id': task-1377866, 'name': PowerOnVM_Task, 'duration_secs': 0.436427} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 954.474427] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-69e9bd4b-8d1d-4678-80b0-4de2eeffdae8 tempest-ServersAaction247Test-1300676990 tempest-ServersAaction247Test-1300676990-project-member] [instance: 44196b4c-1401-40fa-bd14-04a49947ab15] Powered on the VM {{(pid=61923) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 954.474673] env[61923]: INFO nova.compute.manager [None req-69e9bd4b-8d1d-4678-80b0-4de2eeffdae8 tempest-ServersAaction247Test-1300676990 tempest-ServersAaction247Test-1300676990-project-member] [instance: 44196b4c-1401-40fa-bd14-04a49947ab15] Took 7.26 seconds to spawn the instance on the hypervisor. [ 954.474961] env[61923]: DEBUG nova.compute.manager [None req-69e9bd4b-8d1d-4678-80b0-4de2eeffdae8 tempest-ServersAaction247Test-1300676990 tempest-ServersAaction247Test-1300676990-project-member] [instance: 44196b4c-1401-40fa-bd14-04a49947ab15] Checking state {{(pid=61923) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 954.475799] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7bf4c8b3-805b-452f-ab68-5562cea7a099 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.500103] env[61923]: INFO nova.network.neutron [None req-eea05bb2-c691-4804-8c8d-019f8feaa521 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] [instance: a511392e-9ab6-42fb-b07c-f90e9786dcc1] Updating port 0325cd38-bd49-4132-85e3-c66fc9efe49f with attributes {'binding:host_id': 'cpu-1', 'device_owner': 'compute:nova'} [ 954.505101] env[61923]: DEBUG oslo_vmware.api [None req-704a26eb-9a13-4b75-8212-d4cbe21aa5e4 tempest-ServersV294TestFqdnHostnames-833840938 tempest-ServersV294TestFqdnHostnames-833840938-project-member] Task: {'id': task-1377867, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.064379} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 954.505355] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-704a26eb-9a13-4b75-8212-d4cbe21aa5e4 tempest-ServersV294TestFqdnHostnames-833840938 tempest-ServersV294TestFqdnHostnames-833840938-project-member] [instance: aba844c1-d488-4e95-a550-bd1cb7331fb2] Extended root virtual disk {{(pid=61923) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 954.506102] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6006fba-cf7a-4c71-b683-be5be948046f {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.527667] env[61923]: DEBUG nova.virt.vmwareapi.volumeops [None req-704a26eb-9a13-4b75-8212-d4cbe21aa5e4 tempest-ServersV294TestFqdnHostnames-833840938 tempest-ServersV294TestFqdnHostnames-833840938-project-member] [instance: aba844c1-d488-4e95-a550-bd1cb7331fb2] Reconfiguring VM instance instance-0000005b to attach disk [datastore1] aba844c1-d488-4e95-a550-bd1cb7331fb2/aba844c1-d488-4e95-a550-bd1cb7331fb2.vmdk or device None with type sparse {{(pid=61923) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 954.528566] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1356bfc6-01ae-4122-b5ab-78b024889aef {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.547077] env[61923]: DEBUG oslo_vmware.api [None req-704a26eb-9a13-4b75-8212-d4cbe21aa5e4 tempest-ServersV294TestFqdnHostnames-833840938 tempest-ServersV294TestFqdnHostnames-833840938-project-member] Waiting for the task: (returnval){ [ 954.547077] env[61923]: value = "task-1377868" [ 954.547077] env[61923]: _type = "Task" [ 954.547077] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 954.557344] env[61923]: DEBUG oslo_vmware.api [None req-704a26eb-9a13-4b75-8212-d4cbe21aa5e4 tempest-ServersV294TestFqdnHostnames-833840938 tempest-ServersV294TestFqdnHostnames-833840938-project-member] Task: {'id': task-1377868, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.946601] env[61923]: INFO nova.compute.manager [None req-102bab57-0500-4a8d-94a7-052c22c7c916 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: 7867c6cf-2ad6-414f-ab52-581827860836] Took 17.57 seconds to build instance. [ 954.992076] env[61923]: INFO nova.compute.manager [None req-69e9bd4b-8d1d-4678-80b0-4de2eeffdae8 tempest-ServersAaction247Test-1300676990 tempest-ServersAaction247Test-1300676990-project-member] [instance: 44196b4c-1401-40fa-bd14-04a49947ab15] Took 12.01 seconds to build instance. [ 955.037572] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f24e407b-aedf-4227-a628-571da0f4f646 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.045476] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a589e7e-5f29-44b9-a2b0-9c6e096486ff {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.056931] env[61923]: DEBUG oslo_vmware.api [None req-704a26eb-9a13-4b75-8212-d4cbe21aa5e4 tempest-ServersV294TestFqdnHostnames-833840938 tempest-ServersV294TestFqdnHostnames-833840938-project-member] Task: {'id': task-1377868, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.082000] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7840b59-a8a1-4edb-81c5-bb3a908c3042 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.089803] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4193698c-46c0-496f-9765-ee08eb7eeba8 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.103188] env[61923]: DEBUG nova.compute.provider_tree [None req-af395613-edaf-421a-b79b-9579e1c9e371 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 955.447758] env[61923]: DEBUG nova.compute.manager [None req-7a7ec250-e6df-4499-8e79-fb254d66e8e9 tempest-ServersAaction247Test-1300676990 tempest-ServersAaction247Test-1300676990-project-member] [instance: 44196b4c-1401-40fa-bd14-04a49947ab15] Checking state {{(pid=61923) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 955.448758] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d06a1169-2251-4c51-912c-9bef74ae71eb {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.452104] env[61923]: DEBUG oslo_concurrency.lockutils [None req-102bab57-0500-4a8d-94a7-052c22c7c916 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Lock "7867c6cf-2ad6-414f-ab52-581827860836" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 19.083s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 955.496306] env[61923]: DEBUG oslo_concurrency.lockutils [None req-69e9bd4b-8d1d-4678-80b0-4de2eeffdae8 tempest-ServersAaction247Test-1300676990 tempest-ServersAaction247Test-1300676990-project-member] Lock "44196b4c-1401-40fa-bd14-04a49947ab15" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 13.526s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 955.559069] env[61923]: DEBUG oslo_vmware.api [None req-704a26eb-9a13-4b75-8212-d4cbe21aa5e4 tempest-ServersV294TestFqdnHostnames-833840938 tempest-ServersV294TestFqdnHostnames-833840938-project-member] Task: {'id': task-1377868, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.606355] env[61923]: DEBUG nova.scheduler.client.report [None req-af395613-edaf-421a-b79b-9579e1c9e371 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 955.619392] env[61923]: DEBUG oslo_concurrency.lockutils [None req-2605b50b-f7ed-4c40-b6af-0f9b593c3112 tempest-ServersAaction247Test-1300676990 tempest-ServersAaction247Test-1300676990-project-member] Acquiring lock "44196b4c-1401-40fa-bd14-04a49947ab15" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 955.619661] env[61923]: DEBUG oslo_concurrency.lockutils [None req-2605b50b-f7ed-4c40-b6af-0f9b593c3112 tempest-ServersAaction247Test-1300676990 tempest-ServersAaction247Test-1300676990-project-member] Lock "44196b4c-1401-40fa-bd14-04a49947ab15" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 955.619869] env[61923]: DEBUG oslo_concurrency.lockutils [None req-2605b50b-f7ed-4c40-b6af-0f9b593c3112 tempest-ServersAaction247Test-1300676990 tempest-ServersAaction247Test-1300676990-project-member] Acquiring lock "44196b4c-1401-40fa-bd14-04a49947ab15-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 955.620069] env[61923]: DEBUG oslo_concurrency.lockutils [None req-2605b50b-f7ed-4c40-b6af-0f9b593c3112 tempest-ServersAaction247Test-1300676990 tempest-ServersAaction247Test-1300676990-project-member] Lock "44196b4c-1401-40fa-bd14-04a49947ab15-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 955.620244] env[61923]: DEBUG oslo_concurrency.lockutils [None req-2605b50b-f7ed-4c40-b6af-0f9b593c3112 tempest-ServersAaction247Test-1300676990 tempest-ServersAaction247Test-1300676990-project-member] Lock "44196b4c-1401-40fa-bd14-04a49947ab15-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 955.622557] env[61923]: INFO nova.compute.manager [None req-2605b50b-f7ed-4c40-b6af-0f9b593c3112 tempest-ServersAaction247Test-1300676990 tempest-ServersAaction247Test-1300676990-project-member] [instance: 44196b4c-1401-40fa-bd14-04a49947ab15] Terminating instance [ 955.624254] env[61923]: DEBUG oslo_concurrency.lockutils [None req-2605b50b-f7ed-4c40-b6af-0f9b593c3112 tempest-ServersAaction247Test-1300676990 tempest-ServersAaction247Test-1300676990-project-member] Acquiring lock "refresh_cache-44196b4c-1401-40fa-bd14-04a49947ab15" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 955.624434] env[61923]: DEBUG oslo_concurrency.lockutils [None req-2605b50b-f7ed-4c40-b6af-0f9b593c3112 tempest-ServersAaction247Test-1300676990 tempest-ServersAaction247Test-1300676990-project-member] Acquired lock "refresh_cache-44196b4c-1401-40fa-bd14-04a49947ab15" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 955.624633] env[61923]: DEBUG nova.network.neutron [None req-2605b50b-f7ed-4c40-b6af-0f9b593c3112 tempest-ServersAaction247Test-1300676990 tempest-ServersAaction247Test-1300676990-project-member] [instance: 44196b4c-1401-40fa-bd14-04a49947ab15] Building network info cache for instance {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 955.652461] env[61923]: DEBUG oslo_concurrency.lockutils [None req-ac67ac24-c7c3-4924-a5b4-fd1213acaa96 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Acquiring lock "7867c6cf-2ad6-414f-ab52-581827860836" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 955.652703] env[61923]: DEBUG oslo_concurrency.lockutils [None req-ac67ac24-c7c3-4924-a5b4-fd1213acaa96 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Lock "7867c6cf-2ad6-414f-ab52-581827860836" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 955.652908] env[61923]: DEBUG oslo_concurrency.lockutils [None req-ac67ac24-c7c3-4924-a5b4-fd1213acaa96 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Acquiring lock "7867c6cf-2ad6-414f-ab52-581827860836-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 955.653123] env[61923]: DEBUG oslo_concurrency.lockutils [None req-ac67ac24-c7c3-4924-a5b4-fd1213acaa96 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Lock "7867c6cf-2ad6-414f-ab52-581827860836-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 955.653299] env[61923]: DEBUG oslo_concurrency.lockutils [None req-ac67ac24-c7c3-4924-a5b4-fd1213acaa96 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Lock "7867c6cf-2ad6-414f-ab52-581827860836-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 955.655013] env[61923]: INFO nova.compute.manager [None req-ac67ac24-c7c3-4924-a5b4-fd1213acaa96 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: 7867c6cf-2ad6-414f-ab52-581827860836] Terminating instance [ 955.656606] env[61923]: DEBUG nova.compute.manager [None req-ac67ac24-c7c3-4924-a5b4-fd1213acaa96 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: 7867c6cf-2ad6-414f-ab52-581827860836] Start destroying the instance on the hypervisor. {{(pid=61923) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 955.656810] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-ac67ac24-c7c3-4924-a5b4-fd1213acaa96 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: 7867c6cf-2ad6-414f-ab52-581827860836] Destroying instance {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 955.657678] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a198cee-b244-4ccc-bba0-c0bb49fff5ee {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.665271] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-ac67ac24-c7c3-4924-a5b4-fd1213acaa96 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: 7867c6cf-2ad6-414f-ab52-581827860836] Powering off the VM {{(pid=61923) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 955.665503] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-95d45114-bf10-4d8f-b9a6-95eee0378aed {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.671630] env[61923]: DEBUG oslo_vmware.api [None req-ac67ac24-c7c3-4924-a5b4-fd1213acaa96 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Waiting for the task: (returnval){ [ 955.671630] env[61923]: value = "task-1377869" [ 955.671630] env[61923]: _type = "Task" [ 955.671630] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 955.679701] env[61923]: DEBUG oslo_vmware.api [None req-ac67ac24-c7c3-4924-a5b4-fd1213acaa96 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Task: {'id': task-1377869, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.959749] env[61923]: INFO nova.compute.manager [None req-7a7ec250-e6df-4499-8e79-fb254d66e8e9 tempest-ServersAaction247Test-1300676990 tempest-ServersAaction247Test-1300676990-project-member] [instance: 44196b4c-1401-40fa-bd14-04a49947ab15] instance snapshotting [ 955.960383] env[61923]: DEBUG nova.objects.instance [None req-7a7ec250-e6df-4499-8e79-fb254d66e8e9 tempest-ServersAaction247Test-1300676990 tempest-ServersAaction247Test-1300676990-project-member] Lazy-loading 'flavor' on Instance uuid 44196b4c-1401-40fa-bd14-04a49947ab15 {{(pid=61923) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 956.004828] env[61923]: DEBUG nova.compute.manager [req-b2876d24-1ba2-4114-ab4e-51c781f7c1eb req-8edf2952-2c65-480b-bb62-549536fec98b service nova] [instance: a511392e-9ab6-42fb-b07c-f90e9786dcc1] Received event network-vif-plugged-0325cd38-bd49-4132-85e3-c66fc9efe49f {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 956.005071] env[61923]: DEBUG oslo_concurrency.lockutils [req-b2876d24-1ba2-4114-ab4e-51c781f7c1eb req-8edf2952-2c65-480b-bb62-549536fec98b service nova] Acquiring lock "a511392e-9ab6-42fb-b07c-f90e9786dcc1-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 956.005268] env[61923]: DEBUG oslo_concurrency.lockutils [req-b2876d24-1ba2-4114-ab4e-51c781f7c1eb req-8edf2952-2c65-480b-bb62-549536fec98b service nova] Lock "a511392e-9ab6-42fb-b07c-f90e9786dcc1-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 956.005454] env[61923]: DEBUG oslo_concurrency.lockutils [req-b2876d24-1ba2-4114-ab4e-51c781f7c1eb req-8edf2952-2c65-480b-bb62-549536fec98b service nova] Lock "a511392e-9ab6-42fb-b07c-f90e9786dcc1-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 956.005593] env[61923]: DEBUG nova.compute.manager [req-b2876d24-1ba2-4114-ab4e-51c781f7c1eb req-8edf2952-2c65-480b-bb62-549536fec98b service nova] [instance: a511392e-9ab6-42fb-b07c-f90e9786dcc1] No waiting events found dispatching network-vif-plugged-0325cd38-bd49-4132-85e3-c66fc9efe49f {{(pid=61923) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 956.005770] env[61923]: WARNING nova.compute.manager [req-b2876d24-1ba2-4114-ab4e-51c781f7c1eb req-8edf2952-2c65-480b-bb62-549536fec98b service nova] [instance: a511392e-9ab6-42fb-b07c-f90e9786dcc1] Received unexpected event network-vif-plugged-0325cd38-bd49-4132-85e3-c66fc9efe49f for instance with vm_state shelved_offloaded and task_state spawning. [ 956.059901] env[61923]: DEBUG oslo_vmware.api [None req-704a26eb-9a13-4b75-8212-d4cbe21aa5e4 tempest-ServersV294TestFqdnHostnames-833840938 tempest-ServersV294TestFqdnHostnames-833840938-project-member] Task: {'id': task-1377868, 'name': ReconfigVM_Task, 'duration_secs': 1.090092} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 956.060380] env[61923]: DEBUG nova.virt.vmwareapi.volumeops [None req-704a26eb-9a13-4b75-8212-d4cbe21aa5e4 tempest-ServersV294TestFqdnHostnames-833840938 tempest-ServersV294TestFqdnHostnames-833840938-project-member] [instance: aba844c1-d488-4e95-a550-bd1cb7331fb2] Reconfigured VM instance instance-0000005b to attach disk [datastore1] aba844c1-d488-4e95-a550-bd1cb7331fb2/aba844c1-d488-4e95-a550-bd1cb7331fb2.vmdk or device None with type sparse {{(pid=61923) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 956.061145] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-144c7529-8703-4ed3-9435-1a09dd37473d {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.067152] env[61923]: DEBUG oslo_vmware.api [None req-704a26eb-9a13-4b75-8212-d4cbe21aa5e4 tempest-ServersV294TestFqdnHostnames-833840938 tempest-ServersV294TestFqdnHostnames-833840938-project-member] Waiting for the task: (returnval){ [ 956.067152] env[61923]: value = "task-1377870" [ 956.067152] env[61923]: _type = "Task" [ 956.067152] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 956.075186] env[61923]: DEBUG oslo_vmware.api [None req-704a26eb-9a13-4b75-8212-d4cbe21aa5e4 tempest-ServersV294TestFqdnHostnames-833840938 tempest-ServersV294TestFqdnHostnames-833840938-project-member] Task: {'id': task-1377870, 'name': Rename_Task} progress is 5%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.088351] env[61923]: DEBUG oslo_concurrency.lockutils [None req-eea05bb2-c691-4804-8c8d-019f8feaa521 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Acquiring lock "refresh_cache-a511392e-9ab6-42fb-b07c-f90e9786dcc1" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 956.088615] env[61923]: DEBUG oslo_concurrency.lockutils [None req-eea05bb2-c691-4804-8c8d-019f8feaa521 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Acquired lock "refresh_cache-a511392e-9ab6-42fb-b07c-f90e9786dcc1" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 956.088875] env[61923]: DEBUG nova.network.neutron [None req-eea05bb2-c691-4804-8c8d-019f8feaa521 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] [instance: a511392e-9ab6-42fb-b07c-f90e9786dcc1] Building network info cache for instance {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 956.111909] env[61923]: DEBUG oslo_concurrency.lockutils [None req-af395613-edaf-421a-b79b-9579e1c9e371 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.666s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 956.134327] env[61923]: INFO nova.scheduler.client.report [None req-af395613-edaf-421a-b79b-9579e1c9e371 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Deleted allocations for instance 02a8f197-80bb-4cee-bdd6-b07705759986 [ 956.143573] env[61923]: DEBUG nova.network.neutron [None req-2605b50b-f7ed-4c40-b6af-0f9b593c3112 tempest-ServersAaction247Test-1300676990 tempest-ServersAaction247Test-1300676990-project-member] [instance: 44196b4c-1401-40fa-bd14-04a49947ab15] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 956.182660] env[61923]: DEBUG oslo_vmware.api [None req-ac67ac24-c7c3-4924-a5b4-fd1213acaa96 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Task: {'id': task-1377869, 'name': PowerOffVM_Task, 'duration_secs': 0.236945} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 956.183586] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-ac67ac24-c7c3-4924-a5b4-fd1213acaa96 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: 7867c6cf-2ad6-414f-ab52-581827860836] Powered off the VM {{(pid=61923) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 956.183780] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-ac67ac24-c7c3-4924-a5b4-fd1213acaa96 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: 7867c6cf-2ad6-414f-ab52-581827860836] Unregistering the VM {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 956.184091] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-3422bea9-98fd-4341-84f7-18919210fbe8 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.231187] env[61923]: DEBUG nova.network.neutron [None req-2605b50b-f7ed-4c40-b6af-0f9b593c3112 tempest-ServersAaction247Test-1300676990 tempest-ServersAaction247Test-1300676990-project-member] [instance: 44196b4c-1401-40fa-bd14-04a49947ab15] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 956.247696] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-ac67ac24-c7c3-4924-a5b4-fd1213acaa96 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: 7867c6cf-2ad6-414f-ab52-581827860836] Unregistered the VM {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 956.247944] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-ac67ac24-c7c3-4924-a5b4-fd1213acaa96 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: 7867c6cf-2ad6-414f-ab52-581827860836] Deleting contents of the VM from datastore datastore1 {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 956.248149] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-ac67ac24-c7c3-4924-a5b4-fd1213acaa96 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Deleting the datastore file [datastore1] 7867c6cf-2ad6-414f-ab52-581827860836 {{(pid=61923) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 956.248676] env[61923]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-bf2daef8-2f6a-4482-9dc4-07e3d36c61a8 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.256921] env[61923]: DEBUG oslo_vmware.api [None req-ac67ac24-c7c3-4924-a5b4-fd1213acaa96 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Waiting for the task: (returnval){ [ 956.256921] env[61923]: value = "task-1377872" [ 956.256921] env[61923]: _type = "Task" [ 956.256921] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 956.265022] env[61923]: DEBUG oslo_vmware.api [None req-ac67ac24-c7c3-4924-a5b4-fd1213acaa96 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Task: {'id': task-1377872, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.466548] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39d47810-13ab-438d-9904-e9b2ed8d139a {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.482776] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cbf99f26-0177-4a26-ad69-5edb4a0a7e4c {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.576593] env[61923]: DEBUG oslo_vmware.api [None req-704a26eb-9a13-4b75-8212-d4cbe21aa5e4 tempest-ServersV294TestFqdnHostnames-833840938 tempest-ServersV294TestFqdnHostnames-833840938-project-member] Task: {'id': task-1377870, 'name': Rename_Task, 'duration_secs': 0.12742} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 956.576904] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-704a26eb-9a13-4b75-8212-d4cbe21aa5e4 tempest-ServersV294TestFqdnHostnames-833840938 tempest-ServersV294TestFqdnHostnames-833840938-project-member] [instance: aba844c1-d488-4e95-a550-bd1cb7331fb2] Powering on the VM {{(pid=61923) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 956.577162] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5c9153ca-263c-4b9e-8302-3840fe9e3418 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.582992] env[61923]: DEBUG oslo_vmware.api [None req-704a26eb-9a13-4b75-8212-d4cbe21aa5e4 tempest-ServersV294TestFqdnHostnames-833840938 tempest-ServersV294TestFqdnHostnames-833840938-project-member] Waiting for the task: (returnval){ [ 956.582992] env[61923]: value = "task-1377873" [ 956.582992] env[61923]: _type = "Task" [ 956.582992] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 956.597258] env[61923]: DEBUG oslo_vmware.api [None req-704a26eb-9a13-4b75-8212-d4cbe21aa5e4 tempest-ServersV294TestFqdnHostnames-833840938 tempest-ServersV294TestFqdnHostnames-833840938-project-member] Task: {'id': task-1377873, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.644477] env[61923]: DEBUG oslo_concurrency.lockutils [None req-af395613-edaf-421a-b79b-9579e1c9e371 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Lock "02a8f197-80bb-4cee-bdd6-b07705759986" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 9.725s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 956.737822] env[61923]: DEBUG oslo_concurrency.lockutils [None req-2605b50b-f7ed-4c40-b6af-0f9b593c3112 tempest-ServersAaction247Test-1300676990 tempest-ServersAaction247Test-1300676990-project-member] Releasing lock "refresh_cache-44196b4c-1401-40fa-bd14-04a49947ab15" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 956.738201] env[61923]: DEBUG nova.compute.manager [None req-2605b50b-f7ed-4c40-b6af-0f9b593c3112 tempest-ServersAaction247Test-1300676990 tempest-ServersAaction247Test-1300676990-project-member] [instance: 44196b4c-1401-40fa-bd14-04a49947ab15] Start destroying the instance on the hypervisor. {{(pid=61923) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 956.738473] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-2605b50b-f7ed-4c40-b6af-0f9b593c3112 tempest-ServersAaction247Test-1300676990 tempest-ServersAaction247Test-1300676990-project-member] [instance: 44196b4c-1401-40fa-bd14-04a49947ab15] Destroying instance {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 956.739763] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc266dfa-e235-4be9-835d-950f65e09b6b {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.749282] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-2605b50b-f7ed-4c40-b6af-0f9b593c3112 tempest-ServersAaction247Test-1300676990 tempest-ServersAaction247Test-1300676990-project-member] [instance: 44196b4c-1401-40fa-bd14-04a49947ab15] Powering off the VM {{(pid=61923) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 956.749599] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d1cbf118-3be0-49a1-878f-5f154baeb8a9 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.755762] env[61923]: DEBUG oslo_vmware.api [None req-2605b50b-f7ed-4c40-b6af-0f9b593c3112 tempest-ServersAaction247Test-1300676990 tempest-ServersAaction247Test-1300676990-project-member] Waiting for the task: (returnval){ [ 956.755762] env[61923]: value = "task-1377874" [ 956.755762] env[61923]: _type = "Task" [ 956.755762] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 956.771073] env[61923]: DEBUG oslo_vmware.api [None req-2605b50b-f7ed-4c40-b6af-0f9b593c3112 tempest-ServersAaction247Test-1300676990 tempest-ServersAaction247Test-1300676990-project-member] Task: {'id': task-1377874, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.774438] env[61923]: DEBUG oslo_vmware.api [None req-ac67ac24-c7c3-4924-a5b4-fd1213acaa96 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Task: {'id': task-1377872, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.139315} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 956.774858] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-ac67ac24-c7c3-4924-a5b4-fd1213acaa96 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Deleted the datastore file {{(pid=61923) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 956.775105] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-ac67ac24-c7c3-4924-a5b4-fd1213acaa96 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: 7867c6cf-2ad6-414f-ab52-581827860836] Deleted contents of the VM from datastore datastore1 {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 956.775302] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-ac67ac24-c7c3-4924-a5b4-fd1213acaa96 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: 7867c6cf-2ad6-414f-ab52-581827860836] Instance destroyed {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 956.775588] env[61923]: INFO nova.compute.manager [None req-ac67ac24-c7c3-4924-a5b4-fd1213acaa96 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: 7867c6cf-2ad6-414f-ab52-581827860836] Took 1.12 seconds to destroy the instance on the hypervisor. [ 956.775974] env[61923]: DEBUG oslo.service.loopingcall [None req-ac67ac24-c7c3-4924-a5b4-fd1213acaa96 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61923) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 956.776267] env[61923]: DEBUG nova.compute.manager [-] [instance: 7867c6cf-2ad6-414f-ab52-581827860836] Deallocating network for instance {{(pid=61923) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 956.776378] env[61923]: DEBUG nova.network.neutron [-] [instance: 7867c6cf-2ad6-414f-ab52-581827860836] deallocate_for_instance() {{(pid=61923) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 956.919712] env[61923]: DEBUG nova.network.neutron [None req-eea05bb2-c691-4804-8c8d-019f8feaa521 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] [instance: a511392e-9ab6-42fb-b07c-f90e9786dcc1] Updating instance_info_cache with network_info: [{"id": "0325cd38-bd49-4132-85e3-c66fc9efe49f", "address": "fa:16:3e:7f:89:31", "network": {"id": "07fc148e-c3ba-4e53-bec2-36500ec38add", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1374533437-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0d24f4a47d1f4b79a7d999a4dacb99ba", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b2907cce-d529-4809-af05-d29397bed211", "external-id": "nsx-vlan-transportzone-427", "segmentation_id": 427, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0325cd38-bd", "ovs_interfaceid": "0325cd38-bd49-4132-85e3-c66fc9efe49f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 956.992096] env[61923]: DEBUG nova.compute.manager [None req-7a7ec250-e6df-4499-8e79-fb254d66e8e9 tempest-ServersAaction247Test-1300676990 tempest-ServersAaction247Test-1300676990-project-member] [instance: 44196b4c-1401-40fa-bd14-04a49947ab15] Instance disappeared during snapshot {{(pid=61923) _snapshot_instance /opt/stack/nova/nova/compute/manager.py:4494}} [ 957.095414] env[61923]: DEBUG oslo_vmware.api [None req-704a26eb-9a13-4b75-8212-d4cbe21aa5e4 tempest-ServersV294TestFqdnHostnames-833840938 tempest-ServersV294TestFqdnHostnames-833840938-project-member] Task: {'id': task-1377873, 'name': PowerOnVM_Task, 'duration_secs': 0.458623} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 957.095696] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-704a26eb-9a13-4b75-8212-d4cbe21aa5e4 tempest-ServersV294TestFqdnHostnames-833840938 tempest-ServersV294TestFqdnHostnames-833840938-project-member] [instance: aba844c1-d488-4e95-a550-bd1cb7331fb2] Powered on the VM {{(pid=61923) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 957.095903] env[61923]: INFO nova.compute.manager [None req-704a26eb-9a13-4b75-8212-d4cbe21aa5e4 tempest-ServersV294TestFqdnHostnames-833840938 tempest-ServersV294TestFqdnHostnames-833840938-project-member] [instance: aba844c1-d488-4e95-a550-bd1cb7331fb2] Took 7.53 seconds to spawn the instance on the hypervisor. [ 957.096100] env[61923]: DEBUG nova.compute.manager [None req-704a26eb-9a13-4b75-8212-d4cbe21aa5e4 tempest-ServersV294TestFqdnHostnames-833840938 tempest-ServersV294TestFqdnHostnames-833840938-project-member] [instance: aba844c1-d488-4e95-a550-bd1cb7331fb2] Checking state {{(pid=61923) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 957.096895] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-558343fe-281f-41fb-99e6-4db3aad7c50b {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.136719] env[61923]: DEBUG nova.compute.manager [None req-7a7ec250-e6df-4499-8e79-fb254d66e8e9 tempest-ServersAaction247Test-1300676990 tempest-ServersAaction247Test-1300676990-project-member] [instance: 44196b4c-1401-40fa-bd14-04a49947ab15] Found 0 images (rotation: 2) {{(pid=61923) _rotate_backups /opt/stack/nova/nova/compute/manager.py:4554}} [ 957.266948] env[61923]: DEBUG oslo_vmware.api [None req-2605b50b-f7ed-4c40-b6af-0f9b593c3112 tempest-ServersAaction247Test-1300676990 tempest-ServersAaction247Test-1300676990-project-member] Task: {'id': task-1377874, 'name': PowerOffVM_Task, 'duration_secs': 0.170811} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 957.267367] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-2605b50b-f7ed-4c40-b6af-0f9b593c3112 tempest-ServersAaction247Test-1300676990 tempest-ServersAaction247Test-1300676990-project-member] [instance: 44196b4c-1401-40fa-bd14-04a49947ab15] Powered off the VM {{(pid=61923) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 957.267563] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-2605b50b-f7ed-4c40-b6af-0f9b593c3112 tempest-ServersAaction247Test-1300676990 tempest-ServersAaction247Test-1300676990-project-member] [instance: 44196b4c-1401-40fa-bd14-04a49947ab15] Unregistering the VM {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 957.267827] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-89619ac3-7050-4493-a573-f7df22538ec2 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.290465] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-2605b50b-f7ed-4c40-b6af-0f9b593c3112 tempest-ServersAaction247Test-1300676990 tempest-ServersAaction247Test-1300676990-project-member] [instance: 44196b4c-1401-40fa-bd14-04a49947ab15] Unregistered the VM {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 957.290732] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-2605b50b-f7ed-4c40-b6af-0f9b593c3112 tempest-ServersAaction247Test-1300676990 tempest-ServersAaction247Test-1300676990-project-member] [instance: 44196b4c-1401-40fa-bd14-04a49947ab15] Deleting contents of the VM from datastore datastore1 {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 957.290925] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-2605b50b-f7ed-4c40-b6af-0f9b593c3112 tempest-ServersAaction247Test-1300676990 tempest-ServersAaction247Test-1300676990-project-member] Deleting the datastore file [datastore1] 44196b4c-1401-40fa-bd14-04a49947ab15 {{(pid=61923) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 957.291206] env[61923]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-40a27c44-fdb6-4a96-b3cc-c9571d553244 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.297927] env[61923]: DEBUG oslo_vmware.api [None req-2605b50b-f7ed-4c40-b6af-0f9b593c3112 tempest-ServersAaction247Test-1300676990 tempest-ServersAaction247Test-1300676990-project-member] Waiting for the task: (returnval){ [ 957.297927] env[61923]: value = "task-1377876" [ 957.297927] env[61923]: _type = "Task" [ 957.297927] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 957.306715] env[61923]: DEBUG oslo_vmware.api [None req-2605b50b-f7ed-4c40-b6af-0f9b593c3112 tempest-ServersAaction247Test-1300676990 tempest-ServersAaction247Test-1300676990-project-member] Task: {'id': task-1377876, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.426638] env[61923]: DEBUG oslo_concurrency.lockutils [None req-eea05bb2-c691-4804-8c8d-019f8feaa521 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Releasing lock "refresh_cache-a511392e-9ab6-42fb-b07c-f90e9786dcc1" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 957.451322] env[61923]: DEBUG nova.virt.hardware [None req-eea05bb2-c691-4804-8c8d-019f8feaa521 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-29T20:07:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='fb4f21c316ef7c91821c6d6f832a06c6',container_format='bare',created_at=2024-10-29T20:16:26Z,direct_url=,disk_format='vmdk',id=5f919f0b-0e82-4adc-b49b-362960821be5,min_disk=1,min_ram=0,name='tempest-ServersNegativeTestJSON-server-727650487-shelved',owner='0d24f4a47d1f4b79a7d999a4dacb99ba',properties=ImageMetaProps,protected=,size=31667200,status='active',tags=,updated_at=2024-10-29T20:16:39Z,virtual_size=,visibility=), allow threads: False {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 957.451477] env[61923]: DEBUG nova.virt.hardware [None req-eea05bb2-c691-4804-8c8d-019f8feaa521 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Flavor limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 957.451634] env[61923]: DEBUG nova.virt.hardware [None req-eea05bb2-c691-4804-8c8d-019f8feaa521 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Image limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 957.451841] env[61923]: DEBUG nova.virt.hardware [None req-eea05bb2-c691-4804-8c8d-019f8feaa521 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Flavor pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 957.451994] env[61923]: DEBUG nova.virt.hardware [None req-eea05bb2-c691-4804-8c8d-019f8feaa521 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Image pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 957.452158] env[61923]: DEBUG nova.virt.hardware [None req-eea05bb2-c691-4804-8c8d-019f8feaa521 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 957.452372] env[61923]: DEBUG nova.virt.hardware [None req-eea05bb2-c691-4804-8c8d-019f8feaa521 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 957.452578] env[61923]: DEBUG nova.virt.hardware [None req-eea05bb2-c691-4804-8c8d-019f8feaa521 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 957.452692] env[61923]: DEBUG nova.virt.hardware [None req-eea05bb2-c691-4804-8c8d-019f8feaa521 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Got 1 possible topologies {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 957.452850] env[61923]: DEBUG nova.virt.hardware [None req-eea05bb2-c691-4804-8c8d-019f8feaa521 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 957.453031] env[61923]: DEBUG nova.virt.hardware [None req-eea05bb2-c691-4804-8c8d-019f8feaa521 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 957.453910] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68ea9169-2629-4420-8346-fc408b4170dd {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.461940] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1cb2c977-a894-45bb-99e2-ff4eef24f337 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.474878] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-eea05bb2-c691-4804-8c8d-019f8feaa521 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] [instance: a511392e-9ab6-42fb-b07c-f90e9786dcc1] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:7f:89:31', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'b2907cce-d529-4809-af05-d29397bed211', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '0325cd38-bd49-4132-85e3-c66fc9efe49f', 'vif_model': 'vmxnet3'}] {{(pid=61923) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 957.482188] env[61923]: DEBUG oslo.service.loopingcall [None req-eea05bb2-c691-4804-8c8d-019f8feaa521 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61923) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 957.482426] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a511392e-9ab6-42fb-b07c-f90e9786dcc1] Creating VM on the ESX host {{(pid=61923) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 957.482630] env[61923]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-5cf51c8d-bce3-4ce0-b8a2-8a3cfb1bb1ab {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.501094] env[61923]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 957.501094] env[61923]: value = "task-1377877" [ 957.501094] env[61923]: _type = "Task" [ 957.501094] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 957.508070] env[61923]: DEBUG oslo_vmware.api [-] Task: {'id': task-1377877, 'name': CreateVM_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.613953] env[61923]: INFO nova.compute.manager [None req-704a26eb-9a13-4b75-8212-d4cbe21aa5e4 tempest-ServersV294TestFqdnHostnames-833840938 tempest-ServersV294TestFqdnHostnames-833840938-project-member] [instance: aba844c1-d488-4e95-a550-bd1cb7331fb2] Took 12.28 seconds to build instance. [ 957.687183] env[61923]: DEBUG nova.network.neutron [-] [instance: 7867c6cf-2ad6-414f-ab52-581827860836] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 957.780171] env[61923]: DEBUG oslo_concurrency.lockutils [None req-9798b6a1-c446-4119-9842-2560d710a905 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Acquiring lock "638bbbbb-827e-4a73-ac24-a3e1b2cd72eb" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 957.780422] env[61923]: DEBUG oslo_concurrency.lockutils [None req-9798b6a1-c446-4119-9842-2560d710a905 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Lock "638bbbbb-827e-4a73-ac24-a3e1b2cd72eb" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 957.807771] env[61923]: DEBUG oslo_vmware.api [None req-2605b50b-f7ed-4c40-b6af-0f9b593c3112 tempest-ServersAaction247Test-1300676990 tempest-ServersAaction247Test-1300676990-project-member] Task: {'id': task-1377876, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.10773} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 957.808042] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-2605b50b-f7ed-4c40-b6af-0f9b593c3112 tempest-ServersAaction247Test-1300676990 tempest-ServersAaction247Test-1300676990-project-member] Deleted the datastore file {{(pid=61923) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 957.808238] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-2605b50b-f7ed-4c40-b6af-0f9b593c3112 tempest-ServersAaction247Test-1300676990 tempest-ServersAaction247Test-1300676990-project-member] [instance: 44196b4c-1401-40fa-bd14-04a49947ab15] Deleted contents of the VM from datastore datastore1 {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 957.808415] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-2605b50b-f7ed-4c40-b6af-0f9b593c3112 tempest-ServersAaction247Test-1300676990 tempest-ServersAaction247Test-1300676990-project-member] [instance: 44196b4c-1401-40fa-bd14-04a49947ab15] Instance destroyed {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 957.808586] env[61923]: INFO nova.compute.manager [None req-2605b50b-f7ed-4c40-b6af-0f9b593c3112 tempest-ServersAaction247Test-1300676990 tempest-ServersAaction247Test-1300676990-project-member] [instance: 44196b4c-1401-40fa-bd14-04a49947ab15] Took 1.07 seconds to destroy the instance on the hypervisor. [ 957.808828] env[61923]: DEBUG oslo.service.loopingcall [None req-2605b50b-f7ed-4c40-b6af-0f9b593c3112 tempest-ServersAaction247Test-1300676990 tempest-ServersAaction247Test-1300676990-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61923) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 957.809041] env[61923]: DEBUG nova.compute.manager [-] [instance: 44196b4c-1401-40fa-bd14-04a49947ab15] Deallocating network for instance {{(pid=61923) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 957.809312] env[61923]: DEBUG nova.network.neutron [-] [instance: 44196b4c-1401-40fa-bd14-04a49947ab15] deallocate_for_instance() {{(pid=61923) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 957.827880] env[61923]: DEBUG nova.network.neutron [-] [instance: 44196b4c-1401-40fa-bd14-04a49947ab15] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 958.011441] env[61923]: DEBUG oslo_vmware.api [-] Task: {'id': task-1377877, 'name': CreateVM_Task} progress is 99%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.052651] env[61923]: DEBUG nova.compute.manager [req-9dd0324c-843a-4b66-8022-8bb4028610e5 req-a66c817a-ed6f-4eb3-9eda-20cd53f32269 service nova] [instance: a511392e-9ab6-42fb-b07c-f90e9786dcc1] Received event network-changed-0325cd38-bd49-4132-85e3-c66fc9efe49f {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 958.052829] env[61923]: DEBUG nova.compute.manager [req-9dd0324c-843a-4b66-8022-8bb4028610e5 req-a66c817a-ed6f-4eb3-9eda-20cd53f32269 service nova] [instance: a511392e-9ab6-42fb-b07c-f90e9786dcc1] Refreshing instance network info cache due to event network-changed-0325cd38-bd49-4132-85e3-c66fc9efe49f. {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 958.053048] env[61923]: DEBUG oslo_concurrency.lockutils [req-9dd0324c-843a-4b66-8022-8bb4028610e5 req-a66c817a-ed6f-4eb3-9eda-20cd53f32269 service nova] Acquiring lock "refresh_cache-a511392e-9ab6-42fb-b07c-f90e9786dcc1" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 958.053199] env[61923]: DEBUG oslo_concurrency.lockutils [req-9dd0324c-843a-4b66-8022-8bb4028610e5 req-a66c817a-ed6f-4eb3-9eda-20cd53f32269 service nova] Acquired lock "refresh_cache-a511392e-9ab6-42fb-b07c-f90e9786dcc1" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 958.053360] env[61923]: DEBUG nova.network.neutron [req-9dd0324c-843a-4b66-8022-8bb4028610e5 req-a66c817a-ed6f-4eb3-9eda-20cd53f32269 service nova] [instance: a511392e-9ab6-42fb-b07c-f90e9786dcc1] Refreshing network info cache for port 0325cd38-bd49-4132-85e3-c66fc9efe49f {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 958.116402] env[61923]: DEBUG oslo_concurrency.lockutils [None req-704a26eb-9a13-4b75-8212-d4cbe21aa5e4 tempest-ServersV294TestFqdnHostnames-833840938 tempest-ServersV294TestFqdnHostnames-833840938-project-member] Lock "aba844c1-d488-4e95-a550-bd1cb7331fb2" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 13.786s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 958.190143] env[61923]: INFO nova.compute.manager [-] [instance: 7867c6cf-2ad6-414f-ab52-581827860836] Took 1.41 seconds to deallocate network for instance. [ 958.282628] env[61923]: DEBUG nova.compute.manager [None req-9798b6a1-c446-4119-9842-2560d710a905 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: 638bbbbb-827e-4a73-ac24-a3e1b2cd72eb] Starting instance... {{(pid=61923) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 958.331047] env[61923]: DEBUG nova.network.neutron [-] [instance: 44196b4c-1401-40fa-bd14-04a49947ab15] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 958.511675] env[61923]: DEBUG oslo_vmware.api [-] Task: {'id': task-1377877, 'name': CreateVM_Task, 'duration_secs': 1.00802} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 958.513040] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a511392e-9ab6-42fb-b07c-f90e9786dcc1] Created VM on the ESX host {{(pid=61923) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 958.513040] env[61923]: DEBUG oslo_concurrency.lockutils [None req-eea05bb2-c691-4804-8c8d-019f8feaa521 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/5f919f0b-0e82-4adc-b49b-362960821be5" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 958.513197] env[61923]: DEBUG oslo_concurrency.lockutils [None req-eea05bb2-c691-4804-8c8d-019f8feaa521 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Acquired lock "[datastore1] devstack-image-cache_base/5f919f0b-0e82-4adc-b49b-362960821be5" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 958.513571] env[61923]: DEBUG oslo_concurrency.lockutils [None req-eea05bb2-c691-4804-8c8d-019f8feaa521 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/5f919f0b-0e82-4adc-b49b-362960821be5" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 958.513830] env[61923]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b6cc4f55-7393-4700-bef7-ccfc72ba244c {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.518091] env[61923]: DEBUG oslo_vmware.api [None req-eea05bb2-c691-4804-8c8d-019f8feaa521 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Waiting for the task: (returnval){ [ 958.518091] env[61923]: value = "session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]522f5161-8bea-d0a3-afb5-c936b5924577" [ 958.518091] env[61923]: _type = "Task" [ 958.518091] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 958.525244] env[61923]: DEBUG oslo_vmware.api [None req-eea05bb2-c691-4804-8c8d-019f8feaa521 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]522f5161-8bea-d0a3-afb5-c936b5924577, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.696698] env[61923]: DEBUG oslo_concurrency.lockutils [None req-ac67ac24-c7c3-4924-a5b4-fd1213acaa96 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 958.697215] env[61923]: DEBUG oslo_concurrency.lockutils [None req-ac67ac24-c7c3-4924-a5b4-fd1213acaa96 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 958.697459] env[61923]: DEBUG nova.objects.instance [None req-ac67ac24-c7c3-4924-a5b4-fd1213acaa96 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Lazy-loading 'resources' on Instance uuid 7867c6cf-2ad6-414f-ab52-581827860836 {{(pid=61923) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 958.805857] env[61923]: DEBUG oslo_concurrency.lockutils [None req-9798b6a1-c446-4119-9842-2560d710a905 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 958.833253] env[61923]: INFO nova.compute.manager [-] [instance: 44196b4c-1401-40fa-bd14-04a49947ab15] Took 1.02 seconds to deallocate network for instance. [ 958.920954] env[61923]: DEBUG nova.network.neutron [req-9dd0324c-843a-4b66-8022-8bb4028610e5 req-a66c817a-ed6f-4eb3-9eda-20cd53f32269 service nova] [instance: a511392e-9ab6-42fb-b07c-f90e9786dcc1] Updated VIF entry in instance network info cache for port 0325cd38-bd49-4132-85e3-c66fc9efe49f. {{(pid=61923) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 958.921332] env[61923]: DEBUG nova.network.neutron [req-9dd0324c-843a-4b66-8022-8bb4028610e5 req-a66c817a-ed6f-4eb3-9eda-20cd53f32269 service nova] [instance: a511392e-9ab6-42fb-b07c-f90e9786dcc1] Updating instance_info_cache with network_info: [{"id": "0325cd38-bd49-4132-85e3-c66fc9efe49f", "address": "fa:16:3e:7f:89:31", "network": {"id": "07fc148e-c3ba-4e53-bec2-36500ec38add", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1374533437-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0d24f4a47d1f4b79a7d999a4dacb99ba", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b2907cce-d529-4809-af05-d29397bed211", "external-id": "nsx-vlan-transportzone-427", "segmentation_id": 427, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0325cd38-bd", "ovs_interfaceid": "0325cd38-bd49-4132-85e3-c66fc9efe49f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 959.029688] env[61923]: DEBUG oslo_concurrency.lockutils [None req-eea05bb2-c691-4804-8c8d-019f8feaa521 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Releasing lock "[datastore1] devstack-image-cache_base/5f919f0b-0e82-4adc-b49b-362960821be5" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 959.029959] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-eea05bb2-c691-4804-8c8d-019f8feaa521 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] [instance: a511392e-9ab6-42fb-b07c-f90e9786dcc1] Processing image 5f919f0b-0e82-4adc-b49b-362960821be5 {{(pid=61923) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 959.030214] env[61923]: DEBUG oslo_concurrency.lockutils [None req-eea05bb2-c691-4804-8c8d-019f8feaa521 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/5f919f0b-0e82-4adc-b49b-362960821be5/5f919f0b-0e82-4adc-b49b-362960821be5.vmdk" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 959.030365] env[61923]: DEBUG oslo_concurrency.lockutils [None req-eea05bb2-c691-4804-8c8d-019f8feaa521 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Acquired lock "[datastore1] devstack-image-cache_base/5f919f0b-0e82-4adc-b49b-362960821be5/5f919f0b-0e82-4adc-b49b-362960821be5.vmdk" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 959.030548] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-eea05bb2-c691-4804-8c8d-019f8feaa521 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61923) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 959.030811] env[61923]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-da391e8f-36fb-43c7-8f11-090d86cb553f {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.038759] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-eea05bb2-c691-4804-8c8d-019f8feaa521 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61923) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 959.039045] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-eea05bb2-c691-4804-8c8d-019f8feaa521 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61923) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 959.039790] env[61923]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-32a1f950-0251-41fb-a06a-e319b8fb5201 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.044970] env[61923]: DEBUG oslo_vmware.api [None req-eea05bb2-c691-4804-8c8d-019f8feaa521 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Waiting for the task: (returnval){ [ 959.044970] env[61923]: value = "session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]520c8a51-ae4b-ec8c-c321-bfc2cd082e25" [ 959.044970] env[61923]: _type = "Task" [ 959.044970] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 959.052290] env[61923]: DEBUG oslo_vmware.api [None req-eea05bb2-c691-4804-8c8d-019f8feaa521 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]520c8a51-ae4b-ec8c-c321-bfc2cd082e25, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 959.286496] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0865a4d7-e116-4f92-a7a4-76bad7c38912 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.294324] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b18e5b76-7bca-4c83-8d38-07628a7d231f {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.325542] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae471dec-1825-475f-8d41-402ba14ba974 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.333214] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af4ffe90-d51e-484b-8eac-47f36efedd35 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.339179] env[61923]: DEBUG oslo_concurrency.lockutils [None req-2605b50b-f7ed-4c40-b6af-0f9b593c3112 tempest-ServersAaction247Test-1300676990 tempest-ServersAaction247Test-1300676990-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 959.347141] env[61923]: DEBUG nova.compute.provider_tree [None req-ac67ac24-c7c3-4924-a5b4-fd1213acaa96 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 959.423607] env[61923]: DEBUG oslo_concurrency.lockutils [req-9dd0324c-843a-4b66-8022-8bb4028610e5 req-a66c817a-ed6f-4eb3-9eda-20cd53f32269 service nova] Releasing lock "refresh_cache-a511392e-9ab6-42fb-b07c-f90e9786dcc1" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 959.423849] env[61923]: DEBUG nova.compute.manager [req-9dd0324c-843a-4b66-8022-8bb4028610e5 req-a66c817a-ed6f-4eb3-9eda-20cd53f32269 service nova] [instance: 7867c6cf-2ad6-414f-ab52-581827860836] Received event network-vif-deleted-972754b8-9333-4b89-8435-ca8a98aa7f39 {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 959.424073] env[61923]: DEBUG nova.compute.manager [req-9dd0324c-843a-4b66-8022-8bb4028610e5 req-a66c817a-ed6f-4eb3-9eda-20cd53f32269 service nova] [instance: aba844c1-d488-4e95-a550-bd1cb7331fb2] Received event network-changed-6326df12-72bc-44d6-b660-4abdb5ca5cfb {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 959.424257] env[61923]: DEBUG nova.compute.manager [req-9dd0324c-843a-4b66-8022-8bb4028610e5 req-a66c817a-ed6f-4eb3-9eda-20cd53f32269 service nova] [instance: aba844c1-d488-4e95-a550-bd1cb7331fb2] Refreshing instance network info cache due to event network-changed-6326df12-72bc-44d6-b660-4abdb5ca5cfb. {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 959.424469] env[61923]: DEBUG oslo_concurrency.lockutils [req-9dd0324c-843a-4b66-8022-8bb4028610e5 req-a66c817a-ed6f-4eb3-9eda-20cd53f32269 service nova] Acquiring lock "refresh_cache-aba844c1-d488-4e95-a550-bd1cb7331fb2" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 959.424614] env[61923]: DEBUG oslo_concurrency.lockutils [req-9dd0324c-843a-4b66-8022-8bb4028610e5 req-a66c817a-ed6f-4eb3-9eda-20cd53f32269 service nova] Acquired lock "refresh_cache-aba844c1-d488-4e95-a550-bd1cb7331fb2" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 959.424779] env[61923]: DEBUG nova.network.neutron [req-9dd0324c-843a-4b66-8022-8bb4028610e5 req-a66c817a-ed6f-4eb3-9eda-20cd53f32269 service nova] [instance: aba844c1-d488-4e95-a550-bd1cb7331fb2] Refreshing network info cache for port 6326df12-72bc-44d6-b660-4abdb5ca5cfb {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 959.555749] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-eea05bb2-c691-4804-8c8d-019f8feaa521 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] [instance: a511392e-9ab6-42fb-b07c-f90e9786dcc1] Preparing fetch location {{(pid=61923) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 959.556120] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-eea05bb2-c691-4804-8c8d-019f8feaa521 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] [instance: a511392e-9ab6-42fb-b07c-f90e9786dcc1] Fetch image to [datastore1] OSTACK_IMG_0623b35e-63ab-4acd-bf5d-f33dffaf1c79/OSTACK_IMG_0623b35e-63ab-4acd-bf5d-f33dffaf1c79.vmdk {{(pid=61923) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 959.556253] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-eea05bb2-c691-4804-8c8d-019f8feaa521 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] [instance: a511392e-9ab6-42fb-b07c-f90e9786dcc1] Downloading stream optimized image 5f919f0b-0e82-4adc-b49b-362960821be5 to [datastore1] OSTACK_IMG_0623b35e-63ab-4acd-bf5d-f33dffaf1c79/OSTACK_IMG_0623b35e-63ab-4acd-bf5d-f33dffaf1c79.vmdk on the data store datastore1 as vApp {{(pid=61923) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 959.556486] env[61923]: DEBUG nova.virt.vmwareapi.images [None req-eea05bb2-c691-4804-8c8d-019f8feaa521 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] [instance: a511392e-9ab6-42fb-b07c-f90e9786dcc1] Downloading image file data 5f919f0b-0e82-4adc-b49b-362960821be5 to the ESX as VM named 'OSTACK_IMG_0623b35e-63ab-4acd-bf5d-f33dffaf1c79' {{(pid=61923) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 959.635223] env[61923]: DEBUG oslo_vmware.rw_handles [None req-eea05bb2-c691-4804-8c8d-019f8feaa521 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 959.635223] env[61923]: value = "resgroup-9" [ 959.635223] env[61923]: _type = "ResourcePool" [ 959.635223] env[61923]: }. {{(pid=61923) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 959.635582] env[61923]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-d3e31f14-7092-4513-be6d-560b24c1998c {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.663871] env[61923]: DEBUG oslo_vmware.rw_handles [None req-eea05bb2-c691-4804-8c8d-019f8feaa521 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Lease: (returnval){ [ 959.663871] env[61923]: value = "session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52dcdbef-78f5-9796-ed30-8da66b1013c4" [ 959.663871] env[61923]: _type = "HttpNfcLease" [ 959.663871] env[61923]: } obtained for vApp import into resource pool (val){ [ 959.663871] env[61923]: value = "resgroup-9" [ 959.663871] env[61923]: _type = "ResourcePool" [ 959.663871] env[61923]: }. {{(pid=61923) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 959.664282] env[61923]: DEBUG oslo_vmware.api [None req-eea05bb2-c691-4804-8c8d-019f8feaa521 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Waiting for the lease: (returnval){ [ 959.664282] env[61923]: value = "session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52dcdbef-78f5-9796-ed30-8da66b1013c4" [ 959.664282] env[61923]: _type = "HttpNfcLease" [ 959.664282] env[61923]: } to be ready. {{(pid=61923) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 959.673652] env[61923]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 959.673652] env[61923]: value = "session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52dcdbef-78f5-9796-ed30-8da66b1013c4" [ 959.673652] env[61923]: _type = "HttpNfcLease" [ 959.673652] env[61923]: } is initializing. {{(pid=61923) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 959.850213] env[61923]: DEBUG nova.scheduler.client.report [None req-ac67ac24-c7c3-4924-a5b4-fd1213acaa96 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 960.139981] env[61923]: DEBUG nova.network.neutron [req-9dd0324c-843a-4b66-8022-8bb4028610e5 req-a66c817a-ed6f-4eb3-9eda-20cd53f32269 service nova] [instance: aba844c1-d488-4e95-a550-bd1cb7331fb2] Updated VIF entry in instance network info cache for port 6326df12-72bc-44d6-b660-4abdb5ca5cfb. {{(pid=61923) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 960.140418] env[61923]: DEBUG nova.network.neutron [req-9dd0324c-843a-4b66-8022-8bb4028610e5 req-a66c817a-ed6f-4eb3-9eda-20cd53f32269 service nova] [instance: aba844c1-d488-4e95-a550-bd1cb7331fb2] Updating instance_info_cache with network_info: [{"id": "6326df12-72bc-44d6-b660-4abdb5ca5cfb", "address": "fa:16:3e:dd:2b:a6", "network": {"id": "49b0b81e-2f1d-42ae-bc27-de297b6bef9e", "bridge": "br-int", "label": "tempest-ServersV294TestFqdnHostnames-151433926-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.197", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e9e5485c43514e55b88bdd7d2dc2aee7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d275d7c6-2a7b-4ee8-b6f4-fabf1ba1905f", "external-id": "nsx-vlan-transportzone-513", "segmentation_id": 513, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6326df12-72", "ovs_interfaceid": "6326df12-72bc-44d6-b660-4abdb5ca5cfb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 960.173344] env[61923]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 960.173344] env[61923]: value = "session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52dcdbef-78f5-9796-ed30-8da66b1013c4" [ 960.173344] env[61923]: _type = "HttpNfcLease" [ 960.173344] env[61923]: } is initializing. {{(pid=61923) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 960.355855] env[61923]: DEBUG oslo_concurrency.lockutils [None req-ac67ac24-c7c3-4924-a5b4-fd1213acaa96 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.659s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 960.358315] env[61923]: DEBUG oslo_concurrency.lockutils [None req-9798b6a1-c446-4119-9842-2560d710a905 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.553s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 960.359859] env[61923]: INFO nova.compute.claims [None req-9798b6a1-c446-4119-9842-2560d710a905 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: 638bbbbb-827e-4a73-ac24-a3e1b2cd72eb] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 960.387167] env[61923]: INFO nova.scheduler.client.report [None req-ac67ac24-c7c3-4924-a5b4-fd1213acaa96 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Deleted allocations for instance 7867c6cf-2ad6-414f-ab52-581827860836 [ 960.644059] env[61923]: DEBUG oslo_concurrency.lockutils [req-9dd0324c-843a-4b66-8022-8bb4028610e5 req-a66c817a-ed6f-4eb3-9eda-20cd53f32269 service nova] Releasing lock "refresh_cache-aba844c1-d488-4e95-a550-bd1cb7331fb2" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 960.672870] env[61923]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 960.672870] env[61923]: value = "session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52dcdbef-78f5-9796-ed30-8da66b1013c4" [ 960.672870] env[61923]: _type = "HttpNfcLease" [ 960.672870] env[61923]: } is ready. {{(pid=61923) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 960.674409] env[61923]: DEBUG oslo_vmware.rw_handles [None req-eea05bb2-c691-4804-8c8d-019f8feaa521 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 960.674409] env[61923]: value = "session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52dcdbef-78f5-9796-ed30-8da66b1013c4" [ 960.674409] env[61923]: _type = "HttpNfcLease" [ 960.674409] env[61923]: }. {{(pid=61923) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 960.674592] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a549696f-179d-4f66-9669-543f7edf6a88 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.682151] env[61923]: DEBUG oslo_vmware.rw_handles [None req-eea05bb2-c691-4804-8c8d-019f8feaa521 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52d3c346-44f0-e509-badb-8498dd612c00/disk-0.vmdk from lease info. {{(pid=61923) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 960.682340] env[61923]: DEBUG oslo_vmware.rw_handles [None req-eea05bb2-c691-4804-8c8d-019f8feaa521 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Creating HTTP connection to write to file with size = 31667200 and URL = https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52d3c346-44f0-e509-badb-8498dd612c00/disk-0.vmdk. {{(pid=61923) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 960.747854] env[61923]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-80b657f5-4dff-4a11-8abe-0562851079c4 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.895773] env[61923]: DEBUG oslo_concurrency.lockutils [None req-ac67ac24-c7c3-4924-a5b4-fd1213acaa96 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Lock "7867c6cf-2ad6-414f-ab52-581827860836" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.243s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 961.466079] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e066edae-8631-451f-98c8-9d2a28dd0b2a {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.480763] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d420db76-0064-4fcf-9c63-865212e38295 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.518026] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b1500d8-91b4-49cb-be3f-33a4a799be9a {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.528858] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df9c5df7-d3a6-4766-9555-25f8a66264fa {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.545677] env[61923]: DEBUG nova.compute.provider_tree [None req-9798b6a1-c446-4119-9842-2560d710a905 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 961.556082] env[61923]: DEBUG oslo_concurrency.lockutils [None req-cd7c0dc9-d2a5-476c-afbb-7be6ac77c41c tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Acquiring lock "3fcedef5-9040-4b45-9e27-f0700112d528" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 961.556477] env[61923]: DEBUG oslo_concurrency.lockutils [None req-cd7c0dc9-d2a5-476c-afbb-7be6ac77c41c tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Lock "3fcedef5-9040-4b45-9e27-f0700112d528" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 962.051082] env[61923]: DEBUG nova.scheduler.client.report [None req-9798b6a1-c446-4119-9842-2560d710a905 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 962.058262] env[61923]: DEBUG nova.compute.manager [None req-cd7c0dc9-d2a5-476c-afbb-7be6ac77c41c tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: 3fcedef5-9040-4b45-9e27-f0700112d528] Starting instance... {{(pid=61923) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 962.062569] env[61923]: DEBUG oslo_vmware.rw_handles [None req-eea05bb2-c691-4804-8c8d-019f8feaa521 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Completed reading data from the image iterator. {{(pid=61923) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 962.062835] env[61923]: DEBUG oslo_vmware.rw_handles [None req-eea05bb2-c691-4804-8c8d-019f8feaa521 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52d3c346-44f0-e509-badb-8498dd612c00/disk-0.vmdk. {{(pid=61923) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 962.063901] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55a4ac67-371d-4c87-8b78-c62a3448cdea {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.072971] env[61923]: DEBUG oslo_vmware.rw_handles [None req-eea05bb2-c691-4804-8c8d-019f8feaa521 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52d3c346-44f0-e509-badb-8498dd612c00/disk-0.vmdk is in state: ready. {{(pid=61923) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 962.073150] env[61923]: DEBUG oslo_vmware.rw_handles [None req-eea05bb2-c691-4804-8c8d-019f8feaa521 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Releasing lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52d3c346-44f0-e509-badb-8498dd612c00/disk-0.vmdk. {{(pid=61923) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 962.073367] env[61923]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-ce1dfb35-088c-4cfb-abec-8c33d4926af8 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.557386] env[61923]: DEBUG oslo_concurrency.lockutils [None req-9798b6a1-c446-4119-9842-2560d710a905 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.199s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 962.557982] env[61923]: DEBUG nova.compute.manager [None req-9798b6a1-c446-4119-9842-2560d710a905 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: 638bbbbb-827e-4a73-ac24-a3e1b2cd72eb] Start building networks asynchronously for instance. {{(pid=61923) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 962.562699] env[61923]: DEBUG oslo_concurrency.lockutils [None req-2605b50b-f7ed-4c40-b6af-0f9b593c3112 tempest-ServersAaction247Test-1300676990 tempest-ServersAaction247Test-1300676990-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 3.223s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 962.562699] env[61923]: DEBUG nova.objects.instance [None req-2605b50b-f7ed-4c40-b6af-0f9b593c3112 tempest-ServersAaction247Test-1300676990 tempest-ServersAaction247Test-1300676990-project-member] Lazy-loading 'resources' on Instance uuid 44196b4c-1401-40fa-bd14-04a49947ab15 {{(pid=61923) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 962.582060] env[61923]: DEBUG oslo_concurrency.lockutils [None req-cd7c0dc9-d2a5-476c-afbb-7be6ac77c41c tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 962.895650] env[61923]: DEBUG oslo_vmware.rw_handles [None req-eea05bb2-c691-4804-8c8d-019f8feaa521 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Closed VMDK write handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52d3c346-44f0-e509-badb-8498dd612c00/disk-0.vmdk. {{(pid=61923) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 962.895650] env[61923]: INFO nova.virt.vmwareapi.images [None req-eea05bb2-c691-4804-8c8d-019f8feaa521 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] [instance: a511392e-9ab6-42fb-b07c-f90e9786dcc1] Downloaded image file data 5f919f0b-0e82-4adc-b49b-362960821be5 [ 962.896317] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f4352b2-4855-4d91-b423-8e4a7ab3414c {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.912701] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-f26613f8-7fb5-4596-a3d0-fc781f342735 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.972243] env[61923]: INFO nova.virt.vmwareapi.images [None req-eea05bb2-c691-4804-8c8d-019f8feaa521 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] [instance: a511392e-9ab6-42fb-b07c-f90e9786dcc1] The imported VM was unregistered [ 962.974632] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-eea05bb2-c691-4804-8c8d-019f8feaa521 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] [instance: a511392e-9ab6-42fb-b07c-f90e9786dcc1] Caching image {{(pid=61923) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 962.974876] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-eea05bb2-c691-4804-8c8d-019f8feaa521 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Creating directory with path [datastore1] devstack-image-cache_base/5f919f0b-0e82-4adc-b49b-362960821be5 {{(pid=61923) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 962.975173] env[61923]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-19fd47bb-1e47-4a72-b96b-b3bcdd1083f4 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.989356] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-eea05bb2-c691-4804-8c8d-019f8feaa521 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Created directory with path [datastore1] devstack-image-cache_base/5f919f0b-0e82-4adc-b49b-362960821be5 {{(pid=61923) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 962.989539] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-eea05bb2-c691-4804-8c8d-019f8feaa521 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Moving virtual disk from [datastore1] OSTACK_IMG_0623b35e-63ab-4acd-bf5d-f33dffaf1c79/OSTACK_IMG_0623b35e-63ab-4acd-bf5d-f33dffaf1c79.vmdk to [datastore1] devstack-image-cache_base/5f919f0b-0e82-4adc-b49b-362960821be5/5f919f0b-0e82-4adc-b49b-362960821be5.vmdk. {{(pid=61923) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 962.989869] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-dcbfa196-94a2-41a1-bd71-14405e6f3553 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.998289] env[61923]: DEBUG oslo_vmware.api [None req-eea05bb2-c691-4804-8c8d-019f8feaa521 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Waiting for the task: (returnval){ [ 962.998289] env[61923]: value = "task-1377880" [ 962.998289] env[61923]: _type = "Task" [ 962.998289] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 963.007204] env[61923]: DEBUG oslo_vmware.api [None req-eea05bb2-c691-4804-8c8d-019f8feaa521 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Task: {'id': task-1377880, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.063470] env[61923]: DEBUG nova.compute.utils [None req-9798b6a1-c446-4119-9842-2560d710a905 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Using /dev/sd instead of None {{(pid=61923) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 963.064905] env[61923]: DEBUG nova.compute.manager [None req-9798b6a1-c446-4119-9842-2560d710a905 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: 638bbbbb-827e-4a73-ac24-a3e1b2cd72eb] Allocating IP information in the background. {{(pid=61923) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 963.065096] env[61923]: DEBUG nova.network.neutron [None req-9798b6a1-c446-4119-9842-2560d710a905 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: 638bbbbb-827e-4a73-ac24-a3e1b2cd72eb] allocate_for_instance() {{(pid=61923) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 963.121619] env[61923]: DEBUG nova.policy [None req-9798b6a1-c446-4119-9842-2560d710a905 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5b8a639e99914e75857b4571f0d58a3c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '69cc941eb6dd4780ac12aa29656c37f7', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61923) authorize /opt/stack/nova/nova/policy.py:201}} [ 963.152761] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9402a118-5b95-4c64-ae63-5883c9b21a0c {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.161313] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e16c344f-ed23-4745-a545-f495d4cabb4a {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.193746] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e29daea0-98e5-49bb-879b-0500acc0262a {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.201527] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-acddc23f-62b3-4f45-a24c-c722e11aa284 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.215702] env[61923]: DEBUG nova.compute.provider_tree [None req-2605b50b-f7ed-4c40-b6af-0f9b593c3112 tempest-ServersAaction247Test-1300676990 tempest-ServersAaction247Test-1300676990-project-member] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 963.387046] env[61923]: DEBUG nova.network.neutron [None req-9798b6a1-c446-4119-9842-2560d710a905 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: 638bbbbb-827e-4a73-ac24-a3e1b2cd72eb] Successfully created port: 3e6f05d1-ae7a-4aff-9243-20d5dd0e9b4f {{(pid=61923) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 963.510761] env[61923]: DEBUG oslo_vmware.api [None req-eea05bb2-c691-4804-8c8d-019f8feaa521 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Task: {'id': task-1377880, 'name': MoveVirtualDisk_Task} progress is 18%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.568762] env[61923]: DEBUG nova.compute.manager [None req-9798b6a1-c446-4119-9842-2560d710a905 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: 638bbbbb-827e-4a73-ac24-a3e1b2cd72eb] Start building block device mappings for instance. {{(pid=61923) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 963.719305] env[61923]: DEBUG nova.scheduler.client.report [None req-2605b50b-f7ed-4c40-b6af-0f9b593c3112 tempest-ServersAaction247Test-1300676990 tempest-ServersAaction247Test-1300676990-project-member] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 964.009393] env[61923]: DEBUG oslo_vmware.api [None req-eea05bb2-c691-4804-8c8d-019f8feaa521 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Task: {'id': task-1377880, 'name': MoveVirtualDisk_Task} progress is 38%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.226690] env[61923]: DEBUG oslo_concurrency.lockutils [None req-2605b50b-f7ed-4c40-b6af-0f9b593c3112 tempest-ServersAaction247Test-1300676990 tempest-ServersAaction247Test-1300676990-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.665s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 964.229033] env[61923]: DEBUG oslo_concurrency.lockutils [None req-cd7c0dc9-d2a5-476c-afbb-7be6ac77c41c tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.647s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 964.230720] env[61923]: INFO nova.compute.claims [None req-cd7c0dc9-d2a5-476c-afbb-7be6ac77c41c tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: 3fcedef5-9040-4b45-9e27-f0700112d528] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 964.252971] env[61923]: INFO nova.scheduler.client.report [None req-2605b50b-f7ed-4c40-b6af-0f9b593c3112 tempest-ServersAaction247Test-1300676990 tempest-ServersAaction247Test-1300676990-project-member] Deleted allocations for instance 44196b4c-1401-40fa-bd14-04a49947ab15 [ 964.510238] env[61923]: DEBUG oslo_vmware.api [None req-eea05bb2-c691-4804-8c8d-019f8feaa521 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Task: {'id': task-1377880, 'name': MoveVirtualDisk_Task} progress is 40%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.578314] env[61923]: DEBUG nova.compute.manager [None req-9798b6a1-c446-4119-9842-2560d710a905 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: 638bbbbb-827e-4a73-ac24-a3e1b2cd72eb] Start spawning the instance on the hypervisor. {{(pid=61923) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 964.604441] env[61923]: DEBUG nova.virt.hardware [None req-9798b6a1-c446-4119-9842-2560d710a905 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-29T20:07:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-29T20:07:35Z,direct_url=,disk_format='vmdk',id=0f153f63-ae0a-45b1-b7f5-7f9b673c947f,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='4e7b5e3b3c3443c8bd5c70f8a15739f5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-29T20:07:36Z,virtual_size=,visibility=), allow threads: False {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 964.604707] env[61923]: DEBUG nova.virt.hardware [None req-9798b6a1-c446-4119-9842-2560d710a905 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Flavor limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 964.604869] env[61923]: DEBUG nova.virt.hardware [None req-9798b6a1-c446-4119-9842-2560d710a905 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Image limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 964.605066] env[61923]: DEBUG nova.virt.hardware [None req-9798b6a1-c446-4119-9842-2560d710a905 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Flavor pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 964.605219] env[61923]: DEBUG nova.virt.hardware [None req-9798b6a1-c446-4119-9842-2560d710a905 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Image pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 964.605371] env[61923]: DEBUG nova.virt.hardware [None req-9798b6a1-c446-4119-9842-2560d710a905 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 964.605583] env[61923]: DEBUG nova.virt.hardware [None req-9798b6a1-c446-4119-9842-2560d710a905 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 964.605776] env[61923]: DEBUG nova.virt.hardware [None req-9798b6a1-c446-4119-9842-2560d710a905 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 964.605964] env[61923]: DEBUG nova.virt.hardware [None req-9798b6a1-c446-4119-9842-2560d710a905 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Got 1 possible topologies {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 964.606147] env[61923]: DEBUG nova.virt.hardware [None req-9798b6a1-c446-4119-9842-2560d710a905 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 964.606319] env[61923]: DEBUG nova.virt.hardware [None req-9798b6a1-c446-4119-9842-2560d710a905 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 964.607198] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6a81f71-8ac7-44b8-8fa8-63c9ca1e6798 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.615781] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9af5677c-6858-4d83-95a4-3ab3ee5bac0f {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.759890] env[61923]: DEBUG oslo_concurrency.lockutils [None req-2605b50b-f7ed-4c40-b6af-0f9b593c3112 tempest-ServersAaction247Test-1300676990 tempest-ServersAaction247Test-1300676990-project-member] Lock "44196b4c-1401-40fa-bd14-04a49947ab15" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 9.140s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 964.974865] env[61923]: DEBUG nova.compute.manager [req-354a8fc0-8300-4ac3-95fe-1e6bba5faafc req-c9c72eab-c9ea-431d-a88c-7ebd81656e27 service nova] [instance: 638bbbbb-827e-4a73-ac24-a3e1b2cd72eb] Received event network-vif-plugged-3e6f05d1-ae7a-4aff-9243-20d5dd0e9b4f {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 964.974865] env[61923]: DEBUG oslo_concurrency.lockutils [req-354a8fc0-8300-4ac3-95fe-1e6bba5faafc req-c9c72eab-c9ea-431d-a88c-7ebd81656e27 service nova] Acquiring lock "638bbbbb-827e-4a73-ac24-a3e1b2cd72eb-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 964.974969] env[61923]: DEBUG oslo_concurrency.lockutils [req-354a8fc0-8300-4ac3-95fe-1e6bba5faafc req-c9c72eab-c9ea-431d-a88c-7ebd81656e27 service nova] Lock "638bbbbb-827e-4a73-ac24-a3e1b2cd72eb-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 964.975123] env[61923]: DEBUG oslo_concurrency.lockutils [req-354a8fc0-8300-4ac3-95fe-1e6bba5faafc req-c9c72eab-c9ea-431d-a88c-7ebd81656e27 service nova] Lock "638bbbbb-827e-4a73-ac24-a3e1b2cd72eb-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 964.975295] env[61923]: DEBUG nova.compute.manager [req-354a8fc0-8300-4ac3-95fe-1e6bba5faafc req-c9c72eab-c9ea-431d-a88c-7ebd81656e27 service nova] [instance: 638bbbbb-827e-4a73-ac24-a3e1b2cd72eb] No waiting events found dispatching network-vif-plugged-3e6f05d1-ae7a-4aff-9243-20d5dd0e9b4f {{(pid=61923) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 964.975461] env[61923]: WARNING nova.compute.manager [req-354a8fc0-8300-4ac3-95fe-1e6bba5faafc req-c9c72eab-c9ea-431d-a88c-7ebd81656e27 service nova] [instance: 638bbbbb-827e-4a73-ac24-a3e1b2cd72eb] Received unexpected event network-vif-plugged-3e6f05d1-ae7a-4aff-9243-20d5dd0e9b4f for instance with vm_state building and task_state spawning. [ 965.010745] env[61923]: DEBUG oslo_vmware.api [None req-eea05bb2-c691-4804-8c8d-019f8feaa521 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Task: {'id': task-1377880, 'name': MoveVirtualDisk_Task} progress is 40%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.070633] env[61923]: DEBUG nova.network.neutron [None req-9798b6a1-c446-4119-9842-2560d710a905 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: 638bbbbb-827e-4a73-ac24-a3e1b2cd72eb] Successfully updated port: 3e6f05d1-ae7a-4aff-9243-20d5dd0e9b4f {{(pid=61923) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 965.313603] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9855463c-5b93-40da-98e3-5c02f994725e {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.321675] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff054541-3641-45bb-a0ec-b324898cc808 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.352275] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-460c2561-d903-4ca8-9e0f-2ca3cdce44e7 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.359693] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4137f6db-c174-4474-96d8-ff7a9ec33f74 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.373258] env[61923]: DEBUG nova.compute.provider_tree [None req-cd7c0dc9-d2a5-476c-afbb-7be6ac77c41c tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 965.512151] env[61923]: DEBUG oslo_vmware.api [None req-eea05bb2-c691-4804-8c8d-019f8feaa521 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Task: {'id': task-1377880, 'name': MoveVirtualDisk_Task} progress is 43%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.574225] env[61923]: DEBUG oslo_concurrency.lockutils [None req-9798b6a1-c446-4119-9842-2560d710a905 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Acquiring lock "refresh_cache-638bbbbb-827e-4a73-ac24-a3e1b2cd72eb" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 965.574409] env[61923]: DEBUG oslo_concurrency.lockutils [None req-9798b6a1-c446-4119-9842-2560d710a905 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Acquired lock "refresh_cache-638bbbbb-827e-4a73-ac24-a3e1b2cd72eb" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 965.574639] env[61923]: DEBUG nova.network.neutron [None req-9798b6a1-c446-4119-9842-2560d710a905 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: 638bbbbb-827e-4a73-ac24-a3e1b2cd72eb] Building network info cache for instance {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 965.876685] env[61923]: DEBUG nova.scheduler.client.report [None req-cd7c0dc9-d2a5-476c-afbb-7be6ac77c41c tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 966.087358] env[61923]: DEBUG oslo_vmware.api [None req-eea05bb2-c691-4804-8c8d-019f8feaa521 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Task: {'id': task-1377880, 'name': MoveVirtualDisk_Task} progress is 46%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.107304] env[61923]: DEBUG nova.network.neutron [None req-9798b6a1-c446-4119-9842-2560d710a905 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: 638bbbbb-827e-4a73-ac24-a3e1b2cd72eb] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 966.256601] env[61923]: DEBUG nova.network.neutron [None req-9798b6a1-c446-4119-9842-2560d710a905 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: 638bbbbb-827e-4a73-ac24-a3e1b2cd72eb] Updating instance_info_cache with network_info: [{"id": "3e6f05d1-ae7a-4aff-9243-20d5dd0e9b4f", "address": "fa:16:3e:ac:37:ed", "network": {"id": "d1c38272-0461-43e5-93ae-3964346f77d0", "bridge": "br-int", "label": "tempest-ServersTestJSON-151839165-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "69cc941eb6dd4780ac12aa29656c37f7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8ba3bd22-c936-470e-89bd-b3a5587e87a0", "external-id": "nsx-vlan-transportzone-605", "segmentation_id": 605, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3e6f05d1-ae", "ovs_interfaceid": "3e6f05d1-ae7a-4aff-9243-20d5dd0e9b4f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 966.382431] env[61923]: DEBUG oslo_concurrency.lockutils [None req-cd7c0dc9-d2a5-476c-afbb-7be6ac77c41c tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.153s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 966.382973] env[61923]: DEBUG nova.compute.manager [None req-cd7c0dc9-d2a5-476c-afbb-7be6ac77c41c tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: 3fcedef5-9040-4b45-9e27-f0700112d528] Start building networks asynchronously for instance. {{(pid=61923) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 966.513902] env[61923]: DEBUG oslo_vmware.api [None req-eea05bb2-c691-4804-8c8d-019f8feaa521 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Task: {'id': task-1377880, 'name': MoveVirtualDisk_Task} progress is 54%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.760101] env[61923]: DEBUG oslo_concurrency.lockutils [None req-9798b6a1-c446-4119-9842-2560d710a905 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Releasing lock "refresh_cache-638bbbbb-827e-4a73-ac24-a3e1b2cd72eb" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 966.760353] env[61923]: DEBUG nova.compute.manager [None req-9798b6a1-c446-4119-9842-2560d710a905 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: 638bbbbb-827e-4a73-ac24-a3e1b2cd72eb] Instance network_info: |[{"id": "3e6f05d1-ae7a-4aff-9243-20d5dd0e9b4f", "address": "fa:16:3e:ac:37:ed", "network": {"id": "d1c38272-0461-43e5-93ae-3964346f77d0", "bridge": "br-int", "label": "tempest-ServersTestJSON-151839165-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "69cc941eb6dd4780ac12aa29656c37f7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8ba3bd22-c936-470e-89bd-b3a5587e87a0", "external-id": "nsx-vlan-transportzone-605", "segmentation_id": 605, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3e6f05d1-ae", "ovs_interfaceid": "3e6f05d1-ae7a-4aff-9243-20d5dd0e9b4f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61923) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 966.760910] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-9798b6a1-c446-4119-9842-2560d710a905 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: 638bbbbb-827e-4a73-ac24-a3e1b2cd72eb] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ac:37:ed', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '8ba3bd22-c936-470e-89bd-b3a5587e87a0', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '3e6f05d1-ae7a-4aff-9243-20d5dd0e9b4f', 'vif_model': 'vmxnet3'}] {{(pid=61923) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 966.769862] env[61923]: DEBUG oslo.service.loopingcall [None req-9798b6a1-c446-4119-9842-2560d710a905 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61923) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 966.770134] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 638bbbbb-827e-4a73-ac24-a3e1b2cd72eb] Creating VM on the ESX host {{(pid=61923) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 966.770379] env[61923]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-bfa89b61-6167-437f-b034-bc61b2609cfb {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.788457] env[61923]: DEBUG oslo_service.periodic_task [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61923) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 966.788758] env[61923]: DEBUG oslo_service.periodic_task [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=61923) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 966.792482] env[61923]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 966.792482] env[61923]: value = "task-1377881" [ 966.792482] env[61923]: _type = "Task" [ 966.792482] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 966.807847] env[61923]: DEBUG oslo_vmware.api [-] Task: {'id': task-1377881, 'name': CreateVM_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.887763] env[61923]: DEBUG nova.compute.utils [None req-cd7c0dc9-d2a5-476c-afbb-7be6ac77c41c tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Using /dev/sd instead of None {{(pid=61923) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 966.889398] env[61923]: DEBUG nova.compute.manager [None req-cd7c0dc9-d2a5-476c-afbb-7be6ac77c41c tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: 3fcedef5-9040-4b45-9e27-f0700112d528] Allocating IP information in the background. {{(pid=61923) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 966.889698] env[61923]: DEBUG nova.network.neutron [None req-cd7c0dc9-d2a5-476c-afbb-7be6ac77c41c tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: 3fcedef5-9040-4b45-9e27-f0700112d528] allocate_for_instance() {{(pid=61923) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 966.929655] env[61923]: DEBUG nova.policy [None req-cd7c0dc9-d2a5-476c-afbb-7be6ac77c41c tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '69be89006ff141b686261d643b3b9a73', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '780d3541d9604417b977bb62390c4299', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61923) authorize /opt/stack/nova/nova/policy.py:201}} [ 967.001430] env[61923]: DEBUG nova.compute.manager [req-3cd8d5ec-a190-4c42-a736-05072e688b77 req-d774c5a3-83e5-46c4-b9af-eb46516b43a2 service nova] [instance: 638bbbbb-827e-4a73-ac24-a3e1b2cd72eb] Received event network-changed-3e6f05d1-ae7a-4aff-9243-20d5dd0e9b4f {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 967.001656] env[61923]: DEBUG nova.compute.manager [req-3cd8d5ec-a190-4c42-a736-05072e688b77 req-d774c5a3-83e5-46c4-b9af-eb46516b43a2 service nova] [instance: 638bbbbb-827e-4a73-ac24-a3e1b2cd72eb] Refreshing instance network info cache due to event network-changed-3e6f05d1-ae7a-4aff-9243-20d5dd0e9b4f. {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 967.001939] env[61923]: DEBUG oslo_concurrency.lockutils [req-3cd8d5ec-a190-4c42-a736-05072e688b77 req-d774c5a3-83e5-46c4-b9af-eb46516b43a2 service nova] Acquiring lock "refresh_cache-638bbbbb-827e-4a73-ac24-a3e1b2cd72eb" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 967.002156] env[61923]: DEBUG oslo_concurrency.lockutils [req-3cd8d5ec-a190-4c42-a736-05072e688b77 req-d774c5a3-83e5-46c4-b9af-eb46516b43a2 service nova] Acquired lock "refresh_cache-638bbbbb-827e-4a73-ac24-a3e1b2cd72eb" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 967.002337] env[61923]: DEBUG nova.network.neutron [req-3cd8d5ec-a190-4c42-a736-05072e688b77 req-d774c5a3-83e5-46c4-b9af-eb46516b43a2 service nova] [instance: 638bbbbb-827e-4a73-ac24-a3e1b2cd72eb] Refreshing network info cache for port 3e6f05d1-ae7a-4aff-9243-20d5dd0e9b4f {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 967.014827] env[61923]: DEBUG oslo_vmware.api [None req-eea05bb2-c691-4804-8c8d-019f8feaa521 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Task: {'id': task-1377880, 'name': MoveVirtualDisk_Task} progress is 63%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 967.302704] env[61923]: DEBUG oslo_service.periodic_task [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61923) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 967.302825] env[61923]: DEBUG nova.compute.manager [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Starting heal instance info cache {{(pid=61923) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 967.302947] env[61923]: DEBUG nova.compute.manager [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Rebuilding the list of instances to heal {{(pid=61923) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 967.305940] env[61923]: DEBUG nova.network.neutron [None req-cd7c0dc9-d2a5-476c-afbb-7be6ac77c41c tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: 3fcedef5-9040-4b45-9e27-f0700112d528] Successfully created port: a1c503e4-c553-4bfb-b9ae-84d4ebb6937a {{(pid=61923) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 967.311229] env[61923]: DEBUG oslo_vmware.api [-] Task: {'id': task-1377881, 'name': CreateVM_Task} progress is 25%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 967.393630] env[61923]: DEBUG nova.compute.manager [None req-cd7c0dc9-d2a5-476c-afbb-7be6ac77c41c tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: 3fcedef5-9040-4b45-9e27-f0700112d528] Start building block device mappings for instance. {{(pid=61923) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 967.517343] env[61923]: DEBUG oslo_vmware.api [None req-eea05bb2-c691-4804-8c8d-019f8feaa521 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Task: {'id': task-1377880, 'name': MoveVirtualDisk_Task} progress is 77%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 967.809474] env[61923]: DEBUG nova.compute.manager [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] [instance: 638bbbbb-827e-4a73-ac24-a3e1b2cd72eb] Skipping network cache update for instance because it is Building. {{(pid=61923) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 967.809699] env[61923]: DEBUG nova.compute.manager [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] [instance: 3fcedef5-9040-4b45-9e27-f0700112d528] Skipping network cache update for instance because it is Building. {{(pid=61923) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 967.811818] env[61923]: DEBUG oslo_vmware.api [-] Task: {'id': task-1377881, 'name': CreateVM_Task} progress is 25%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 967.830550] env[61923]: DEBUG nova.network.neutron [req-3cd8d5ec-a190-4c42-a736-05072e688b77 req-d774c5a3-83e5-46c4-b9af-eb46516b43a2 service nova] [instance: 638bbbbb-827e-4a73-ac24-a3e1b2cd72eb] Updated VIF entry in instance network info cache for port 3e6f05d1-ae7a-4aff-9243-20d5dd0e9b4f. {{(pid=61923) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 967.830996] env[61923]: DEBUG nova.network.neutron [req-3cd8d5ec-a190-4c42-a736-05072e688b77 req-d774c5a3-83e5-46c4-b9af-eb46516b43a2 service nova] [instance: 638bbbbb-827e-4a73-ac24-a3e1b2cd72eb] Updating instance_info_cache with network_info: [{"id": "3e6f05d1-ae7a-4aff-9243-20d5dd0e9b4f", "address": "fa:16:3e:ac:37:ed", "network": {"id": "d1c38272-0461-43e5-93ae-3964346f77d0", "bridge": "br-int", "label": "tempest-ServersTestJSON-151839165-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "69cc941eb6dd4780ac12aa29656c37f7", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8ba3bd22-c936-470e-89bd-b3a5587e87a0", "external-id": "nsx-vlan-transportzone-605", "segmentation_id": 605, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3e6f05d1-ae", "ovs_interfaceid": "3e6f05d1-ae7a-4aff-9243-20d5dd0e9b4f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 967.842523] env[61923]: DEBUG oslo_concurrency.lockutils [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Acquiring lock "refresh_cache-b779d183-89ae-4e4d-ae99-e514e145ed43" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 967.842670] env[61923]: DEBUG oslo_concurrency.lockutils [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Acquired lock "refresh_cache-b779d183-89ae-4e4d-ae99-e514e145ed43" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 967.842809] env[61923]: DEBUG nova.network.neutron [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] [instance: b779d183-89ae-4e4d-ae99-e514e145ed43] Forcefully refreshing network info cache for instance {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 967.842968] env[61923]: DEBUG nova.objects.instance [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Lazy-loading 'info_cache' on Instance uuid b779d183-89ae-4e4d-ae99-e514e145ed43 {{(pid=61923) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 968.017893] env[61923]: DEBUG oslo_vmware.api [None req-eea05bb2-c691-4804-8c8d-019f8feaa521 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Task: {'id': task-1377880, 'name': MoveVirtualDisk_Task} progress is 94%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.305560] env[61923]: DEBUG oslo_vmware.api [-] Task: {'id': task-1377881, 'name': CreateVM_Task} progress is 99%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.334172] env[61923]: DEBUG oslo_concurrency.lockutils [req-3cd8d5ec-a190-4c42-a736-05072e688b77 req-d774c5a3-83e5-46c4-b9af-eb46516b43a2 service nova] Releasing lock "refresh_cache-638bbbbb-827e-4a73-ac24-a3e1b2cd72eb" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 968.406378] env[61923]: DEBUG nova.compute.manager [None req-cd7c0dc9-d2a5-476c-afbb-7be6ac77c41c tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: 3fcedef5-9040-4b45-9e27-f0700112d528] Start spawning the instance on the hypervisor. {{(pid=61923) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 968.431738] env[61923]: DEBUG nova.virt.hardware [None req-cd7c0dc9-d2a5-476c-afbb-7be6ac77c41c tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-29T20:07:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-29T20:07:35Z,direct_url=,disk_format='vmdk',id=0f153f63-ae0a-45b1-b7f5-7f9b673c947f,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='4e7b5e3b3c3443c8bd5c70f8a15739f5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-29T20:07:36Z,virtual_size=,visibility=), allow threads: False {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 968.432070] env[61923]: DEBUG nova.virt.hardware [None req-cd7c0dc9-d2a5-476c-afbb-7be6ac77c41c tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Flavor limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 968.432262] env[61923]: DEBUG nova.virt.hardware [None req-cd7c0dc9-d2a5-476c-afbb-7be6ac77c41c tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Image limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 968.432461] env[61923]: DEBUG nova.virt.hardware [None req-cd7c0dc9-d2a5-476c-afbb-7be6ac77c41c tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Flavor pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 968.432614] env[61923]: DEBUG nova.virt.hardware [None req-cd7c0dc9-d2a5-476c-afbb-7be6ac77c41c tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Image pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 968.432765] env[61923]: DEBUG nova.virt.hardware [None req-cd7c0dc9-d2a5-476c-afbb-7be6ac77c41c tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 968.432981] env[61923]: DEBUG nova.virt.hardware [None req-cd7c0dc9-d2a5-476c-afbb-7be6ac77c41c tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 968.433159] env[61923]: DEBUG nova.virt.hardware [None req-cd7c0dc9-d2a5-476c-afbb-7be6ac77c41c tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 968.433335] env[61923]: DEBUG nova.virt.hardware [None req-cd7c0dc9-d2a5-476c-afbb-7be6ac77c41c tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Got 1 possible topologies {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 968.433499] env[61923]: DEBUG nova.virt.hardware [None req-cd7c0dc9-d2a5-476c-afbb-7be6ac77c41c tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 968.433676] env[61923]: DEBUG nova.virt.hardware [None req-cd7c0dc9-d2a5-476c-afbb-7be6ac77c41c tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 968.434632] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a7f6d19-bcb1-4a15-91cb-0d4e9f448d03 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.445806] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2fe9efd-7feb-4b1b-a87e-d50d6ef486b0 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.517974] env[61923]: DEBUG oslo_vmware.api [None req-eea05bb2-c691-4804-8c8d-019f8feaa521 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Task: {'id': task-1377880, 'name': MoveVirtualDisk_Task, 'duration_secs': 5.21096} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 968.518261] env[61923]: INFO nova.virt.vmwareapi.ds_util [None req-eea05bb2-c691-4804-8c8d-019f8feaa521 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Moved virtual disk from [datastore1] OSTACK_IMG_0623b35e-63ab-4acd-bf5d-f33dffaf1c79/OSTACK_IMG_0623b35e-63ab-4acd-bf5d-f33dffaf1c79.vmdk to [datastore1] devstack-image-cache_base/5f919f0b-0e82-4adc-b49b-362960821be5/5f919f0b-0e82-4adc-b49b-362960821be5.vmdk. [ 968.518461] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-eea05bb2-c691-4804-8c8d-019f8feaa521 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] [instance: a511392e-9ab6-42fb-b07c-f90e9786dcc1] Cleaning up location [datastore1] OSTACK_IMG_0623b35e-63ab-4acd-bf5d-f33dffaf1c79 {{(pid=61923) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 968.518626] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-eea05bb2-c691-4804-8c8d-019f8feaa521 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Deleting the datastore file [datastore1] OSTACK_IMG_0623b35e-63ab-4acd-bf5d-f33dffaf1c79 {{(pid=61923) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 968.518900] env[61923]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-9c51a230-1ec8-4c9a-b4b2-8fb389150a71 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.525982] env[61923]: DEBUG oslo_vmware.api [None req-eea05bb2-c691-4804-8c8d-019f8feaa521 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Waiting for the task: (returnval){ [ 968.525982] env[61923]: value = "task-1377882" [ 968.525982] env[61923]: _type = "Task" [ 968.525982] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 968.533988] env[61923]: DEBUG oslo_vmware.api [None req-eea05bb2-c691-4804-8c8d-019f8feaa521 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Task: {'id': task-1377882, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.808303] env[61923]: DEBUG oslo_vmware.api [-] Task: {'id': task-1377881, 'name': CreateVM_Task, 'duration_secs': 1.527443} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 968.808778] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 638bbbbb-827e-4a73-ac24-a3e1b2cd72eb] Created VM on the ESX host {{(pid=61923) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 968.809282] env[61923]: DEBUG oslo_concurrency.lockutils [None req-9798b6a1-c446-4119-9842-2560d710a905 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 968.809482] env[61923]: DEBUG oslo_concurrency.lockutils [None req-9798b6a1-c446-4119-9842-2560d710a905 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 968.809926] env[61923]: DEBUG oslo_concurrency.lockutils [None req-9798b6a1-c446-4119-9842-2560d710a905 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 968.810211] env[61923]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3154e38e-39b9-4a21-97ed-383bf216e2d1 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.815136] env[61923]: DEBUG oslo_vmware.api [None req-9798b6a1-c446-4119-9842-2560d710a905 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Waiting for the task: (returnval){ [ 968.815136] env[61923]: value = "session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52b2d54c-b9d5-ab8c-8273-fcd3bfc69bd7" [ 968.815136] env[61923]: _type = "Task" [ 968.815136] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 968.822567] env[61923]: DEBUG oslo_vmware.api [None req-9798b6a1-c446-4119-9842-2560d710a905 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52b2d54c-b9d5-ab8c-8273-fcd3bfc69bd7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.005177] env[61923]: DEBUG nova.network.neutron [None req-cd7c0dc9-d2a5-476c-afbb-7be6ac77c41c tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: 3fcedef5-9040-4b45-9e27-f0700112d528] Successfully updated port: a1c503e4-c553-4bfb-b9ae-84d4ebb6937a {{(pid=61923) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 969.025902] env[61923]: DEBUG nova.compute.manager [req-2c447bf7-f751-48ec-82fd-34e28be2bb05 req-64d36f8b-a4b6-4403-b493-9bcff1225af5 service nova] [instance: 3fcedef5-9040-4b45-9e27-f0700112d528] Received event network-vif-plugged-a1c503e4-c553-4bfb-b9ae-84d4ebb6937a {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 969.026191] env[61923]: DEBUG oslo_concurrency.lockutils [req-2c447bf7-f751-48ec-82fd-34e28be2bb05 req-64d36f8b-a4b6-4403-b493-9bcff1225af5 service nova] Acquiring lock "3fcedef5-9040-4b45-9e27-f0700112d528-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 969.026383] env[61923]: DEBUG oslo_concurrency.lockutils [req-2c447bf7-f751-48ec-82fd-34e28be2bb05 req-64d36f8b-a4b6-4403-b493-9bcff1225af5 service nova] Lock "3fcedef5-9040-4b45-9e27-f0700112d528-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 969.026516] env[61923]: DEBUG oslo_concurrency.lockutils [req-2c447bf7-f751-48ec-82fd-34e28be2bb05 req-64d36f8b-a4b6-4403-b493-9bcff1225af5 service nova] Lock "3fcedef5-9040-4b45-9e27-f0700112d528-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 969.026691] env[61923]: DEBUG nova.compute.manager [req-2c447bf7-f751-48ec-82fd-34e28be2bb05 req-64d36f8b-a4b6-4403-b493-9bcff1225af5 service nova] [instance: 3fcedef5-9040-4b45-9e27-f0700112d528] No waiting events found dispatching network-vif-plugged-a1c503e4-c553-4bfb-b9ae-84d4ebb6937a {{(pid=61923) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 969.026861] env[61923]: WARNING nova.compute.manager [req-2c447bf7-f751-48ec-82fd-34e28be2bb05 req-64d36f8b-a4b6-4403-b493-9bcff1225af5 service nova] [instance: 3fcedef5-9040-4b45-9e27-f0700112d528] Received unexpected event network-vif-plugged-a1c503e4-c553-4bfb-b9ae-84d4ebb6937a for instance with vm_state building and task_state spawning. [ 969.027128] env[61923]: DEBUG nova.compute.manager [req-2c447bf7-f751-48ec-82fd-34e28be2bb05 req-64d36f8b-a4b6-4403-b493-9bcff1225af5 service nova] [instance: 3fcedef5-9040-4b45-9e27-f0700112d528] Received event network-changed-a1c503e4-c553-4bfb-b9ae-84d4ebb6937a {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 969.027348] env[61923]: DEBUG nova.compute.manager [req-2c447bf7-f751-48ec-82fd-34e28be2bb05 req-64d36f8b-a4b6-4403-b493-9bcff1225af5 service nova] [instance: 3fcedef5-9040-4b45-9e27-f0700112d528] Refreshing instance network info cache due to event network-changed-a1c503e4-c553-4bfb-b9ae-84d4ebb6937a. {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 969.027545] env[61923]: DEBUG oslo_concurrency.lockutils [req-2c447bf7-f751-48ec-82fd-34e28be2bb05 req-64d36f8b-a4b6-4403-b493-9bcff1225af5 service nova] Acquiring lock "refresh_cache-3fcedef5-9040-4b45-9e27-f0700112d528" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 969.027709] env[61923]: DEBUG oslo_concurrency.lockutils [req-2c447bf7-f751-48ec-82fd-34e28be2bb05 req-64d36f8b-a4b6-4403-b493-9bcff1225af5 service nova] Acquired lock "refresh_cache-3fcedef5-9040-4b45-9e27-f0700112d528" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 969.027881] env[61923]: DEBUG nova.network.neutron [req-2c447bf7-f751-48ec-82fd-34e28be2bb05 req-64d36f8b-a4b6-4403-b493-9bcff1225af5 service nova] [instance: 3fcedef5-9040-4b45-9e27-f0700112d528] Refreshing network info cache for port a1c503e4-c553-4bfb-b9ae-84d4ebb6937a {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 969.039373] env[61923]: DEBUG oslo_vmware.api [None req-eea05bb2-c691-4804-8c8d-019f8feaa521 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Task: {'id': task-1377882, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.042574} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 969.039653] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-eea05bb2-c691-4804-8c8d-019f8feaa521 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Deleted the datastore file {{(pid=61923) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 969.039823] env[61923]: DEBUG oslo_concurrency.lockutils [None req-eea05bb2-c691-4804-8c8d-019f8feaa521 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Releasing lock "[datastore1] devstack-image-cache_base/5f919f0b-0e82-4adc-b49b-362960821be5/5f919f0b-0e82-4adc-b49b-362960821be5.vmdk" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 969.040072] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-eea05bb2-c691-4804-8c8d-019f8feaa521 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/5f919f0b-0e82-4adc-b49b-362960821be5/5f919f0b-0e82-4adc-b49b-362960821be5.vmdk to [datastore1] a511392e-9ab6-42fb-b07c-f90e9786dcc1/a511392e-9ab6-42fb-b07c-f90e9786dcc1.vmdk {{(pid=61923) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 969.040316] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-03036901-7151-4f45-b94f-497702cf5c41 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.048101] env[61923]: DEBUG oslo_vmware.api [None req-eea05bb2-c691-4804-8c8d-019f8feaa521 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Waiting for the task: (returnval){ [ 969.048101] env[61923]: value = "task-1377883" [ 969.048101] env[61923]: _type = "Task" [ 969.048101] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 969.056012] env[61923]: DEBUG oslo_vmware.api [None req-eea05bb2-c691-4804-8c8d-019f8feaa521 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Task: {'id': task-1377883, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.146162] env[61923]: DEBUG oslo_concurrency.lockutils [None req-aab527a1-a3a6-41ef-8ac0-681040f3cb52 tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Acquiring lock "b779d183-89ae-4e4d-ae99-e514e145ed43" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 969.146446] env[61923]: DEBUG oslo_concurrency.lockutils [None req-aab527a1-a3a6-41ef-8ac0-681040f3cb52 tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Lock "b779d183-89ae-4e4d-ae99-e514e145ed43" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.001s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 969.326559] env[61923]: DEBUG oslo_vmware.api [None req-9798b6a1-c446-4119-9842-2560d710a905 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52b2d54c-b9d5-ab8c-8273-fcd3bfc69bd7, 'name': SearchDatastore_Task, 'duration_secs': 0.009983} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 969.326894] env[61923]: DEBUG oslo_concurrency.lockutils [None req-9798b6a1-c446-4119-9842-2560d710a905 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 969.327146] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-9798b6a1-c446-4119-9842-2560d710a905 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: 638bbbbb-827e-4a73-ac24-a3e1b2cd72eb] Processing image 0f153f63-ae0a-45b1-b7f5-7f9b673c947f {{(pid=61923) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 969.327395] env[61923]: DEBUG oslo_concurrency.lockutils [None req-9798b6a1-c446-4119-9842-2560d710a905 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 969.327548] env[61923]: DEBUG oslo_concurrency.lockutils [None req-9798b6a1-c446-4119-9842-2560d710a905 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 969.327767] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-9798b6a1-c446-4119-9842-2560d710a905 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61923) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 969.328060] env[61923]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-67ccdc1b-1e4b-48f8-bd57-b770439d35b0 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.339656] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-9798b6a1-c446-4119-9842-2560d710a905 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61923) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 969.339975] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-9798b6a1-c446-4119-9842-2560d710a905 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61923) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 969.340869] env[61923]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-62505ae6-0376-461c-a1a4-46d6a7fa2661 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.347419] env[61923]: DEBUG oslo_vmware.api [None req-9798b6a1-c446-4119-9842-2560d710a905 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Waiting for the task: (returnval){ [ 969.347419] env[61923]: value = "session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52764905-06ec-ddfa-1fc0-5ad72f91982f" [ 969.347419] env[61923]: _type = "Task" [ 969.347419] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 969.356563] env[61923]: DEBUG oslo_vmware.api [None req-9798b6a1-c446-4119-9842-2560d710a905 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52764905-06ec-ddfa-1fc0-5ad72f91982f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.505650] env[61923]: DEBUG oslo_concurrency.lockutils [None req-cd7c0dc9-d2a5-476c-afbb-7be6ac77c41c tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Acquiring lock "refresh_cache-3fcedef5-9040-4b45-9e27-f0700112d528" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 969.563099] env[61923]: DEBUG oslo_vmware.api [None req-eea05bb2-c691-4804-8c8d-019f8feaa521 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Task: {'id': task-1377883, 'name': CopyVirtualDisk_Task} progress is 24%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.563099] env[61923]: DEBUG nova.network.neutron [req-2c447bf7-f751-48ec-82fd-34e28be2bb05 req-64d36f8b-a4b6-4403-b493-9bcff1225af5 service nova] [instance: 3fcedef5-9040-4b45-9e27-f0700112d528] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 969.627165] env[61923]: DEBUG nova.network.neutron [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] [instance: b779d183-89ae-4e4d-ae99-e514e145ed43] Updating instance_info_cache with network_info: [{"id": "d49a3010-c84a-4bef-beed-9ca4aa7d5dc2", "address": "fa:16:3e:48:e6:a2", "network": {"id": "d2905082-7ca4-4f92-a142-bb85211bb7c4", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-885218622-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.251", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a13d827933fa4597984afdb91dbbdd39", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "03ac2c9c-6ad2-4a85-bfab-c7e336df859a", "external-id": "nsx-vlan-transportzone-379", "segmentation_id": 379, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd49a3010-c8", "ovs_interfaceid": "d49a3010-c84a-4bef-beed-9ca4aa7d5dc2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 969.644020] env[61923]: DEBUG nova.network.neutron [req-2c447bf7-f751-48ec-82fd-34e28be2bb05 req-64d36f8b-a4b6-4403-b493-9bcff1225af5 service nova] [instance: 3fcedef5-9040-4b45-9e27-f0700112d528] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 969.649401] env[61923]: INFO nova.compute.manager [None req-aab527a1-a3a6-41ef-8ac0-681040f3cb52 tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] [instance: b779d183-89ae-4e4d-ae99-e514e145ed43] Detaching volume 84fb906f-352f-4eae-b044-c86e8dda3254 [ 969.686227] env[61923]: INFO nova.virt.block_device [None req-aab527a1-a3a6-41ef-8ac0-681040f3cb52 tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] [instance: b779d183-89ae-4e4d-ae99-e514e145ed43] Attempting to driver detach volume 84fb906f-352f-4eae-b044-c86e8dda3254 from mountpoint /dev/sdb [ 969.686227] env[61923]: DEBUG nova.virt.vmwareapi.volumeops [None req-aab527a1-a3a6-41ef-8ac0-681040f3cb52 tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] [instance: b779d183-89ae-4e4d-ae99-e514e145ed43] Volume detach. Driver type: vmdk {{(pid=61923) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 969.686443] env[61923]: DEBUG nova.virt.vmwareapi.volumeops [None req-aab527a1-a3a6-41ef-8ac0-681040f3cb52 tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] [instance: b779d183-89ae-4e4d-ae99-e514e145ed43] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-292719', 'volume_id': '84fb906f-352f-4eae-b044-c86e8dda3254', 'name': 'volume-84fb906f-352f-4eae-b044-c86e8dda3254', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'b779d183-89ae-4e4d-ae99-e514e145ed43', 'attached_at': '', 'detached_at': '', 'volume_id': '84fb906f-352f-4eae-b044-c86e8dda3254', 'serial': '84fb906f-352f-4eae-b044-c86e8dda3254'} {{(pid=61923) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 969.687851] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4eff2a1-e8b7-4043-86c4-a649c42c908d {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.710689] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f6b5e89-3cc2-4a26-8241-f5b261dfb37f {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.722213] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db5ec962-f6d8-4174-bff1-3a15e5975596 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.747547] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0216750a-dd6a-4aa7-9438-e774c09c5d8d {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.764741] env[61923]: DEBUG nova.virt.vmwareapi.volumeops [None req-aab527a1-a3a6-41ef-8ac0-681040f3cb52 tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] The volume has not been displaced from its original location: [datastore1] volume-84fb906f-352f-4eae-b044-c86e8dda3254/volume-84fb906f-352f-4eae-b044-c86e8dda3254.vmdk. No consolidation needed. {{(pid=61923) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 969.770830] env[61923]: DEBUG nova.virt.vmwareapi.volumeops [None req-aab527a1-a3a6-41ef-8ac0-681040f3cb52 tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] [instance: b779d183-89ae-4e4d-ae99-e514e145ed43] Reconfiguring VM instance instance-0000004d to detach disk 2001 {{(pid=61923) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 969.771292] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-cb5b0d58-d581-4198-a992-7919b7f1f3ae {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.792799] env[61923]: DEBUG oslo_vmware.api [None req-aab527a1-a3a6-41ef-8ac0-681040f3cb52 tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Waiting for the task: (returnval){ [ 969.792799] env[61923]: value = "task-1377884" [ 969.792799] env[61923]: _type = "Task" [ 969.792799] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 969.805178] env[61923]: DEBUG oslo_vmware.api [None req-aab527a1-a3a6-41ef-8ac0-681040f3cb52 tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Task: {'id': task-1377884, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.859160] env[61923]: DEBUG oslo_vmware.api [None req-9798b6a1-c446-4119-9842-2560d710a905 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52764905-06ec-ddfa-1fc0-5ad72f91982f, 'name': SearchDatastore_Task, 'duration_secs': 0.096465} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 969.860173] env[61923]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ca856194-ba3a-4982-84ce-8056cdc830e3 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.870139] env[61923]: DEBUG oslo_vmware.api [None req-9798b6a1-c446-4119-9842-2560d710a905 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Waiting for the task: (returnval){ [ 969.870139] env[61923]: value = "session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52cb4eaa-6ce7-8cc0-e234-bb9e1b138d19" [ 969.870139] env[61923]: _type = "Task" [ 969.870139] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 969.881871] env[61923]: DEBUG oslo_vmware.api [None req-9798b6a1-c446-4119-9842-2560d710a905 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52cb4eaa-6ce7-8cc0-e234-bb9e1b138d19, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.060991] env[61923]: DEBUG oslo_vmware.api [None req-eea05bb2-c691-4804-8c8d-019f8feaa521 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Task: {'id': task-1377883, 'name': CopyVirtualDisk_Task} progress is 43%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.130093] env[61923]: DEBUG oslo_concurrency.lockutils [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Releasing lock "refresh_cache-b779d183-89ae-4e4d-ae99-e514e145ed43" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 970.130457] env[61923]: DEBUG nova.compute.manager [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] [instance: b779d183-89ae-4e4d-ae99-e514e145ed43] Updated the network info_cache for instance {{(pid=61923) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9999}} [ 970.130763] env[61923]: DEBUG oslo_service.periodic_task [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61923) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 970.131065] env[61923]: DEBUG oslo_service.periodic_task [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61923) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 970.131341] env[61923]: DEBUG oslo_service.periodic_task [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61923) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 970.131605] env[61923]: DEBUG oslo_service.periodic_task [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61923) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 970.131852] env[61923]: DEBUG oslo_service.periodic_task [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61923) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 970.132075] env[61923]: DEBUG oslo_service.periodic_task [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61923) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 970.132323] env[61923]: DEBUG nova.compute.manager [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61923) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 970.132579] env[61923]: DEBUG oslo_service.periodic_task [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Running periodic task ComputeManager.update_available_resource {{(pid=61923) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 970.146387] env[61923]: DEBUG oslo_concurrency.lockutils [req-2c447bf7-f751-48ec-82fd-34e28be2bb05 req-64d36f8b-a4b6-4403-b493-9bcff1225af5 service nova] Releasing lock "refresh_cache-3fcedef5-9040-4b45-9e27-f0700112d528" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 970.147155] env[61923]: DEBUG oslo_concurrency.lockutils [None req-cd7c0dc9-d2a5-476c-afbb-7be6ac77c41c tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Acquired lock "refresh_cache-3fcedef5-9040-4b45-9e27-f0700112d528" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 970.147367] env[61923]: DEBUG nova.network.neutron [None req-cd7c0dc9-d2a5-476c-afbb-7be6ac77c41c tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: 3fcedef5-9040-4b45-9e27-f0700112d528] Building network info cache for instance {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 970.304186] env[61923]: DEBUG oslo_vmware.api [None req-aab527a1-a3a6-41ef-8ac0-681040f3cb52 tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Task: {'id': task-1377884, 'name': ReconfigVM_Task, 'duration_secs': 0.475877} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 970.304509] env[61923]: DEBUG nova.virt.vmwareapi.volumeops [None req-aab527a1-a3a6-41ef-8ac0-681040f3cb52 tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] [instance: b779d183-89ae-4e4d-ae99-e514e145ed43] Reconfigured VM instance instance-0000004d to detach disk 2001 {{(pid=61923) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 970.309454] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-03713779-692e-47e6-a702-2e62d8147ba5 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.325999] env[61923]: DEBUG oslo_vmware.api [None req-aab527a1-a3a6-41ef-8ac0-681040f3cb52 tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Waiting for the task: (returnval){ [ 970.325999] env[61923]: value = "task-1377885" [ 970.325999] env[61923]: _type = "Task" [ 970.325999] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 970.334966] env[61923]: DEBUG oslo_vmware.api [None req-aab527a1-a3a6-41ef-8ac0-681040f3cb52 tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Task: {'id': task-1377885, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.381892] env[61923]: DEBUG oslo_vmware.api [None req-9798b6a1-c446-4119-9842-2560d710a905 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52cb4eaa-6ce7-8cc0-e234-bb9e1b138d19, 'name': SearchDatastore_Task, 'duration_secs': 0.094358} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 970.382254] env[61923]: DEBUG oslo_concurrency.lockutils [None req-9798b6a1-c446-4119-9842-2560d710a905 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 970.382532] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-9798b6a1-c446-4119-9842-2560d710a905 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk to [datastore1] 638bbbbb-827e-4a73-ac24-a3e1b2cd72eb/638bbbbb-827e-4a73-ac24-a3e1b2cd72eb.vmdk {{(pid=61923) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 970.382844] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-98845b08-d587-41eb-af8b-15f38e02fc21 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.391453] env[61923]: DEBUG oslo_vmware.api [None req-9798b6a1-c446-4119-9842-2560d710a905 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Waiting for the task: (returnval){ [ 970.391453] env[61923]: value = "task-1377886" [ 970.391453] env[61923]: _type = "Task" [ 970.391453] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 970.400319] env[61923]: DEBUG oslo_vmware.api [None req-9798b6a1-c446-4119-9842-2560d710a905 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Task: {'id': task-1377886, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.561059] env[61923]: DEBUG oslo_vmware.api [None req-eea05bb2-c691-4804-8c8d-019f8feaa521 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Task: {'id': task-1377883, 'name': CopyVirtualDisk_Task} progress is 63%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.636286] env[61923]: DEBUG oslo_concurrency.lockutils [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 970.636536] env[61923]: DEBUG oslo_concurrency.lockutils [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 970.636702] env[61923]: DEBUG oslo_concurrency.lockutils [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 970.636900] env[61923]: DEBUG nova.compute.resource_tracker [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61923) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 970.637940] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f445391-1c0a-4560-83f5-72faf092fb4e {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.648889] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5db88ca-be6f-4777-ac6a-0274af8042be {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.672771] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f25b8f6-6826-485e-92b6-1448a7b15c1d {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.682395] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46e94f64-f30e-4099-a2e8-05cfc0edbc54 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.687672] env[61923]: DEBUG nova.network.neutron [None req-cd7c0dc9-d2a5-476c-afbb-7be6ac77c41c tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: 3fcedef5-9040-4b45-9e27-f0700112d528] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 970.717235] env[61923]: DEBUG nova.compute.resource_tracker [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=179704MB free_disk=178GB free_vcpus=48 pci_devices=None {{(pid=61923) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 970.717407] env[61923]: DEBUG oslo_concurrency.lockutils [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 970.717648] env[61923]: DEBUG oslo_concurrency.lockutils [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 970.836616] env[61923]: DEBUG oslo_vmware.api [None req-aab527a1-a3a6-41ef-8ac0-681040f3cb52 tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Task: {'id': task-1377885, 'name': ReconfigVM_Task, 'duration_secs': 0.16864} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 970.836955] env[61923]: DEBUG nova.virt.vmwareapi.volumeops [None req-aab527a1-a3a6-41ef-8ac0-681040f3cb52 tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] [instance: b779d183-89ae-4e4d-ae99-e514e145ed43] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-292719', 'volume_id': '84fb906f-352f-4eae-b044-c86e8dda3254', 'name': 'volume-84fb906f-352f-4eae-b044-c86e8dda3254', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'b779d183-89ae-4e4d-ae99-e514e145ed43', 'attached_at': '', 'detached_at': '', 'volume_id': '84fb906f-352f-4eae-b044-c86e8dda3254', 'serial': '84fb906f-352f-4eae-b044-c86e8dda3254'} {{(pid=61923) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 970.857510] env[61923]: DEBUG nova.network.neutron [None req-cd7c0dc9-d2a5-476c-afbb-7be6ac77c41c tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: 3fcedef5-9040-4b45-9e27-f0700112d528] Updating instance_info_cache with network_info: [{"id": "a1c503e4-c553-4bfb-b9ae-84d4ebb6937a", "address": "fa:16:3e:e1:bd:1a", "network": {"id": "9533e79d-fde2-4c10-86a0-86a36845a8cf", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1736743137-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "780d3541d9604417b977bb62390c4299", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6d650b26-c3e7-4de7-98db-5e4b816d123a", "external-id": "nsx-vlan-transportzone-757", "segmentation_id": 757, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa1c503e4-c5", "ovs_interfaceid": "a1c503e4-c553-4bfb-b9ae-84d4ebb6937a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 970.903023] env[61923]: DEBUG oslo_vmware.api [None req-9798b6a1-c446-4119-9842-2560d710a905 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Task: {'id': task-1377886, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.062150] env[61923]: DEBUG oslo_vmware.api [None req-eea05bb2-c691-4804-8c8d-019f8feaa521 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Task: {'id': task-1377883, 'name': CopyVirtualDisk_Task} progress is 83%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.360244] env[61923]: DEBUG oslo_concurrency.lockutils [None req-cd7c0dc9-d2a5-476c-afbb-7be6ac77c41c tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Releasing lock "refresh_cache-3fcedef5-9040-4b45-9e27-f0700112d528" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 971.360580] env[61923]: DEBUG nova.compute.manager [None req-cd7c0dc9-d2a5-476c-afbb-7be6ac77c41c tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: 3fcedef5-9040-4b45-9e27-f0700112d528] Instance network_info: |[{"id": "a1c503e4-c553-4bfb-b9ae-84d4ebb6937a", "address": "fa:16:3e:e1:bd:1a", "network": {"id": "9533e79d-fde2-4c10-86a0-86a36845a8cf", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1736743137-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "780d3541d9604417b977bb62390c4299", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6d650b26-c3e7-4de7-98db-5e4b816d123a", "external-id": "nsx-vlan-transportzone-757", "segmentation_id": 757, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa1c503e4-c5", "ovs_interfaceid": "a1c503e4-c553-4bfb-b9ae-84d4ebb6937a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61923) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 971.361203] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-cd7c0dc9-d2a5-476c-afbb-7be6ac77c41c tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: 3fcedef5-9040-4b45-9e27-f0700112d528] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:e1:bd:1a', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '6d650b26-c3e7-4de7-98db-5e4b816d123a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'a1c503e4-c553-4bfb-b9ae-84d4ebb6937a', 'vif_model': 'vmxnet3'}] {{(pid=61923) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 971.369305] env[61923]: DEBUG oslo.service.loopingcall [None req-cd7c0dc9-d2a5-476c-afbb-7be6ac77c41c tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61923) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 971.370440] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3fcedef5-9040-4b45-9e27-f0700112d528] Creating VM on the ESX host {{(pid=61923) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 971.370741] env[61923]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-6f349fcb-f453-43c2-a411-882083f862d8 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.387280] env[61923]: DEBUG nova.objects.instance [None req-aab527a1-a3a6-41ef-8ac0-681040f3cb52 tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Lazy-loading 'flavor' on Instance uuid b779d183-89ae-4e4d-ae99-e514e145ed43 {{(pid=61923) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 971.394590] env[61923]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 971.394590] env[61923]: value = "task-1377887" [ 971.394590] env[61923]: _type = "Task" [ 971.394590] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 971.409729] env[61923]: DEBUG oslo_vmware.api [None req-9798b6a1-c446-4119-9842-2560d710a905 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Task: {'id': task-1377886, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.412638] env[61923]: DEBUG oslo_vmware.api [-] Task: {'id': task-1377887, 'name': CreateVM_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.564960] env[61923]: DEBUG oslo_vmware.api [None req-eea05bb2-c691-4804-8c8d-019f8feaa521 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Task: {'id': task-1377883, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.354568} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 971.565382] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-eea05bb2-c691-4804-8c8d-019f8feaa521 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/5f919f0b-0e82-4adc-b49b-362960821be5/5f919f0b-0e82-4adc-b49b-362960821be5.vmdk to [datastore1] a511392e-9ab6-42fb-b07c-f90e9786dcc1/a511392e-9ab6-42fb-b07c-f90e9786dcc1.vmdk {{(pid=61923) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 971.566299] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98e2ba06-7f25-4736-aa25-24bd8b9f0b4b {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.592522] env[61923]: DEBUG nova.virt.vmwareapi.volumeops [None req-eea05bb2-c691-4804-8c8d-019f8feaa521 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] [instance: a511392e-9ab6-42fb-b07c-f90e9786dcc1] Reconfiguring VM instance instance-0000004f to attach disk [datastore1] a511392e-9ab6-42fb-b07c-f90e9786dcc1/a511392e-9ab6-42fb-b07c-f90e9786dcc1.vmdk or device None with type streamOptimized {{(pid=61923) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 971.592944] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a0ddbd42-9b42-4a18-830b-56c18f4b7fb8 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.615756] env[61923]: DEBUG oslo_vmware.api [None req-eea05bb2-c691-4804-8c8d-019f8feaa521 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Waiting for the task: (returnval){ [ 971.615756] env[61923]: value = "task-1377888" [ 971.615756] env[61923]: _type = "Task" [ 971.615756] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 971.625693] env[61923]: DEBUG oslo_vmware.api [None req-eea05bb2-c691-4804-8c8d-019f8feaa521 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Task: {'id': task-1377888, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.752155] env[61923]: DEBUG nova.compute.resource_tracker [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Instance b779d183-89ae-4e4d-ae99-e514e145ed43 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61923) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 971.752566] env[61923]: DEBUG nova.compute.resource_tracker [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Instance aba844c1-d488-4e95-a550-bd1cb7331fb2 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61923) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 971.752566] env[61923]: DEBUG nova.compute.resource_tracker [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Instance a511392e-9ab6-42fb-b07c-f90e9786dcc1 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61923) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 971.752566] env[61923]: DEBUG nova.compute.resource_tracker [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Instance 638bbbbb-827e-4a73-ac24-a3e1b2cd72eb actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61923) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 971.752702] env[61923]: DEBUG nova.compute.resource_tracker [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Instance 3fcedef5-9040-4b45-9e27-f0700112d528 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61923) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 971.752921] env[61923]: DEBUG nova.compute.resource_tracker [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Total usable vcpus: 48, total allocated vcpus: 5 {{(pid=61923) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 971.753072] env[61923]: DEBUG nova.compute.resource_tracker [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1472MB phys_disk=200GB used_disk=5GB total_vcpus=48 used_vcpus=5 pci_stats=[] {{(pid=61923) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 971.837664] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c64e0ada-9aba-4c2d-a1a2-27f618c5ad77 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.845955] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a055dd96-140c-40ab-adeb-d67471c24307 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.877346] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-167b1974-7089-4a51-8363-d3668e001369 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.887955] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ebee79fc-8bd6-43cd-acfe-715c7043318a {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.909582] env[61923]: DEBUG nova.compute.provider_tree [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 971.918239] env[61923]: DEBUG oslo_vmware.api [None req-9798b6a1-c446-4119-9842-2560d710a905 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Task: {'id': task-1377886, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.419447} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 971.918867] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-9798b6a1-c446-4119-9842-2560d710a905 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk to [datastore1] 638bbbbb-827e-4a73-ac24-a3e1b2cd72eb/638bbbbb-827e-4a73-ac24-a3e1b2cd72eb.vmdk {{(pid=61923) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 971.919130] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-9798b6a1-c446-4119-9842-2560d710a905 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: 638bbbbb-827e-4a73-ac24-a3e1b2cd72eb] Extending root virtual disk to 1048576 {{(pid=61923) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 971.919386] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-a4997e45-705f-4e2b-b3d8-1a7639576f10 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.925697] env[61923]: DEBUG oslo_vmware.api [-] Task: {'id': task-1377887, 'name': CreateVM_Task, 'duration_secs': 0.42019} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 971.925697] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3fcedef5-9040-4b45-9e27-f0700112d528] Created VM on the ESX host {{(pid=61923) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 971.926309] env[61923]: DEBUG oslo_concurrency.lockutils [None req-cd7c0dc9-d2a5-476c-afbb-7be6ac77c41c tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 971.926535] env[61923]: DEBUG oslo_concurrency.lockutils [None req-cd7c0dc9-d2a5-476c-afbb-7be6ac77c41c tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 971.927066] env[61923]: DEBUG oslo_concurrency.lockutils [None req-cd7c0dc9-d2a5-476c-afbb-7be6ac77c41c tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 971.927526] env[61923]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c902d883-651b-408e-ae59-1abee6b18910 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.931028] env[61923]: DEBUG oslo_vmware.api [None req-9798b6a1-c446-4119-9842-2560d710a905 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Waiting for the task: (returnval){ [ 971.931028] env[61923]: value = "task-1377889" [ 971.931028] env[61923]: _type = "Task" [ 971.931028] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 971.939023] env[61923]: DEBUG oslo_vmware.api [None req-cd7c0dc9-d2a5-476c-afbb-7be6ac77c41c tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Waiting for the task: (returnval){ [ 971.939023] env[61923]: value = "session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]5220297f-ddfa-013e-283d-9fcdc18148da" [ 971.939023] env[61923]: _type = "Task" [ 971.939023] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 971.943603] env[61923]: DEBUG oslo_vmware.api [None req-9798b6a1-c446-4119-9842-2560d710a905 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Task: {'id': task-1377889, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.950536] env[61923]: DEBUG oslo_vmware.api [None req-cd7c0dc9-d2a5-476c-afbb-7be6ac77c41c tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]5220297f-ddfa-013e-283d-9fcdc18148da, 'name': SearchDatastore_Task, 'duration_secs': 0.011352} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 971.950884] env[61923]: DEBUG oslo_concurrency.lockutils [None req-cd7c0dc9-d2a5-476c-afbb-7be6ac77c41c tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 971.951138] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-cd7c0dc9-d2a5-476c-afbb-7be6ac77c41c tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: 3fcedef5-9040-4b45-9e27-f0700112d528] Processing image 0f153f63-ae0a-45b1-b7f5-7f9b673c947f {{(pid=61923) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 971.951377] env[61923]: DEBUG oslo_concurrency.lockutils [None req-cd7c0dc9-d2a5-476c-afbb-7be6ac77c41c tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 971.951528] env[61923]: DEBUG oslo_concurrency.lockutils [None req-cd7c0dc9-d2a5-476c-afbb-7be6ac77c41c tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 971.951705] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-cd7c0dc9-d2a5-476c-afbb-7be6ac77c41c tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61923) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 971.951974] env[61923]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a7d831c1-fd66-48b2-9fa6-794594122bed {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.982744] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-cd7c0dc9-d2a5-476c-afbb-7be6ac77c41c tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61923) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 971.982956] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-cd7c0dc9-d2a5-476c-afbb-7be6ac77c41c tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61923) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 971.983729] env[61923]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d51b925b-c0fb-463f-acf1-a924ad0c28db {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.990676] env[61923]: DEBUG oslo_vmware.api [None req-cd7c0dc9-d2a5-476c-afbb-7be6ac77c41c tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Waiting for the task: (returnval){ [ 971.990676] env[61923]: value = "session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52d2ebd8-fee1-6551-456b-de709ceaa845" [ 971.990676] env[61923]: _type = "Task" [ 971.990676] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 971.998801] env[61923]: DEBUG oslo_vmware.api [None req-cd7c0dc9-d2a5-476c-afbb-7be6ac77c41c tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52d2ebd8-fee1-6551-456b-de709ceaa845, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.127641] env[61923]: DEBUG oslo_vmware.api [None req-eea05bb2-c691-4804-8c8d-019f8feaa521 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Task: {'id': task-1377888, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.388470] env[61923]: DEBUG oslo_concurrency.lockutils [None req-38f5bd31-2045-4be1-a885-010753cd933f tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Acquiring lock "b779d183-89ae-4e4d-ae99-e514e145ed43" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 972.409248] env[61923]: DEBUG oslo_concurrency.lockutils [None req-aab527a1-a3a6-41ef-8ac0-681040f3cb52 tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Lock "b779d183-89ae-4e4d-ae99-e514e145ed43" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.263s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 972.410952] env[61923]: DEBUG oslo_concurrency.lockutils [None req-38f5bd31-2045-4be1-a885-010753cd933f tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Lock "b779d183-89ae-4e4d-ae99-e514e145ed43" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.023s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 972.411228] env[61923]: DEBUG nova.compute.manager [None req-38f5bd31-2045-4be1-a885-010753cd933f tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] [instance: b779d183-89ae-4e4d-ae99-e514e145ed43] Checking state {{(pid=61923) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 972.412072] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46b71990-8bce-43c0-be8a-a19dad62d7a9 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.415648] env[61923]: DEBUG nova.scheduler.client.report [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 972.424164] env[61923]: DEBUG nova.compute.manager [None req-38f5bd31-2045-4be1-a885-010753cd933f tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] [instance: b779d183-89ae-4e4d-ae99-e514e145ed43] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=61923) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 972.424771] env[61923]: DEBUG nova.objects.instance [None req-38f5bd31-2045-4be1-a885-010753cd933f tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Lazy-loading 'flavor' on Instance uuid b779d183-89ae-4e4d-ae99-e514e145ed43 {{(pid=61923) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 972.442238] env[61923]: DEBUG oslo_vmware.api [None req-9798b6a1-c446-4119-9842-2560d710a905 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Task: {'id': task-1377889, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.238763} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 972.442508] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-9798b6a1-c446-4119-9842-2560d710a905 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: 638bbbbb-827e-4a73-ac24-a3e1b2cd72eb] Extended root virtual disk {{(pid=61923) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 972.443339] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-936cb896-db58-49e5-a67b-3fa7928e62e8 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.465936] env[61923]: DEBUG nova.virt.vmwareapi.volumeops [None req-9798b6a1-c446-4119-9842-2560d710a905 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: 638bbbbb-827e-4a73-ac24-a3e1b2cd72eb] Reconfiguring VM instance instance-0000005c to attach disk [datastore1] 638bbbbb-827e-4a73-ac24-a3e1b2cd72eb/638bbbbb-827e-4a73-ac24-a3e1b2cd72eb.vmdk or device None with type sparse {{(pid=61923) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 972.466206] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0bba5f46-0271-49fe-a469-1ad9510fed02 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.486542] env[61923]: DEBUG oslo_vmware.api [None req-9798b6a1-c446-4119-9842-2560d710a905 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Waiting for the task: (returnval){ [ 972.486542] env[61923]: value = "task-1377890" [ 972.486542] env[61923]: _type = "Task" [ 972.486542] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 972.497262] env[61923]: DEBUG oslo_vmware.api [None req-9798b6a1-c446-4119-9842-2560d710a905 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Task: {'id': task-1377890, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.502301] env[61923]: DEBUG oslo_vmware.api [None req-cd7c0dc9-d2a5-476c-afbb-7be6ac77c41c tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52d2ebd8-fee1-6551-456b-de709ceaa845, 'name': SearchDatastore_Task, 'duration_secs': 0.013019} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 972.503072] env[61923]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f7541fe7-22e9-49f7-adb9-185a1418afdd {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.508292] env[61923]: DEBUG oslo_vmware.api [None req-cd7c0dc9-d2a5-476c-afbb-7be6ac77c41c tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Waiting for the task: (returnval){ [ 972.508292] env[61923]: value = "session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52313e27-3741-c4ee-0b65-7623da79e81b" [ 972.508292] env[61923]: _type = "Task" [ 972.508292] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 972.518736] env[61923]: DEBUG oslo_vmware.api [None req-cd7c0dc9-d2a5-476c-afbb-7be6ac77c41c tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52313e27-3741-c4ee-0b65-7623da79e81b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.627867] env[61923]: DEBUG oslo_vmware.api [None req-eea05bb2-c691-4804-8c8d-019f8feaa521 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Task: {'id': task-1377888, 'name': ReconfigVM_Task, 'duration_secs': 0.913015} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 972.628263] env[61923]: DEBUG nova.virt.vmwareapi.volumeops [None req-eea05bb2-c691-4804-8c8d-019f8feaa521 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] [instance: a511392e-9ab6-42fb-b07c-f90e9786dcc1] Reconfigured VM instance instance-0000004f to attach disk [datastore1] a511392e-9ab6-42fb-b07c-f90e9786dcc1/a511392e-9ab6-42fb-b07c-f90e9786dcc1.vmdk or device None with type streamOptimized {{(pid=61923) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 972.628927] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e6447316-7016-4c28-bfbc-16367d2264f6 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.635934] env[61923]: DEBUG oslo_vmware.api [None req-eea05bb2-c691-4804-8c8d-019f8feaa521 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Waiting for the task: (returnval){ [ 972.635934] env[61923]: value = "task-1377891" [ 972.635934] env[61923]: _type = "Task" [ 972.635934] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 972.645307] env[61923]: DEBUG oslo_vmware.api [None req-eea05bb2-c691-4804-8c8d-019f8feaa521 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Task: {'id': task-1377891, 'name': Rename_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.921654] env[61923]: DEBUG nova.compute.resource_tracker [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61923) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 972.921824] env[61923]: DEBUG oslo_concurrency.lockutils [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.204s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 972.930020] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-38f5bd31-2045-4be1-a885-010753cd933f tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] [instance: b779d183-89ae-4e4d-ae99-e514e145ed43] Powering off the VM {{(pid=61923) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 972.930020] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9bd0f744-eec3-4b0b-b6a4-f5fb5510caa6 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.938465] env[61923]: DEBUG oslo_vmware.api [None req-38f5bd31-2045-4be1-a885-010753cd933f tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Waiting for the task: (returnval){ [ 972.938465] env[61923]: value = "task-1377892" [ 972.938465] env[61923]: _type = "Task" [ 972.938465] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 972.947607] env[61923]: DEBUG oslo_vmware.api [None req-38f5bd31-2045-4be1-a885-010753cd933f tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Task: {'id': task-1377892, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.996833] env[61923]: DEBUG oslo_vmware.api [None req-9798b6a1-c446-4119-9842-2560d710a905 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Task: {'id': task-1377890, 'name': ReconfigVM_Task, 'duration_secs': 0.358715} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 972.997118] env[61923]: DEBUG nova.virt.vmwareapi.volumeops [None req-9798b6a1-c446-4119-9842-2560d710a905 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: 638bbbbb-827e-4a73-ac24-a3e1b2cd72eb] Reconfigured VM instance instance-0000005c to attach disk [datastore1] 638bbbbb-827e-4a73-ac24-a3e1b2cd72eb/638bbbbb-827e-4a73-ac24-a3e1b2cd72eb.vmdk or device None with type sparse {{(pid=61923) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 972.997757] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-350ec447-978d-4ad7-9aba-061b604663bd {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.005161] env[61923]: DEBUG oslo_vmware.api [None req-9798b6a1-c446-4119-9842-2560d710a905 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Waiting for the task: (returnval){ [ 973.005161] env[61923]: value = "task-1377893" [ 973.005161] env[61923]: _type = "Task" [ 973.005161] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 973.015901] env[61923]: DEBUG oslo_vmware.api [None req-9798b6a1-c446-4119-9842-2560d710a905 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Task: {'id': task-1377893, 'name': Rename_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 973.021751] env[61923]: DEBUG oslo_vmware.api [None req-cd7c0dc9-d2a5-476c-afbb-7be6ac77c41c tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52313e27-3741-c4ee-0b65-7623da79e81b, 'name': SearchDatastore_Task, 'duration_secs': 0.013492} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 973.022030] env[61923]: DEBUG oslo_concurrency.lockutils [None req-cd7c0dc9-d2a5-476c-afbb-7be6ac77c41c tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 973.022306] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-cd7c0dc9-d2a5-476c-afbb-7be6ac77c41c tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk to [datastore1] 3fcedef5-9040-4b45-9e27-f0700112d528/3fcedef5-9040-4b45-9e27-f0700112d528.vmdk {{(pid=61923) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 973.022565] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d05b47f8-c0f6-45a3-a877-c6dda7711a55 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.029811] env[61923]: DEBUG oslo_vmware.api [None req-cd7c0dc9-d2a5-476c-afbb-7be6ac77c41c tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Waiting for the task: (returnval){ [ 973.029811] env[61923]: value = "task-1377894" [ 973.029811] env[61923]: _type = "Task" [ 973.029811] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 973.038556] env[61923]: DEBUG oslo_vmware.api [None req-cd7c0dc9-d2a5-476c-afbb-7be6ac77c41c tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Task: {'id': task-1377894, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 973.146811] env[61923]: DEBUG oslo_vmware.api [None req-eea05bb2-c691-4804-8c8d-019f8feaa521 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Task: {'id': task-1377891, 'name': Rename_Task, 'duration_secs': 0.153313} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 973.147137] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-eea05bb2-c691-4804-8c8d-019f8feaa521 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] [instance: a511392e-9ab6-42fb-b07c-f90e9786dcc1] Powering on the VM {{(pid=61923) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 973.147431] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-086734af-a9bd-4292-ab2f-73232acbb738 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.155138] env[61923]: DEBUG oslo_vmware.api [None req-eea05bb2-c691-4804-8c8d-019f8feaa521 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Waiting for the task: (returnval){ [ 973.155138] env[61923]: value = "task-1377895" [ 973.155138] env[61923]: _type = "Task" [ 973.155138] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 973.163216] env[61923]: DEBUG oslo_vmware.api [None req-eea05bb2-c691-4804-8c8d-019f8feaa521 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Task: {'id': task-1377895, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 973.450042] env[61923]: DEBUG oslo_vmware.api [None req-38f5bd31-2045-4be1-a885-010753cd933f tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Task: {'id': task-1377892, 'name': PowerOffVM_Task, 'duration_secs': 0.174983} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 973.450377] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-38f5bd31-2045-4be1-a885-010753cd933f tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] [instance: b779d183-89ae-4e4d-ae99-e514e145ed43] Powered off the VM {{(pid=61923) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 973.450582] env[61923]: DEBUG nova.compute.manager [None req-38f5bd31-2045-4be1-a885-010753cd933f tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] [instance: b779d183-89ae-4e4d-ae99-e514e145ed43] Checking state {{(pid=61923) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 973.451624] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6dbd7ff2-0639-4f2c-b616-07211d84bd1e {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.514641] env[61923]: DEBUG oslo_vmware.api [None req-9798b6a1-c446-4119-9842-2560d710a905 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Task: {'id': task-1377893, 'name': Rename_Task, 'duration_secs': 0.155232} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 973.514912] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-9798b6a1-c446-4119-9842-2560d710a905 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: 638bbbbb-827e-4a73-ac24-a3e1b2cd72eb] Powering on the VM {{(pid=61923) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 973.515171] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-46ce7b0d-b39a-47b8-90a4-ef5ebe87a95f {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.522365] env[61923]: DEBUG oslo_vmware.api [None req-9798b6a1-c446-4119-9842-2560d710a905 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Waiting for the task: (returnval){ [ 973.522365] env[61923]: value = "task-1377896" [ 973.522365] env[61923]: _type = "Task" [ 973.522365] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 973.530758] env[61923]: DEBUG oslo_vmware.api [None req-9798b6a1-c446-4119-9842-2560d710a905 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Task: {'id': task-1377896, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 973.539188] env[61923]: DEBUG oslo_vmware.api [None req-cd7c0dc9-d2a5-476c-afbb-7be6ac77c41c tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Task: {'id': task-1377894, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.455588} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 973.539469] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-cd7c0dc9-d2a5-476c-afbb-7be6ac77c41c tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk to [datastore1] 3fcedef5-9040-4b45-9e27-f0700112d528/3fcedef5-9040-4b45-9e27-f0700112d528.vmdk {{(pid=61923) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 973.539785] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-cd7c0dc9-d2a5-476c-afbb-7be6ac77c41c tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: 3fcedef5-9040-4b45-9e27-f0700112d528] Extending root virtual disk to 1048576 {{(pid=61923) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 973.540063] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-3c0b9db6-9d0a-427b-b77c-a9450e7ac3bf {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.546285] env[61923]: DEBUG oslo_vmware.api [None req-cd7c0dc9-d2a5-476c-afbb-7be6ac77c41c tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Waiting for the task: (returnval){ [ 973.546285] env[61923]: value = "task-1377897" [ 973.546285] env[61923]: _type = "Task" [ 973.546285] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 973.556346] env[61923]: DEBUG oslo_vmware.api [None req-cd7c0dc9-d2a5-476c-afbb-7be6ac77c41c tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Task: {'id': task-1377897, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 973.665704] env[61923]: DEBUG oslo_vmware.api [None req-eea05bb2-c691-4804-8c8d-019f8feaa521 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Task: {'id': task-1377895, 'name': PowerOnVM_Task} progress is 89%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 973.966531] env[61923]: DEBUG oslo_concurrency.lockutils [None req-38f5bd31-2045-4be1-a885-010753cd933f tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Lock "b779d183-89ae-4e4d-ae99-e514e145ed43" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.556s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 974.034349] env[61923]: DEBUG oslo_vmware.api [None req-9798b6a1-c446-4119-9842-2560d710a905 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Task: {'id': task-1377896, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.055307] env[61923]: DEBUG oslo_vmware.api [None req-cd7c0dc9-d2a5-476c-afbb-7be6ac77c41c tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Task: {'id': task-1377897, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.071123} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 974.055582] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-cd7c0dc9-d2a5-476c-afbb-7be6ac77c41c tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: 3fcedef5-9040-4b45-9e27-f0700112d528] Extended root virtual disk {{(pid=61923) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 974.056421] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4aabfa11-6e4e-48f5-bb59-49d8fe984be4 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.078367] env[61923]: DEBUG nova.virt.vmwareapi.volumeops [None req-cd7c0dc9-d2a5-476c-afbb-7be6ac77c41c tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: 3fcedef5-9040-4b45-9e27-f0700112d528] Reconfiguring VM instance instance-0000005d to attach disk [datastore1] 3fcedef5-9040-4b45-9e27-f0700112d528/3fcedef5-9040-4b45-9e27-f0700112d528.vmdk or device None with type sparse {{(pid=61923) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 974.078666] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0ac79184-8817-45f8-989f-4740aa73d2d1 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.100218] env[61923]: DEBUG oslo_vmware.api [None req-cd7c0dc9-d2a5-476c-afbb-7be6ac77c41c tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Waiting for the task: (returnval){ [ 974.100218] env[61923]: value = "task-1377898" [ 974.100218] env[61923]: _type = "Task" [ 974.100218] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 974.108884] env[61923]: DEBUG oslo_vmware.api [None req-cd7c0dc9-d2a5-476c-afbb-7be6ac77c41c tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Task: {'id': task-1377898, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.167062] env[61923]: DEBUG oslo_vmware.api [None req-eea05bb2-c691-4804-8c8d-019f8feaa521 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Task: {'id': task-1377895, 'name': PowerOnVM_Task, 'duration_secs': 0.582473} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 974.167368] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-eea05bb2-c691-4804-8c8d-019f8feaa521 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] [instance: a511392e-9ab6-42fb-b07c-f90e9786dcc1] Powered on the VM {{(pid=61923) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 974.273351] env[61923]: DEBUG nova.compute.manager [None req-eea05bb2-c691-4804-8c8d-019f8feaa521 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] [instance: a511392e-9ab6-42fb-b07c-f90e9786dcc1] Checking state {{(pid=61923) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 974.274494] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a8a973c-0927-4e09-80f6-d79927221ca3 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.533316] env[61923]: DEBUG oslo_vmware.api [None req-9798b6a1-c446-4119-9842-2560d710a905 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Task: {'id': task-1377896, 'name': PowerOnVM_Task, 'duration_secs': 0.518303} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 974.533608] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-9798b6a1-c446-4119-9842-2560d710a905 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: 638bbbbb-827e-4a73-ac24-a3e1b2cd72eb] Powered on the VM {{(pid=61923) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 974.533735] env[61923]: INFO nova.compute.manager [None req-9798b6a1-c446-4119-9842-2560d710a905 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: 638bbbbb-827e-4a73-ac24-a3e1b2cd72eb] Took 9.96 seconds to spawn the instance on the hypervisor. [ 974.533959] env[61923]: DEBUG nova.compute.manager [None req-9798b6a1-c446-4119-9842-2560d710a905 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: 638bbbbb-827e-4a73-ac24-a3e1b2cd72eb] Checking state {{(pid=61923) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 974.534752] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e540351b-5e7b-44c4-86e1-8dd1192ad6e6 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.610349] env[61923]: DEBUG oslo_vmware.api [None req-cd7c0dc9-d2a5-476c-afbb-7be6ac77c41c tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Task: {'id': task-1377898, 'name': ReconfigVM_Task, 'duration_secs': 0.299885} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 974.610349] env[61923]: DEBUG nova.virt.vmwareapi.volumeops [None req-cd7c0dc9-d2a5-476c-afbb-7be6ac77c41c tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: 3fcedef5-9040-4b45-9e27-f0700112d528] Reconfigured VM instance instance-0000005d to attach disk [datastore1] 3fcedef5-9040-4b45-9e27-f0700112d528/3fcedef5-9040-4b45-9e27-f0700112d528.vmdk or device None with type sparse {{(pid=61923) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 974.611260] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-cb8ef325-6f63-4251-89a1-f2eec6b09123 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.618517] env[61923]: DEBUG oslo_vmware.api [None req-cd7c0dc9-d2a5-476c-afbb-7be6ac77c41c tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Waiting for the task: (returnval){ [ 974.618517] env[61923]: value = "task-1377899" [ 974.618517] env[61923]: _type = "Task" [ 974.618517] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 974.627178] env[61923]: DEBUG oslo_vmware.api [None req-cd7c0dc9-d2a5-476c-afbb-7be6ac77c41c tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Task: {'id': task-1377899, 'name': Rename_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.792306] env[61923]: DEBUG nova.objects.instance [None req-e924267f-c245-4db8-af4c-6f10168a1fcb tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Lazy-loading 'flavor' on Instance uuid b779d183-89ae-4e4d-ae99-e514e145ed43 {{(pid=61923) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 974.795122] env[61923]: DEBUG oslo_concurrency.lockutils [None req-eea05bb2-c691-4804-8c8d-019f8feaa521 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Lock "a511392e-9ab6-42fb-b07c-f90e9786dcc1" "released" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: held 24.574s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 975.052612] env[61923]: INFO nova.compute.manager [None req-9798b6a1-c446-4119-9842-2560d710a905 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: 638bbbbb-827e-4a73-ac24-a3e1b2cd72eb] Took 16.27 seconds to build instance. [ 975.131905] env[61923]: DEBUG oslo_vmware.api [None req-cd7c0dc9-d2a5-476c-afbb-7be6ac77c41c tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Task: {'id': task-1377899, 'name': Rename_Task, 'duration_secs': 0.14521} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 975.132251] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-cd7c0dc9-d2a5-476c-afbb-7be6ac77c41c tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: 3fcedef5-9040-4b45-9e27-f0700112d528] Powering on the VM {{(pid=61923) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 975.132505] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-edce34ad-3002-48bb-8cab-42a4ded026df {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.145655] env[61923]: DEBUG oslo_vmware.api [None req-cd7c0dc9-d2a5-476c-afbb-7be6ac77c41c tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Waiting for the task: (returnval){ [ 975.145655] env[61923]: value = "task-1377900" [ 975.145655] env[61923]: _type = "Task" [ 975.145655] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 975.158057] env[61923]: DEBUG oslo_vmware.api [None req-cd7c0dc9-d2a5-476c-afbb-7be6ac77c41c tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Task: {'id': task-1377900, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 975.298291] env[61923]: DEBUG oslo_concurrency.lockutils [None req-e924267f-c245-4db8-af4c-6f10168a1fcb tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Acquiring lock "refresh_cache-b779d183-89ae-4e4d-ae99-e514e145ed43" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 975.298573] env[61923]: DEBUG oslo_concurrency.lockutils [None req-e924267f-c245-4db8-af4c-6f10168a1fcb tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Acquired lock "refresh_cache-b779d183-89ae-4e4d-ae99-e514e145ed43" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 975.298769] env[61923]: DEBUG nova.network.neutron [None req-e924267f-c245-4db8-af4c-6f10168a1fcb tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] [instance: b779d183-89ae-4e4d-ae99-e514e145ed43] Building network info cache for instance {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 975.298957] env[61923]: DEBUG nova.objects.instance [None req-e924267f-c245-4db8-af4c-6f10168a1fcb tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Lazy-loading 'info_cache' on Instance uuid b779d183-89ae-4e4d-ae99-e514e145ed43 {{(pid=61923) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 975.554299] env[61923]: DEBUG oslo_concurrency.lockutils [None req-9798b6a1-c446-4119-9842-2560d710a905 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Lock "638bbbbb-827e-4a73-ac24-a3e1b2cd72eb" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 17.774s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 975.656484] env[61923]: DEBUG oslo_vmware.api [None req-cd7c0dc9-d2a5-476c-afbb-7be6ac77c41c tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Task: {'id': task-1377900, 'name': PowerOnVM_Task, 'duration_secs': 0.449261} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 975.656754] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-cd7c0dc9-d2a5-476c-afbb-7be6ac77c41c tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: 3fcedef5-9040-4b45-9e27-f0700112d528] Powered on the VM {{(pid=61923) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 975.656955] env[61923]: INFO nova.compute.manager [None req-cd7c0dc9-d2a5-476c-afbb-7be6ac77c41c tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: 3fcedef5-9040-4b45-9e27-f0700112d528] Took 7.25 seconds to spawn the instance on the hypervisor. [ 975.657152] env[61923]: DEBUG nova.compute.manager [None req-cd7c0dc9-d2a5-476c-afbb-7be6ac77c41c tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: 3fcedef5-9040-4b45-9e27-f0700112d528] Checking state {{(pid=61923) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 975.657919] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd6ffd32-ddc2-4984-8793-ab9a47a44fe7 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.802842] env[61923]: DEBUG nova.objects.base [None req-e924267f-c245-4db8-af4c-6f10168a1fcb tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Object Instance lazy-loaded attributes: flavor,info_cache {{(pid=61923) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 975.892489] env[61923]: DEBUG oslo_concurrency.lockutils [None req-0be1ab9a-5f88-4a43-b081-39c4a84daac9 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Acquiring lock "638bbbbb-827e-4a73-ac24-a3e1b2cd72eb" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 975.892765] env[61923]: DEBUG oslo_concurrency.lockutils [None req-0be1ab9a-5f88-4a43-b081-39c4a84daac9 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Lock "638bbbbb-827e-4a73-ac24-a3e1b2cd72eb" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 975.892943] env[61923]: DEBUG nova.compute.manager [None req-0be1ab9a-5f88-4a43-b081-39c4a84daac9 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: 638bbbbb-827e-4a73-ac24-a3e1b2cd72eb] Checking state {{(pid=61923) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 975.894319] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-331ed5e2-d4e8-4367-865a-aaac926fabed {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.901816] env[61923]: DEBUG nova.compute.manager [None req-0be1ab9a-5f88-4a43-b081-39c4a84daac9 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: 638bbbbb-827e-4a73-ac24-a3e1b2cd72eb] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=61923) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 975.902385] env[61923]: DEBUG nova.objects.instance [None req-0be1ab9a-5f88-4a43-b081-39c4a84daac9 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Lazy-loading 'flavor' on Instance uuid 638bbbbb-827e-4a73-ac24-a3e1b2cd72eb {{(pid=61923) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 976.175387] env[61923]: INFO nova.compute.manager [None req-cd7c0dc9-d2a5-476c-afbb-7be6ac77c41c tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: 3fcedef5-9040-4b45-9e27-f0700112d528] Took 13.61 seconds to build instance. [ 976.377695] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12a63494-0d60-4d0e-b7af-19765d7e9ac5 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.385451] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-08ecbe02-36f6-40cd-93d7-c54939989001 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] [instance: a511392e-9ab6-42fb-b07c-f90e9786dcc1] Suspending the VM {{(pid=61923) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1162}} [ 976.385768] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.SuspendVM_Task with opID=oslo.vmware-0320c23f-560b-4d35-8e17-0e21c4efe849 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.393342] env[61923]: DEBUG oslo_vmware.api [None req-08ecbe02-36f6-40cd-93d7-c54939989001 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Waiting for the task: (returnval){ [ 976.393342] env[61923]: value = "task-1377901" [ 976.393342] env[61923]: _type = "Task" [ 976.393342] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 976.402547] env[61923]: DEBUG oslo_vmware.api [None req-08ecbe02-36f6-40cd-93d7-c54939989001 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Task: {'id': task-1377901, 'name': SuspendVM_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 976.407487] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-0be1ab9a-5f88-4a43-b081-39c4a84daac9 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: 638bbbbb-827e-4a73-ac24-a3e1b2cd72eb] Powering off the VM {{(pid=61923) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 976.407834] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-3a809147-2636-4d51-8fd6-1e59f7cec8c2 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.416227] env[61923]: DEBUG oslo_vmware.api [None req-0be1ab9a-5f88-4a43-b081-39c4a84daac9 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Waiting for the task: (returnval){ [ 976.416227] env[61923]: value = "task-1377902" [ 976.416227] env[61923]: _type = "Task" [ 976.416227] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 976.426099] env[61923]: DEBUG oslo_vmware.api [None req-0be1ab9a-5f88-4a43-b081-39c4a84daac9 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Task: {'id': task-1377902, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 976.542309] env[61923]: DEBUG nova.network.neutron [None req-e924267f-c245-4db8-af4c-6f10168a1fcb tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] [instance: b779d183-89ae-4e4d-ae99-e514e145ed43] Updating instance_info_cache with network_info: [{"id": "d49a3010-c84a-4bef-beed-9ca4aa7d5dc2", "address": "fa:16:3e:48:e6:a2", "network": {"id": "d2905082-7ca4-4f92-a142-bb85211bb7c4", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-885218622-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.251", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a13d827933fa4597984afdb91dbbdd39", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "03ac2c9c-6ad2-4a85-bfab-c7e336df859a", "external-id": "nsx-vlan-transportzone-379", "segmentation_id": 379, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd49a3010-c8", "ovs_interfaceid": "d49a3010-c84a-4bef-beed-9ca4aa7d5dc2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 976.678109] env[61923]: DEBUG oslo_concurrency.lockutils [None req-cd7c0dc9-d2a5-476c-afbb-7be6ac77c41c tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Lock "3fcedef5-9040-4b45-9e27-f0700112d528" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.121s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 976.906697] env[61923]: DEBUG oslo_vmware.api [None req-08ecbe02-36f6-40cd-93d7-c54939989001 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Task: {'id': task-1377901, 'name': SuspendVM_Task} progress is 54%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 976.927134] env[61923]: DEBUG oslo_vmware.api [None req-0be1ab9a-5f88-4a43-b081-39c4a84daac9 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Task: {'id': task-1377902, 'name': PowerOffVM_Task, 'duration_secs': 0.250792} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 976.927566] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-0be1ab9a-5f88-4a43-b081-39c4a84daac9 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: 638bbbbb-827e-4a73-ac24-a3e1b2cd72eb] Powered off the VM {{(pid=61923) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 976.927802] env[61923]: DEBUG nova.compute.manager [None req-0be1ab9a-5f88-4a43-b081-39c4a84daac9 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: 638bbbbb-827e-4a73-ac24-a3e1b2cd72eb] Checking state {{(pid=61923) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 976.928605] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99721f12-9375-4f92-ac6b-a7c8b9c413a9 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.045072] env[61923]: DEBUG oslo_concurrency.lockutils [None req-e924267f-c245-4db8-af4c-6f10168a1fcb tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Releasing lock "refresh_cache-b779d183-89ae-4e4d-ae99-e514e145ed43" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 977.406051] env[61923]: DEBUG oslo_vmware.api [None req-08ecbe02-36f6-40cd-93d7-c54939989001 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Task: {'id': task-1377901, 'name': SuspendVM_Task, 'duration_secs': 0.65138} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 977.406051] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-08ecbe02-36f6-40cd-93d7-c54939989001 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] [instance: a511392e-9ab6-42fb-b07c-f90e9786dcc1] Suspended the VM {{(pid=61923) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1166}} [ 977.406051] env[61923]: DEBUG nova.compute.manager [None req-08ecbe02-36f6-40cd-93d7-c54939989001 tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] [instance: a511392e-9ab6-42fb-b07c-f90e9786dcc1] Checking state {{(pid=61923) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 977.406472] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dce2077e-e171-4f64-a78e-c0d49095665f {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.441726] env[61923]: DEBUG oslo_concurrency.lockutils [None req-0be1ab9a-5f88-4a43-b081-39c4a84daac9 tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Lock "638bbbbb-827e-4a73-ac24-a3e1b2cd72eb" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.548s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 977.549172] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-e924267f-c245-4db8-af4c-6f10168a1fcb tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] [instance: b779d183-89ae-4e4d-ae99-e514e145ed43] Powering on the VM {{(pid=61923) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 977.549510] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-29f34548-b655-4cd9-ad08-1535064db448 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.556919] env[61923]: DEBUG oslo_vmware.api [None req-e924267f-c245-4db8-af4c-6f10168a1fcb tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Waiting for the task: (returnval){ [ 977.556919] env[61923]: value = "task-1377903" [ 977.556919] env[61923]: _type = "Task" [ 977.556919] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 977.565603] env[61923]: DEBUG oslo_vmware.api [None req-e924267f-c245-4db8-af4c-6f10168a1fcb tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Task: {'id': task-1377903, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 977.733966] env[61923]: DEBUG oslo_concurrency.lockutils [None req-fd8faa6e-b3a9-4ab5-afac-c7b4d3d4b57d tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Acquiring lock "3fcedef5-9040-4b45-9e27-f0700112d528" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 977.734246] env[61923]: DEBUG oslo_concurrency.lockutils [None req-fd8faa6e-b3a9-4ab5-afac-c7b4d3d4b57d tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Lock "3fcedef5-9040-4b45-9e27-f0700112d528" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 978.070523] env[61923]: DEBUG oslo_vmware.api [None req-e924267f-c245-4db8-af4c-6f10168a1fcb tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Task: {'id': task-1377903, 'name': PowerOnVM_Task, 'duration_secs': 0.376178} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 978.070850] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-e924267f-c245-4db8-af4c-6f10168a1fcb tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] [instance: b779d183-89ae-4e4d-ae99-e514e145ed43] Powered on the VM {{(pid=61923) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 978.070981] env[61923]: DEBUG nova.compute.manager [None req-e924267f-c245-4db8-af4c-6f10168a1fcb tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] [instance: b779d183-89ae-4e4d-ae99-e514e145ed43] Checking state {{(pid=61923) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 978.071817] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8814bc9a-7061-4e4e-8749-fb2a0244ff1f {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.238135] env[61923]: DEBUG nova.compute.utils [None req-fd8faa6e-b3a9-4ab5-afac-c7b4d3d4b57d tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Using /dev/sd instead of None {{(pid=61923) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 978.623875] env[61923]: DEBUG oslo_concurrency.lockutils [None req-d6738f12-6cd9-4792-98bf-452343c4873f tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Acquiring lock "638bbbbb-827e-4a73-ac24-a3e1b2cd72eb" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 978.624651] env[61923]: DEBUG oslo_concurrency.lockutils [None req-d6738f12-6cd9-4792-98bf-452343c4873f tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Lock "638bbbbb-827e-4a73-ac24-a3e1b2cd72eb" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 978.624882] env[61923]: DEBUG oslo_concurrency.lockutils [None req-d6738f12-6cd9-4792-98bf-452343c4873f tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Acquiring lock "638bbbbb-827e-4a73-ac24-a3e1b2cd72eb-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 978.625090] env[61923]: DEBUG oslo_concurrency.lockutils [None req-d6738f12-6cd9-4792-98bf-452343c4873f tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Lock "638bbbbb-827e-4a73-ac24-a3e1b2cd72eb-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 978.625266] env[61923]: DEBUG oslo_concurrency.lockutils [None req-d6738f12-6cd9-4792-98bf-452343c4873f tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Lock "638bbbbb-827e-4a73-ac24-a3e1b2cd72eb-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 978.631094] env[61923]: INFO nova.compute.manager [None req-d6738f12-6cd9-4792-98bf-452343c4873f tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: 638bbbbb-827e-4a73-ac24-a3e1b2cd72eb] Terminating instance [ 978.632748] env[61923]: DEBUG nova.compute.manager [None req-d6738f12-6cd9-4792-98bf-452343c4873f tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: 638bbbbb-827e-4a73-ac24-a3e1b2cd72eb] Start destroying the instance on the hypervisor. {{(pid=61923) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 978.632947] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-d6738f12-6cd9-4792-98bf-452343c4873f tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: 638bbbbb-827e-4a73-ac24-a3e1b2cd72eb] Destroying instance {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 978.633785] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b5c4243-3711-4a61-a384-9c0af07681ac {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.642143] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-d6738f12-6cd9-4792-98bf-452343c4873f tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: 638bbbbb-827e-4a73-ac24-a3e1b2cd72eb] Unregistering the VM {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 978.642389] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-92a8a1cd-c7d5-43fc-b87b-8204f2efa30a {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.716416] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-d6738f12-6cd9-4792-98bf-452343c4873f tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: 638bbbbb-827e-4a73-ac24-a3e1b2cd72eb] Unregistered the VM {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 978.716612] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-d6738f12-6cd9-4792-98bf-452343c4873f tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: 638bbbbb-827e-4a73-ac24-a3e1b2cd72eb] Deleting contents of the VM from datastore datastore1 {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 978.716778] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-d6738f12-6cd9-4792-98bf-452343c4873f tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Deleting the datastore file [datastore1] 638bbbbb-827e-4a73-ac24-a3e1b2cd72eb {{(pid=61923) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 978.717060] env[61923]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-2a268cc4-c330-49fb-84e0-c5662dcf8f95 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.725317] env[61923]: DEBUG oslo_vmware.api [None req-d6738f12-6cd9-4792-98bf-452343c4873f tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Waiting for the task: (returnval){ [ 978.725317] env[61923]: value = "task-1377905" [ 978.725317] env[61923]: _type = "Task" [ 978.725317] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 978.734206] env[61923]: DEBUG oslo_vmware.api [None req-d6738f12-6cd9-4792-98bf-452343c4873f tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Task: {'id': task-1377905, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 978.741173] env[61923]: DEBUG oslo_concurrency.lockutils [None req-fd8faa6e-b3a9-4ab5-afac-c7b4d3d4b57d tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Lock "3fcedef5-9040-4b45-9e27-f0700112d528" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.007s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 978.907606] env[61923]: INFO nova.compute.manager [None req-604e4018-6e74-42a9-8cba-fd4a4ec7b92c tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] [instance: a511392e-9ab6-42fb-b07c-f90e9786dcc1] Resuming [ 978.908610] env[61923]: DEBUG nova.objects.instance [None req-604e4018-6e74-42a9-8cba-fd4a4ec7b92c tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Lazy-loading 'flavor' on Instance uuid a511392e-9ab6-42fb-b07c-f90e9786dcc1 {{(pid=61923) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 979.238061] env[61923]: DEBUG oslo_vmware.api [None req-d6738f12-6cd9-4792-98bf-452343c4873f tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Task: {'id': task-1377905, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.161281} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 979.238061] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-d6738f12-6cd9-4792-98bf-452343c4873f tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Deleted the datastore file {{(pid=61923) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 979.238621] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-d6738f12-6cd9-4792-98bf-452343c4873f tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: 638bbbbb-827e-4a73-ac24-a3e1b2cd72eb] Deleted contents of the VM from datastore datastore1 {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 979.238621] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-d6738f12-6cd9-4792-98bf-452343c4873f tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: 638bbbbb-827e-4a73-ac24-a3e1b2cd72eb] Instance destroyed {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 979.238621] env[61923]: INFO nova.compute.manager [None req-d6738f12-6cd9-4792-98bf-452343c4873f tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] [instance: 638bbbbb-827e-4a73-ac24-a3e1b2cd72eb] Took 0.61 seconds to destroy the instance on the hypervisor. [ 979.238786] env[61923]: DEBUG oslo.service.loopingcall [None req-d6738f12-6cd9-4792-98bf-452343c4873f tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61923) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 979.238940] env[61923]: DEBUG nova.compute.manager [-] [instance: 638bbbbb-827e-4a73-ac24-a3e1b2cd72eb] Deallocating network for instance {{(pid=61923) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 979.239032] env[61923]: DEBUG nova.network.neutron [-] [instance: 638bbbbb-827e-4a73-ac24-a3e1b2cd72eb] deallocate_for_instance() {{(pid=61923) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 979.509153] env[61923]: DEBUG nova.compute.manager [req-832485cb-2134-4c02-b92c-0182ecde3a8b req-9b732e1b-f0b3-4629-a073-7d69d50af46f service nova] [instance: 638bbbbb-827e-4a73-ac24-a3e1b2cd72eb] Received event network-vif-deleted-3e6f05d1-ae7a-4aff-9243-20d5dd0e9b4f {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 979.509153] env[61923]: INFO nova.compute.manager [req-832485cb-2134-4c02-b92c-0182ecde3a8b req-9b732e1b-f0b3-4629-a073-7d69d50af46f service nova] [instance: 638bbbbb-827e-4a73-ac24-a3e1b2cd72eb] Neutron deleted interface 3e6f05d1-ae7a-4aff-9243-20d5dd0e9b4f; detaching it from the instance and deleting it from the info cache [ 979.509343] env[61923]: DEBUG nova.network.neutron [req-832485cb-2134-4c02-b92c-0182ecde3a8b req-9b732e1b-f0b3-4629-a073-7d69d50af46f service nova] [instance: 638bbbbb-827e-4a73-ac24-a3e1b2cd72eb] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 979.808681] env[61923]: DEBUG oslo_concurrency.lockutils [None req-fd8faa6e-b3a9-4ab5-afac-c7b4d3d4b57d tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Acquiring lock "3fcedef5-9040-4b45-9e27-f0700112d528" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 979.809021] env[61923]: DEBUG oslo_concurrency.lockutils [None req-fd8faa6e-b3a9-4ab5-afac-c7b4d3d4b57d tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Lock "3fcedef5-9040-4b45-9e27-f0700112d528" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.001s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 979.809296] env[61923]: INFO nova.compute.manager [None req-fd8faa6e-b3a9-4ab5-afac-c7b4d3d4b57d tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: 3fcedef5-9040-4b45-9e27-f0700112d528] Attaching volume 85aa34e0-345d-403c-ab7a-58822ce81f4e to /dev/sdb [ 979.844616] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13bb9dfc-cd22-44fd-b88c-d296b56f3939 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.852325] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da4363b7-19dd-467b-af6a-60f516734387 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.866908] env[61923]: DEBUG nova.virt.block_device [None req-fd8faa6e-b3a9-4ab5-afac-c7b4d3d4b57d tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: 3fcedef5-9040-4b45-9e27-f0700112d528] Updating existing volume attachment record: 96d50ed1-85e7-45b2-85fa-f08fdf559d10 {{(pid=61923) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 979.917159] env[61923]: DEBUG oslo_concurrency.lockutils [None req-604e4018-6e74-42a9-8cba-fd4a4ec7b92c tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Acquiring lock "refresh_cache-a511392e-9ab6-42fb-b07c-f90e9786dcc1" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 979.917371] env[61923]: DEBUG oslo_concurrency.lockutils [None req-604e4018-6e74-42a9-8cba-fd4a4ec7b92c tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Acquired lock "refresh_cache-a511392e-9ab6-42fb-b07c-f90e9786dcc1" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 979.917549] env[61923]: DEBUG nova.network.neutron [None req-604e4018-6e74-42a9-8cba-fd4a4ec7b92c tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] [instance: a511392e-9ab6-42fb-b07c-f90e9786dcc1] Building network info cache for instance {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 979.988675] env[61923]: DEBUG nova.network.neutron [-] [instance: 638bbbbb-827e-4a73-ac24-a3e1b2cd72eb] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 980.011629] env[61923]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d1abda11-05d7-49e2-817a-2db420b266d8 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.021081] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e40b6941-ff26-4617-8443-e64d30508c6a {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.048297] env[61923]: DEBUG nova.compute.manager [req-832485cb-2134-4c02-b92c-0182ecde3a8b req-9b732e1b-f0b3-4629-a073-7d69d50af46f service nova] [instance: 638bbbbb-827e-4a73-ac24-a3e1b2cd72eb] Detach interface failed, port_id=3e6f05d1-ae7a-4aff-9243-20d5dd0e9b4f, reason: Instance 638bbbbb-827e-4a73-ac24-a3e1b2cd72eb could not be found. {{(pid=61923) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 980.491894] env[61923]: INFO nova.compute.manager [-] [instance: 638bbbbb-827e-4a73-ac24-a3e1b2cd72eb] Took 1.25 seconds to deallocate network for instance. [ 980.635978] env[61923]: DEBUG nova.network.neutron [None req-604e4018-6e74-42a9-8cba-fd4a4ec7b92c tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] [instance: a511392e-9ab6-42fb-b07c-f90e9786dcc1] Updating instance_info_cache with network_info: [{"id": "0325cd38-bd49-4132-85e3-c66fc9efe49f", "address": "fa:16:3e:7f:89:31", "network": {"id": "07fc148e-c3ba-4e53-bec2-36500ec38add", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1374533437-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "0d24f4a47d1f4b79a7d999a4dacb99ba", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "b2907cce-d529-4809-af05-d29397bed211", "external-id": "nsx-vlan-transportzone-427", "segmentation_id": 427, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0325cd38-bd", "ovs_interfaceid": "0325cd38-bd49-4132-85e3-c66fc9efe49f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 980.998963] env[61923]: DEBUG oslo_concurrency.lockutils [None req-d6738f12-6cd9-4792-98bf-452343c4873f tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 980.999385] env[61923]: DEBUG oslo_concurrency.lockutils [None req-d6738f12-6cd9-4792-98bf-452343c4873f tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 980.999705] env[61923]: DEBUG nova.objects.instance [None req-d6738f12-6cd9-4792-98bf-452343c4873f tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Lazy-loading 'resources' on Instance uuid 638bbbbb-827e-4a73-ac24-a3e1b2cd72eb {{(pid=61923) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 981.138292] env[61923]: DEBUG oslo_concurrency.lockutils [None req-604e4018-6e74-42a9-8cba-fd4a4ec7b92c tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Releasing lock "refresh_cache-a511392e-9ab6-42fb-b07c-f90e9786dcc1" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 981.139280] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71162250-804b-416a-abc6-b7913bafab97 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.146627] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-604e4018-6e74-42a9-8cba-fd4a4ec7b92c tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] [instance: a511392e-9ab6-42fb-b07c-f90e9786dcc1] Resuming the VM {{(pid=61923) resume /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1183}} [ 981.146859] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-fd71f7bc-7661-4f9e-b9a3-ae5f80e1b071 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.153178] env[61923]: DEBUG oslo_vmware.api [None req-604e4018-6e74-42a9-8cba-fd4a4ec7b92c tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Waiting for the task: (returnval){ [ 981.153178] env[61923]: value = "task-1377909" [ 981.153178] env[61923]: _type = "Task" [ 981.153178] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 981.160483] env[61923]: DEBUG oslo_vmware.api [None req-604e4018-6e74-42a9-8cba-fd4a4ec7b92c tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Task: {'id': task-1377909, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 981.591453] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a5305c8-3dc1-4c86-8bd2-48ee6610013c {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.600631] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d42bc18b-98fd-474e-ada8-e6130470c047 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.633228] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90413d50-25e2-488c-a00e-f850991dbb63 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.641811] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c97a92a-b043-42ef-9c03-a82a9c2c04f5 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.657618] env[61923]: DEBUG nova.compute.provider_tree [None req-d6738f12-6cd9-4792-98bf-452343c4873f tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 981.670323] env[61923]: DEBUG oslo_vmware.api [None req-604e4018-6e74-42a9-8cba-fd4a4ec7b92c tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Task: {'id': task-1377909, 'name': PowerOnVM_Task} progress is 100%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 982.164265] env[61923]: DEBUG nova.scheduler.client.report [None req-d6738f12-6cd9-4792-98bf-452343c4873f tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 982.170521] env[61923]: DEBUG oslo_vmware.api [None req-604e4018-6e74-42a9-8cba-fd4a4ec7b92c tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Task: {'id': task-1377909, 'name': PowerOnVM_Task, 'duration_secs': 0.545177} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 982.171019] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-604e4018-6e74-42a9-8cba-fd4a4ec7b92c tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] [instance: a511392e-9ab6-42fb-b07c-f90e9786dcc1] Resumed the VM {{(pid=61923) resume /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1188}} [ 982.171242] env[61923]: DEBUG nova.compute.manager [None req-604e4018-6e74-42a9-8cba-fd4a4ec7b92c tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] [instance: a511392e-9ab6-42fb-b07c-f90e9786dcc1] Checking state {{(pid=61923) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 982.172067] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79567359-7371-4357-93b2-35ae99a18134 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.672523] env[61923]: DEBUG oslo_concurrency.lockutils [None req-d6738f12-6cd9-4792-98bf-452343c4873f tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.673s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 982.696338] env[61923]: INFO nova.scheduler.client.report [None req-d6738f12-6cd9-4792-98bf-452343c4873f tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Deleted allocations for instance 638bbbbb-827e-4a73-ac24-a3e1b2cd72eb [ 983.207044] env[61923]: DEBUG oslo_concurrency.lockutils [None req-d6738f12-6cd9-4792-98bf-452343c4873f tempest-ServersTestJSON-1823530610 tempest-ServersTestJSON-1823530610-project-member] Lock "638bbbbb-827e-4a73-ac24-a3e1b2cd72eb" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 4.582s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 984.413119] env[61923]: DEBUG nova.virt.vmwareapi.volumeops [None req-fd8faa6e-b3a9-4ab5-afac-c7b4d3d4b57d tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: 3fcedef5-9040-4b45-9e27-f0700112d528] Volume attach. Driver type: vmdk {{(pid=61923) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 984.413385] env[61923]: DEBUG nova.virt.vmwareapi.volumeops [None req-fd8faa6e-b3a9-4ab5-afac-c7b4d3d4b57d tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: 3fcedef5-9040-4b45-9e27-f0700112d528] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-292739', 'volume_id': '85aa34e0-345d-403c-ab7a-58822ce81f4e', 'name': 'volume-85aa34e0-345d-403c-ab7a-58822ce81f4e', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '3fcedef5-9040-4b45-9e27-f0700112d528', 'attached_at': '', 'detached_at': '', 'volume_id': '85aa34e0-345d-403c-ab7a-58822ce81f4e', 'serial': '85aa34e0-345d-403c-ab7a-58822ce81f4e'} {{(pid=61923) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 984.417019] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a00da27-99a9-401d-9956-468051bb1add {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.431576] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a39b4044-1285-4411-8de1-7dd5e253bcca {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.458996] env[61923]: DEBUG nova.virt.vmwareapi.volumeops [None req-fd8faa6e-b3a9-4ab5-afac-c7b4d3d4b57d tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: 3fcedef5-9040-4b45-9e27-f0700112d528] Reconfiguring VM instance instance-0000005d to attach disk [datastore2] volume-85aa34e0-345d-403c-ab7a-58822ce81f4e/volume-85aa34e0-345d-403c-ab7a-58822ce81f4e.vmdk or device None with type thin {{(pid=61923) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 984.459803] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-566e406d-63bb-4d4b-af47-8681c15fd065 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.478500] env[61923]: DEBUG oslo_vmware.api [None req-fd8faa6e-b3a9-4ab5-afac-c7b4d3d4b57d tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Waiting for the task: (returnval){ [ 984.478500] env[61923]: value = "task-1377911" [ 984.478500] env[61923]: _type = "Task" [ 984.478500] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 984.487335] env[61923]: DEBUG oslo_vmware.api [None req-fd8faa6e-b3a9-4ab5-afac-c7b4d3d4b57d tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Task: {'id': task-1377911, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 984.989019] env[61923]: DEBUG oslo_vmware.api [None req-fd8faa6e-b3a9-4ab5-afac-c7b4d3d4b57d tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Task: {'id': task-1377911, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.491249] env[61923]: DEBUG oslo_vmware.api [None req-fd8faa6e-b3a9-4ab5-afac-c7b4d3d4b57d tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Task: {'id': task-1377911, 'name': ReconfigVM_Task, 'duration_secs': 0.623367} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 985.491545] env[61923]: DEBUG nova.virt.vmwareapi.volumeops [None req-fd8faa6e-b3a9-4ab5-afac-c7b4d3d4b57d tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: 3fcedef5-9040-4b45-9e27-f0700112d528] Reconfigured VM instance instance-0000005d to attach disk [datastore2] volume-85aa34e0-345d-403c-ab7a-58822ce81f4e/volume-85aa34e0-345d-403c-ab7a-58822ce81f4e.vmdk or device None with type thin {{(pid=61923) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 985.496963] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8090754d-64f4-4678-9ac4-404d6c538429 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.514485] env[61923]: DEBUG oslo_vmware.api [None req-fd8faa6e-b3a9-4ab5-afac-c7b4d3d4b57d tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Waiting for the task: (returnval){ [ 985.514485] env[61923]: value = "task-1377912" [ 985.514485] env[61923]: _type = "Task" [ 985.514485] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 985.523757] env[61923]: DEBUG oslo_vmware.api [None req-fd8faa6e-b3a9-4ab5-afac-c7b4d3d4b57d tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Task: {'id': task-1377912, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.024395] env[61923]: DEBUG oslo_vmware.api [None req-fd8faa6e-b3a9-4ab5-afac-c7b4d3d4b57d tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Task: {'id': task-1377912, 'name': ReconfigVM_Task, 'duration_secs': 0.143473} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 986.024668] env[61923]: DEBUG nova.virt.vmwareapi.volumeops [None req-fd8faa6e-b3a9-4ab5-afac-c7b4d3d4b57d tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: 3fcedef5-9040-4b45-9e27-f0700112d528] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-292739', 'volume_id': '85aa34e0-345d-403c-ab7a-58822ce81f4e', 'name': 'volume-85aa34e0-345d-403c-ab7a-58822ce81f4e', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '3fcedef5-9040-4b45-9e27-f0700112d528', 'attached_at': '', 'detached_at': '', 'volume_id': '85aa34e0-345d-403c-ab7a-58822ce81f4e', 'serial': '85aa34e0-345d-403c-ab7a-58822ce81f4e'} {{(pid=61923) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 987.063090] env[61923]: DEBUG nova.objects.instance [None req-fd8faa6e-b3a9-4ab5-afac-c7b4d3d4b57d tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Lazy-loading 'flavor' on Instance uuid 3fcedef5-9040-4b45-9e27-f0700112d528 {{(pid=61923) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 987.372445] env[61923]: DEBUG oslo_concurrency.lockutils [None req-5f0ffb82-8d31-4829-b8c1-ae776078cdda tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Acquiring lock "a511392e-9ab6-42fb-b07c-f90e9786dcc1" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 987.372713] env[61923]: DEBUG oslo_concurrency.lockutils [None req-5f0ffb82-8d31-4829-b8c1-ae776078cdda tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Lock "a511392e-9ab6-42fb-b07c-f90e9786dcc1" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 987.372934] env[61923]: DEBUG oslo_concurrency.lockutils [None req-5f0ffb82-8d31-4829-b8c1-ae776078cdda tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Acquiring lock "a511392e-9ab6-42fb-b07c-f90e9786dcc1-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 987.373142] env[61923]: DEBUG oslo_concurrency.lockutils [None req-5f0ffb82-8d31-4829-b8c1-ae776078cdda tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Lock "a511392e-9ab6-42fb-b07c-f90e9786dcc1-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 987.373320] env[61923]: DEBUG oslo_concurrency.lockutils [None req-5f0ffb82-8d31-4829-b8c1-ae776078cdda tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Lock "a511392e-9ab6-42fb-b07c-f90e9786dcc1-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 987.375590] env[61923]: INFO nova.compute.manager [None req-5f0ffb82-8d31-4829-b8c1-ae776078cdda tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] [instance: a511392e-9ab6-42fb-b07c-f90e9786dcc1] Terminating instance [ 987.377881] env[61923]: DEBUG nova.compute.manager [None req-5f0ffb82-8d31-4829-b8c1-ae776078cdda tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] [instance: a511392e-9ab6-42fb-b07c-f90e9786dcc1] Start destroying the instance on the hypervisor. {{(pid=61923) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 987.378134] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-5f0ffb82-8d31-4829-b8c1-ae776078cdda tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] [instance: a511392e-9ab6-42fb-b07c-f90e9786dcc1] Destroying instance {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 987.378970] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54dfb5ec-3f65-409c-a552-49dd16cdcd83 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.387348] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-5f0ffb82-8d31-4829-b8c1-ae776078cdda tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] [instance: a511392e-9ab6-42fb-b07c-f90e9786dcc1] Powering off the VM {{(pid=61923) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 987.387616] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c8bd9eb8-ff6d-490a-9f10-23992415baf0 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.394539] env[61923]: DEBUG oslo_vmware.api [None req-5f0ffb82-8d31-4829-b8c1-ae776078cdda tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Waiting for the task: (returnval){ [ 987.394539] env[61923]: value = "task-1377913" [ 987.394539] env[61923]: _type = "Task" [ 987.394539] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 987.402778] env[61923]: DEBUG oslo_vmware.api [None req-5f0ffb82-8d31-4829-b8c1-ae776078cdda tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Task: {'id': task-1377913, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.568084] env[61923]: DEBUG oslo_concurrency.lockutils [None req-fd8faa6e-b3a9-4ab5-afac-c7b4d3d4b57d tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Lock "3fcedef5-9040-4b45-9e27-f0700112d528" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.759s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 987.905588] env[61923]: DEBUG oslo_vmware.api [None req-5f0ffb82-8d31-4829-b8c1-ae776078cdda tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Task: {'id': task-1377913, 'name': PowerOffVM_Task, 'duration_secs': 0.196203} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 987.905865] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-5f0ffb82-8d31-4829-b8c1-ae776078cdda tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] [instance: a511392e-9ab6-42fb-b07c-f90e9786dcc1] Powered off the VM {{(pid=61923) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 987.906049] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-5f0ffb82-8d31-4829-b8c1-ae776078cdda tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] [instance: a511392e-9ab6-42fb-b07c-f90e9786dcc1] Unregistering the VM {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 987.906305] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-ed94a599-34ca-4951-aaaf-507f3127f580 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.997322] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-5f0ffb82-8d31-4829-b8c1-ae776078cdda tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] [instance: a511392e-9ab6-42fb-b07c-f90e9786dcc1] Unregistered the VM {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 987.997642] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-5f0ffb82-8d31-4829-b8c1-ae776078cdda tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] [instance: a511392e-9ab6-42fb-b07c-f90e9786dcc1] Deleting contents of the VM from datastore datastore1 {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 987.997855] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-5f0ffb82-8d31-4829-b8c1-ae776078cdda tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Deleting the datastore file [datastore1] a511392e-9ab6-42fb-b07c-f90e9786dcc1 {{(pid=61923) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 987.998157] env[61923]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ae1ed3bf-68a1-45ec-8965-3cd0075e0a5a {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.005756] env[61923]: DEBUG oslo_vmware.api [None req-5f0ffb82-8d31-4829-b8c1-ae776078cdda tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Waiting for the task: (returnval){ [ 988.005756] env[61923]: value = "task-1377915" [ 988.005756] env[61923]: _type = "Task" [ 988.005756] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 988.014055] env[61923]: DEBUG oslo_vmware.api [None req-5f0ffb82-8d31-4829-b8c1-ae776078cdda tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Task: {'id': task-1377915, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 988.186641] env[61923]: DEBUG oslo_concurrency.lockutils [None req-a2366500-67be-46eb-90d6-0efe293c8fc5 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Acquiring lock "3fcedef5-9040-4b45-9e27-f0700112d528" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 988.187036] env[61923]: DEBUG oslo_concurrency.lockutils [None req-a2366500-67be-46eb-90d6-0efe293c8fc5 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Lock "3fcedef5-9040-4b45-9e27-f0700112d528" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 988.187215] env[61923]: DEBUG oslo_concurrency.lockutils [None req-a2366500-67be-46eb-90d6-0efe293c8fc5 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Acquiring lock "3fcedef5-9040-4b45-9e27-f0700112d528-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 988.187451] env[61923]: DEBUG oslo_concurrency.lockutils [None req-a2366500-67be-46eb-90d6-0efe293c8fc5 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Lock "3fcedef5-9040-4b45-9e27-f0700112d528-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 988.187742] env[61923]: DEBUG oslo_concurrency.lockutils [None req-a2366500-67be-46eb-90d6-0efe293c8fc5 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Lock "3fcedef5-9040-4b45-9e27-f0700112d528-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 988.190252] env[61923]: INFO nova.compute.manager [None req-a2366500-67be-46eb-90d6-0efe293c8fc5 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: 3fcedef5-9040-4b45-9e27-f0700112d528] Terminating instance [ 988.192575] env[61923]: DEBUG nova.compute.manager [None req-a2366500-67be-46eb-90d6-0efe293c8fc5 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: 3fcedef5-9040-4b45-9e27-f0700112d528] Start destroying the instance on the hypervisor. {{(pid=61923) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 988.192827] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-a2366500-67be-46eb-90d6-0efe293c8fc5 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: 3fcedef5-9040-4b45-9e27-f0700112d528] Powering off the VM {{(pid=61923) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 988.193132] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1660d41a-54f3-478c-934e-5b53a886e462 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.202631] env[61923]: DEBUG oslo_vmware.api [None req-a2366500-67be-46eb-90d6-0efe293c8fc5 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Waiting for the task: (returnval){ [ 988.202631] env[61923]: value = "task-1377916" [ 988.202631] env[61923]: _type = "Task" [ 988.202631] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 988.212320] env[61923]: DEBUG oslo_vmware.api [None req-a2366500-67be-46eb-90d6-0efe293c8fc5 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Task: {'id': task-1377916, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 988.515366] env[61923]: DEBUG oslo_vmware.api [None req-5f0ffb82-8d31-4829-b8c1-ae776078cdda tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Task: {'id': task-1377915, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.139299} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 988.515607] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-5f0ffb82-8d31-4829-b8c1-ae776078cdda tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Deleted the datastore file {{(pid=61923) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 988.515745] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-5f0ffb82-8d31-4829-b8c1-ae776078cdda tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] [instance: a511392e-9ab6-42fb-b07c-f90e9786dcc1] Deleted contents of the VM from datastore datastore1 {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 988.515923] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-5f0ffb82-8d31-4829-b8c1-ae776078cdda tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] [instance: a511392e-9ab6-42fb-b07c-f90e9786dcc1] Instance destroyed {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 988.516114] env[61923]: INFO nova.compute.manager [None req-5f0ffb82-8d31-4829-b8c1-ae776078cdda tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] [instance: a511392e-9ab6-42fb-b07c-f90e9786dcc1] Took 1.14 seconds to destroy the instance on the hypervisor. [ 988.516369] env[61923]: DEBUG oslo.service.loopingcall [None req-5f0ffb82-8d31-4829-b8c1-ae776078cdda tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61923) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 988.516565] env[61923]: DEBUG nova.compute.manager [-] [instance: a511392e-9ab6-42fb-b07c-f90e9786dcc1] Deallocating network for instance {{(pid=61923) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 988.516660] env[61923]: DEBUG nova.network.neutron [-] [instance: a511392e-9ab6-42fb-b07c-f90e9786dcc1] deallocate_for_instance() {{(pid=61923) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 988.712686] env[61923]: DEBUG oslo_vmware.api [None req-a2366500-67be-46eb-90d6-0efe293c8fc5 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Task: {'id': task-1377916, 'name': PowerOffVM_Task, 'duration_secs': 0.193743} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 988.712975] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-a2366500-67be-46eb-90d6-0efe293c8fc5 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: 3fcedef5-9040-4b45-9e27-f0700112d528] Powered off the VM {{(pid=61923) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 988.713193] env[61923]: DEBUG nova.virt.vmwareapi.volumeops [None req-a2366500-67be-46eb-90d6-0efe293c8fc5 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: 3fcedef5-9040-4b45-9e27-f0700112d528] Volume detach. Driver type: vmdk {{(pid=61923) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 988.713391] env[61923]: DEBUG nova.virt.vmwareapi.volumeops [None req-a2366500-67be-46eb-90d6-0efe293c8fc5 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: 3fcedef5-9040-4b45-9e27-f0700112d528] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-292739', 'volume_id': '85aa34e0-345d-403c-ab7a-58822ce81f4e', 'name': 'volume-85aa34e0-345d-403c-ab7a-58822ce81f4e', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '3fcedef5-9040-4b45-9e27-f0700112d528', 'attached_at': '', 'detached_at': '', 'volume_id': '85aa34e0-345d-403c-ab7a-58822ce81f4e', 'serial': '85aa34e0-345d-403c-ab7a-58822ce81f4e'} {{(pid=61923) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 988.714174] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-653567d9-d5f2-478d-baea-21fa0897b364 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.735618] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6bd95f75-821c-4e3d-aac4-10c4ed71a1f5 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.742633] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d73f240-9ef2-47e6-9392-377799513d6c {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.762861] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc61716a-e55a-433b-850b-27175f51502b {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.777489] env[61923]: DEBUG nova.virt.vmwareapi.volumeops [None req-a2366500-67be-46eb-90d6-0efe293c8fc5 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] The volume has not been displaced from its original location: [datastore2] volume-85aa34e0-345d-403c-ab7a-58822ce81f4e/volume-85aa34e0-345d-403c-ab7a-58822ce81f4e.vmdk. No consolidation needed. {{(pid=61923) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 988.782812] env[61923]: DEBUG nova.virt.vmwareapi.volumeops [None req-a2366500-67be-46eb-90d6-0efe293c8fc5 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: 3fcedef5-9040-4b45-9e27-f0700112d528] Reconfiguring VM instance instance-0000005d to detach disk 2001 {{(pid=61923) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 988.784511] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-40bdc212-7c46-4f96-8057-8e31ef5ecd42 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.807930] env[61923]: DEBUG nova.compute.manager [req-161204e8-a5a8-4d02-8b15-0a4b735ec12f req-1f958734-2449-46fd-9cd5-12e0be04b0a6 service nova] [instance: a511392e-9ab6-42fb-b07c-f90e9786dcc1] Received event network-vif-deleted-0325cd38-bd49-4132-85e3-c66fc9efe49f {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 988.808166] env[61923]: INFO nova.compute.manager [req-161204e8-a5a8-4d02-8b15-0a4b735ec12f req-1f958734-2449-46fd-9cd5-12e0be04b0a6 service nova] [instance: a511392e-9ab6-42fb-b07c-f90e9786dcc1] Neutron deleted interface 0325cd38-bd49-4132-85e3-c66fc9efe49f; detaching it from the instance and deleting it from the info cache [ 988.808341] env[61923]: DEBUG nova.network.neutron [req-161204e8-a5a8-4d02-8b15-0a4b735ec12f req-1f958734-2449-46fd-9cd5-12e0be04b0a6 service nova] [instance: a511392e-9ab6-42fb-b07c-f90e9786dcc1] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 988.810573] env[61923]: DEBUG oslo_vmware.api [None req-a2366500-67be-46eb-90d6-0efe293c8fc5 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Waiting for the task: (returnval){ [ 988.810573] env[61923]: value = "task-1377917" [ 988.810573] env[61923]: _type = "Task" [ 988.810573] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 988.820093] env[61923]: DEBUG oslo_vmware.api [None req-a2366500-67be-46eb-90d6-0efe293c8fc5 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Task: {'id': task-1377917, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 989.298265] env[61923]: DEBUG nova.network.neutron [-] [instance: a511392e-9ab6-42fb-b07c-f90e9786dcc1] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 989.312230] env[61923]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ff20ab95-5184-41f7-bdbd-d0273506b61c {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.323501] env[61923]: DEBUG oslo_vmware.api [None req-a2366500-67be-46eb-90d6-0efe293c8fc5 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Task: {'id': task-1377917, 'name': ReconfigVM_Task, 'duration_secs': 0.201932} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 989.324675] env[61923]: DEBUG nova.virt.vmwareapi.volumeops [None req-a2366500-67be-46eb-90d6-0efe293c8fc5 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: 3fcedef5-9040-4b45-9e27-f0700112d528] Reconfigured VM instance instance-0000005d to detach disk 2001 {{(pid=61923) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 989.329675] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9ed15f09-763e-484f-bef0-ca1d83114558 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.341605] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c66e8ae-0c55-4b38-8bf6-3f9e8cb65ac1 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.357385] env[61923]: DEBUG oslo_vmware.api [None req-a2366500-67be-46eb-90d6-0efe293c8fc5 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Waiting for the task: (returnval){ [ 989.357385] env[61923]: value = "task-1377918" [ 989.357385] env[61923]: _type = "Task" [ 989.357385] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 989.368184] env[61923]: DEBUG nova.compute.manager [req-161204e8-a5a8-4d02-8b15-0a4b735ec12f req-1f958734-2449-46fd-9cd5-12e0be04b0a6 service nova] [instance: a511392e-9ab6-42fb-b07c-f90e9786dcc1] Detach interface failed, port_id=0325cd38-bd49-4132-85e3-c66fc9efe49f, reason: Instance a511392e-9ab6-42fb-b07c-f90e9786dcc1 could not be found. {{(pid=61923) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 989.371632] env[61923]: DEBUG oslo_vmware.api [None req-a2366500-67be-46eb-90d6-0efe293c8fc5 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Task: {'id': task-1377918, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 989.800780] env[61923]: INFO nova.compute.manager [-] [instance: a511392e-9ab6-42fb-b07c-f90e9786dcc1] Took 1.28 seconds to deallocate network for instance. [ 989.868516] env[61923]: DEBUG oslo_vmware.api [None req-a2366500-67be-46eb-90d6-0efe293c8fc5 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Task: {'id': task-1377918, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 990.307600] env[61923]: DEBUG oslo_concurrency.lockutils [None req-5f0ffb82-8d31-4829-b8c1-ae776078cdda tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 990.308073] env[61923]: DEBUG oslo_concurrency.lockutils [None req-5f0ffb82-8d31-4829-b8c1-ae776078cdda tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 990.308189] env[61923]: DEBUG nova.objects.instance [None req-5f0ffb82-8d31-4829-b8c1-ae776078cdda tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Lazy-loading 'resources' on Instance uuid a511392e-9ab6-42fb-b07c-f90e9786dcc1 {{(pid=61923) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 990.367544] env[61923]: DEBUG oslo_vmware.api [None req-a2366500-67be-46eb-90d6-0efe293c8fc5 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Task: {'id': task-1377918, 'name': ReconfigVM_Task, 'duration_secs': 0.585811} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 990.367830] env[61923]: DEBUG nova.virt.vmwareapi.volumeops [None req-a2366500-67be-46eb-90d6-0efe293c8fc5 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: 3fcedef5-9040-4b45-9e27-f0700112d528] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-292739', 'volume_id': '85aa34e0-345d-403c-ab7a-58822ce81f4e', 'name': 'volume-85aa34e0-345d-403c-ab7a-58822ce81f4e', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '3fcedef5-9040-4b45-9e27-f0700112d528', 'attached_at': '', 'detached_at': '', 'volume_id': '85aa34e0-345d-403c-ab7a-58822ce81f4e', 'serial': '85aa34e0-345d-403c-ab7a-58822ce81f4e'} {{(pid=61923) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 990.368200] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-a2366500-67be-46eb-90d6-0efe293c8fc5 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: 3fcedef5-9040-4b45-9e27-f0700112d528] Destroying instance {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 990.368962] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05e29bb7-f6d7-43a3-8666-d8a634b76ce8 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.375189] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-a2366500-67be-46eb-90d6-0efe293c8fc5 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: 3fcedef5-9040-4b45-9e27-f0700112d528] Unregistering the VM {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 990.375412] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-3c0e1188-0d90-4c93-bfc5-113c04755c21 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.445133] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-a2366500-67be-46eb-90d6-0efe293c8fc5 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: 3fcedef5-9040-4b45-9e27-f0700112d528] Unregistered the VM {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 990.445500] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-a2366500-67be-46eb-90d6-0efe293c8fc5 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: 3fcedef5-9040-4b45-9e27-f0700112d528] Deleting contents of the VM from datastore datastore1 {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 990.445747] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-a2366500-67be-46eb-90d6-0efe293c8fc5 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Deleting the datastore file [datastore1] 3fcedef5-9040-4b45-9e27-f0700112d528 {{(pid=61923) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 990.446037] env[61923]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-9713905d-cfeb-44dd-94cc-f46fe264eae7 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.452539] env[61923]: DEBUG oslo_vmware.api [None req-a2366500-67be-46eb-90d6-0efe293c8fc5 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Waiting for the task: (returnval){ [ 990.452539] env[61923]: value = "task-1377920" [ 990.452539] env[61923]: _type = "Task" [ 990.452539] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 990.459746] env[61923]: DEBUG oslo_vmware.api [None req-a2366500-67be-46eb-90d6-0efe293c8fc5 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Task: {'id': task-1377920, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 990.871180] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1125189-df3a-4cdb-94e9-a4ce2f30e70f {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.878988] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3eb485b1-874a-4788-9adf-9cfe38e93dd3 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.911274] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9c521dd-feab-487c-ae08-b0a8ae273111 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.918178] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-552769c8-b4f5-445e-b939-08d363b06e8f {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.930878] env[61923]: DEBUG nova.compute.provider_tree [None req-5f0ffb82-8d31-4829-b8c1-ae776078cdda tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 990.961148] env[61923]: DEBUG oslo_vmware.api [None req-a2366500-67be-46eb-90d6-0efe293c8fc5 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Task: {'id': task-1377920, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 991.433660] env[61923]: DEBUG nova.scheduler.client.report [None req-5f0ffb82-8d31-4829-b8c1-ae776078cdda tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 991.462385] env[61923]: DEBUG oslo_vmware.api [None req-a2366500-67be-46eb-90d6-0efe293c8fc5 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Task: {'id': task-1377920, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 991.938985] env[61923]: DEBUG oslo_concurrency.lockutils [None req-5f0ffb82-8d31-4829-b8c1-ae776078cdda tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.631s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 991.965385] env[61923]: DEBUG oslo_vmware.api [None req-a2366500-67be-46eb-90d6-0efe293c8fc5 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Task: {'id': task-1377920, 'name': DeleteDatastoreFile_Task, 'duration_secs': 1.414643} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 991.966266] env[61923]: INFO nova.scheduler.client.report [None req-5f0ffb82-8d31-4829-b8c1-ae776078cdda tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Deleted allocations for instance a511392e-9ab6-42fb-b07c-f90e9786dcc1 [ 991.967282] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-a2366500-67be-46eb-90d6-0efe293c8fc5 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Deleted the datastore file {{(pid=61923) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 991.967423] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-a2366500-67be-46eb-90d6-0efe293c8fc5 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: 3fcedef5-9040-4b45-9e27-f0700112d528] Deleted contents of the VM from datastore datastore1 {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 991.967593] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-a2366500-67be-46eb-90d6-0efe293c8fc5 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: 3fcedef5-9040-4b45-9e27-f0700112d528] Instance destroyed {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 991.967772] env[61923]: INFO nova.compute.manager [None req-a2366500-67be-46eb-90d6-0efe293c8fc5 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: 3fcedef5-9040-4b45-9e27-f0700112d528] Took 3.78 seconds to destroy the instance on the hypervisor. [ 991.968032] env[61923]: DEBUG oslo.service.loopingcall [None req-a2366500-67be-46eb-90d6-0efe293c8fc5 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61923) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 991.970355] env[61923]: DEBUG nova.compute.manager [-] [instance: 3fcedef5-9040-4b45-9e27-f0700112d528] Deallocating network for instance {{(pid=61923) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 991.970355] env[61923]: DEBUG nova.network.neutron [-] [instance: 3fcedef5-9040-4b45-9e27-f0700112d528] deallocate_for_instance() {{(pid=61923) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 992.415573] env[61923]: DEBUG nova.compute.manager [req-a81d9535-49d8-4ad2-bd9e-a46802fd406e req-4267c251-214f-46b6-b761-c755ae8dbcbd service nova] [instance: 3fcedef5-9040-4b45-9e27-f0700112d528] Received event network-vif-deleted-a1c503e4-c553-4bfb-b9ae-84d4ebb6937a {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 992.415573] env[61923]: INFO nova.compute.manager [req-a81d9535-49d8-4ad2-bd9e-a46802fd406e req-4267c251-214f-46b6-b761-c755ae8dbcbd service nova] [instance: 3fcedef5-9040-4b45-9e27-f0700112d528] Neutron deleted interface a1c503e4-c553-4bfb-b9ae-84d4ebb6937a; detaching it from the instance and deleting it from the info cache [ 992.415746] env[61923]: DEBUG nova.network.neutron [req-a81d9535-49d8-4ad2-bd9e-a46802fd406e req-4267c251-214f-46b6-b761-c755ae8dbcbd service nova] [instance: 3fcedef5-9040-4b45-9e27-f0700112d528] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 992.475119] env[61923]: DEBUG oslo_concurrency.lockutils [None req-5f0ffb82-8d31-4829-b8c1-ae776078cdda tempest-ServersNegativeTestJSON-882234871 tempest-ServersNegativeTestJSON-882234871-project-member] Lock "a511392e-9ab6-42fb-b07c-f90e9786dcc1" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.102s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 992.894199] env[61923]: DEBUG nova.network.neutron [-] [instance: 3fcedef5-9040-4b45-9e27-f0700112d528] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 992.920409] env[61923]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ca9a81d3-7659-4022-b2ec-27e2f73d4a42 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.931837] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8bec1b08-c59e-4749-b3b5-c246174b0e80 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.956403] env[61923]: DEBUG nova.compute.manager [req-a81d9535-49d8-4ad2-bd9e-a46802fd406e req-4267c251-214f-46b6-b761-c755ae8dbcbd service nova] [instance: 3fcedef5-9040-4b45-9e27-f0700112d528] Detach interface failed, port_id=a1c503e4-c553-4bfb-b9ae-84d4ebb6937a, reason: Instance 3fcedef5-9040-4b45-9e27-f0700112d528 could not be found. {{(pid=61923) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 993.398704] env[61923]: INFO nova.compute.manager [-] [instance: 3fcedef5-9040-4b45-9e27-f0700112d528] Took 1.43 seconds to deallocate network for instance. [ 993.945394] env[61923]: INFO nova.compute.manager [None req-a2366500-67be-46eb-90d6-0efe293c8fc5 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: 3fcedef5-9040-4b45-9e27-f0700112d528] Took 0.55 seconds to detach 1 volumes for instance. [ 994.451549] env[61923]: DEBUG oslo_concurrency.lockutils [None req-a2366500-67be-46eb-90d6-0efe293c8fc5 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 994.451824] env[61923]: DEBUG oslo_concurrency.lockutils [None req-a2366500-67be-46eb-90d6-0efe293c8fc5 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 994.452061] env[61923]: DEBUG nova.objects.instance [None req-a2366500-67be-46eb-90d6-0efe293c8fc5 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Lazy-loading 'resources' on Instance uuid 3fcedef5-9040-4b45-9e27-f0700112d528 {{(pid=61923) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 995.005058] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ba46b4b-62d0-4158-b138-7c4201b2aa6d {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.012639] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-282daad4-69d8-4ae6-91e3-063127af157a {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.042726] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ef108c2-1712-4cac-be18-37d3c87add69 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.049520] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-692573cd-5db4-4c3d-b343-dd6ea71a2074 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.062060] env[61923]: DEBUG nova.compute.provider_tree [None req-a2366500-67be-46eb-90d6-0efe293c8fc5 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 995.564784] env[61923]: DEBUG nova.scheduler.client.report [None req-a2366500-67be-46eb-90d6-0efe293c8fc5 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 996.069562] env[61923]: DEBUG oslo_concurrency.lockutils [None req-a2366500-67be-46eb-90d6-0efe293c8fc5 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.618s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 996.091766] env[61923]: INFO nova.scheduler.client.report [None req-a2366500-67be-46eb-90d6-0efe293c8fc5 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Deleted allocations for instance 3fcedef5-9040-4b45-9e27-f0700112d528 [ 996.599294] env[61923]: DEBUG oslo_concurrency.lockutils [None req-a2366500-67be-46eb-90d6-0efe293c8fc5 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Lock "3fcedef5-9040-4b45-9e27-f0700112d528" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 8.412s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 997.013024] env[61923]: DEBUG oslo_concurrency.lockutils [None req-441b5d3c-129b-46cf-84bc-aff8092ca7a7 tempest-ServersV294TestFqdnHostnames-833840938 tempest-ServersV294TestFqdnHostnames-833840938-project-member] Acquiring lock "aba844c1-d488-4e95-a550-bd1cb7331fb2" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 997.013024] env[61923]: DEBUG oslo_concurrency.lockutils [None req-441b5d3c-129b-46cf-84bc-aff8092ca7a7 tempest-ServersV294TestFqdnHostnames-833840938 tempest-ServersV294TestFqdnHostnames-833840938-project-member] Lock "aba844c1-d488-4e95-a550-bd1cb7331fb2" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 997.013024] env[61923]: DEBUG oslo_concurrency.lockutils [None req-441b5d3c-129b-46cf-84bc-aff8092ca7a7 tempest-ServersV294TestFqdnHostnames-833840938 tempest-ServersV294TestFqdnHostnames-833840938-project-member] Acquiring lock "aba844c1-d488-4e95-a550-bd1cb7331fb2-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 997.013024] env[61923]: DEBUG oslo_concurrency.lockutils [None req-441b5d3c-129b-46cf-84bc-aff8092ca7a7 tempest-ServersV294TestFqdnHostnames-833840938 tempest-ServersV294TestFqdnHostnames-833840938-project-member] Lock "aba844c1-d488-4e95-a550-bd1cb7331fb2-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 997.013024] env[61923]: DEBUG oslo_concurrency.lockutils [None req-441b5d3c-129b-46cf-84bc-aff8092ca7a7 tempest-ServersV294TestFqdnHostnames-833840938 tempest-ServersV294TestFqdnHostnames-833840938-project-member] Lock "aba844c1-d488-4e95-a550-bd1cb7331fb2-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 997.014674] env[61923]: INFO nova.compute.manager [None req-441b5d3c-129b-46cf-84bc-aff8092ca7a7 tempest-ServersV294TestFqdnHostnames-833840938 tempest-ServersV294TestFqdnHostnames-833840938-project-member] [instance: aba844c1-d488-4e95-a550-bd1cb7331fb2] Terminating instance [ 997.016684] env[61923]: DEBUG nova.compute.manager [None req-441b5d3c-129b-46cf-84bc-aff8092ca7a7 tempest-ServersV294TestFqdnHostnames-833840938 tempest-ServersV294TestFqdnHostnames-833840938-project-member] [instance: aba844c1-d488-4e95-a550-bd1cb7331fb2] Start destroying the instance on the hypervisor. {{(pid=61923) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 997.016919] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-441b5d3c-129b-46cf-84bc-aff8092ca7a7 tempest-ServersV294TestFqdnHostnames-833840938 tempest-ServersV294TestFqdnHostnames-833840938-project-member] [instance: aba844c1-d488-4e95-a550-bd1cb7331fb2] Destroying instance {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 997.017780] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b20dae6-4022-40a3-9d86-24820ab60504 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.026451] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-441b5d3c-129b-46cf-84bc-aff8092ca7a7 tempest-ServersV294TestFqdnHostnames-833840938 tempest-ServersV294TestFqdnHostnames-833840938-project-member] [instance: aba844c1-d488-4e95-a550-bd1cb7331fb2] Powering off the VM {{(pid=61923) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 997.026451] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-834d36cc-10fd-4ad3-83f4-3595bb4d3efc {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.031839] env[61923]: DEBUG oslo_vmware.api [None req-441b5d3c-129b-46cf-84bc-aff8092ca7a7 tempest-ServersV294TestFqdnHostnames-833840938 tempest-ServersV294TestFqdnHostnames-833840938-project-member] Waiting for the task: (returnval){ [ 997.031839] env[61923]: value = "task-1377921" [ 997.031839] env[61923]: _type = "Task" [ 997.031839] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 997.039833] env[61923]: DEBUG oslo_vmware.api [None req-441b5d3c-129b-46cf-84bc-aff8092ca7a7 tempest-ServersV294TestFqdnHostnames-833840938 tempest-ServersV294TestFqdnHostnames-833840938-project-member] Task: {'id': task-1377921, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.542895] env[61923]: DEBUG oslo_vmware.api [None req-441b5d3c-129b-46cf-84bc-aff8092ca7a7 tempest-ServersV294TestFqdnHostnames-833840938 tempest-ServersV294TestFqdnHostnames-833840938-project-member] Task: {'id': task-1377921, 'name': PowerOffVM_Task, 'duration_secs': 0.184867} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 997.543181] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-441b5d3c-129b-46cf-84bc-aff8092ca7a7 tempest-ServersV294TestFqdnHostnames-833840938 tempest-ServersV294TestFqdnHostnames-833840938-project-member] [instance: aba844c1-d488-4e95-a550-bd1cb7331fb2] Powered off the VM {{(pid=61923) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 997.543295] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-441b5d3c-129b-46cf-84bc-aff8092ca7a7 tempest-ServersV294TestFqdnHostnames-833840938 tempest-ServersV294TestFqdnHostnames-833840938-project-member] [instance: aba844c1-d488-4e95-a550-bd1cb7331fb2] Unregistering the VM {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 997.543595] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-462ab7ae-0027-4109-b3c7-d467f9584c77 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.604344] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-441b5d3c-129b-46cf-84bc-aff8092ca7a7 tempest-ServersV294TestFqdnHostnames-833840938 tempest-ServersV294TestFqdnHostnames-833840938-project-member] [instance: aba844c1-d488-4e95-a550-bd1cb7331fb2] Unregistered the VM {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 997.604575] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-441b5d3c-129b-46cf-84bc-aff8092ca7a7 tempest-ServersV294TestFqdnHostnames-833840938 tempest-ServersV294TestFqdnHostnames-833840938-project-member] [instance: aba844c1-d488-4e95-a550-bd1cb7331fb2] Deleting contents of the VM from datastore datastore1 {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 997.604762] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-441b5d3c-129b-46cf-84bc-aff8092ca7a7 tempest-ServersV294TestFqdnHostnames-833840938 tempest-ServersV294TestFqdnHostnames-833840938-project-member] Deleting the datastore file [datastore1] aba844c1-d488-4e95-a550-bd1cb7331fb2 {{(pid=61923) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 997.605127] env[61923]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f3319896-1797-47ca-9b09-37d6ec54c61b {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.612046] env[61923]: DEBUG oslo_vmware.api [None req-441b5d3c-129b-46cf-84bc-aff8092ca7a7 tempest-ServersV294TestFqdnHostnames-833840938 tempest-ServersV294TestFqdnHostnames-833840938-project-member] Waiting for the task: (returnval){ [ 997.612046] env[61923]: value = "task-1377923" [ 997.612046] env[61923]: _type = "Task" [ 997.612046] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 997.619054] env[61923]: DEBUG oslo_vmware.api [None req-441b5d3c-129b-46cf-84bc-aff8092ca7a7 tempest-ServersV294TestFqdnHostnames-833840938 tempest-ServersV294TestFqdnHostnames-833840938-project-member] Task: {'id': task-1377923, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.850723] env[61923]: DEBUG oslo_concurrency.lockutils [None req-fdfd1e98-519f-4911-9320-f9cd337334e8 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Acquiring lock "a2f97947-6506-40b3-a24a-4ac03b88c5f5" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 997.850964] env[61923]: DEBUG oslo_concurrency.lockutils [None req-fdfd1e98-519f-4911-9320-f9cd337334e8 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Lock "a2f97947-6506-40b3-a24a-4ac03b88c5f5" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 998.122660] env[61923]: DEBUG oslo_vmware.api [None req-441b5d3c-129b-46cf-84bc-aff8092ca7a7 tempest-ServersV294TestFqdnHostnames-833840938 tempest-ServersV294TestFqdnHostnames-833840938-project-member] Task: {'id': task-1377923, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.137622} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 998.122938] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-441b5d3c-129b-46cf-84bc-aff8092ca7a7 tempest-ServersV294TestFqdnHostnames-833840938 tempest-ServersV294TestFqdnHostnames-833840938-project-member] Deleted the datastore file {{(pid=61923) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 998.123171] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-441b5d3c-129b-46cf-84bc-aff8092ca7a7 tempest-ServersV294TestFqdnHostnames-833840938 tempest-ServersV294TestFqdnHostnames-833840938-project-member] [instance: aba844c1-d488-4e95-a550-bd1cb7331fb2] Deleted contents of the VM from datastore datastore1 {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 998.123397] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-441b5d3c-129b-46cf-84bc-aff8092ca7a7 tempest-ServersV294TestFqdnHostnames-833840938 tempest-ServersV294TestFqdnHostnames-833840938-project-member] [instance: aba844c1-d488-4e95-a550-bd1cb7331fb2] Instance destroyed {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 998.123608] env[61923]: INFO nova.compute.manager [None req-441b5d3c-129b-46cf-84bc-aff8092ca7a7 tempest-ServersV294TestFqdnHostnames-833840938 tempest-ServersV294TestFqdnHostnames-833840938-project-member] [instance: aba844c1-d488-4e95-a550-bd1cb7331fb2] Took 1.11 seconds to destroy the instance on the hypervisor. [ 998.123863] env[61923]: DEBUG oslo.service.loopingcall [None req-441b5d3c-129b-46cf-84bc-aff8092ca7a7 tempest-ServersV294TestFqdnHostnames-833840938 tempest-ServersV294TestFqdnHostnames-833840938-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61923) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 998.124085] env[61923]: DEBUG nova.compute.manager [-] [instance: aba844c1-d488-4e95-a550-bd1cb7331fb2] Deallocating network for instance {{(pid=61923) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 998.124186] env[61923]: DEBUG nova.network.neutron [-] [instance: aba844c1-d488-4e95-a550-bd1cb7331fb2] deallocate_for_instance() {{(pid=61923) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 998.353081] env[61923]: DEBUG nova.compute.manager [None req-fdfd1e98-519f-4911-9320-f9cd337334e8 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: a2f97947-6506-40b3-a24a-4ac03b88c5f5] Starting instance... {{(pid=61923) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 998.569938] env[61923]: DEBUG nova.compute.manager [req-2fa8695b-97e1-41f1-9af2-3851e0b44f38 req-c60419d0-56dd-437a-b4d8-51dea6b23a14 service nova] [instance: aba844c1-d488-4e95-a550-bd1cb7331fb2] Received event network-vif-deleted-6326df12-72bc-44d6-b660-4abdb5ca5cfb {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 998.570186] env[61923]: INFO nova.compute.manager [req-2fa8695b-97e1-41f1-9af2-3851e0b44f38 req-c60419d0-56dd-437a-b4d8-51dea6b23a14 service nova] [instance: aba844c1-d488-4e95-a550-bd1cb7331fb2] Neutron deleted interface 6326df12-72bc-44d6-b660-4abdb5ca5cfb; detaching it from the instance and deleting it from the info cache [ 998.570309] env[61923]: DEBUG nova.network.neutron [req-2fa8695b-97e1-41f1-9af2-3851e0b44f38 req-c60419d0-56dd-437a-b4d8-51dea6b23a14 service nova] [instance: aba844c1-d488-4e95-a550-bd1cb7331fb2] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 998.875151] env[61923]: DEBUG oslo_concurrency.lockutils [None req-fdfd1e98-519f-4911-9320-f9cd337334e8 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 998.875418] env[61923]: DEBUG oslo_concurrency.lockutils [None req-fdfd1e98-519f-4911-9320-f9cd337334e8 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 998.876994] env[61923]: INFO nova.compute.claims [None req-fdfd1e98-519f-4911-9320-f9cd337334e8 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: a2f97947-6506-40b3-a24a-4ac03b88c5f5] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 999.048810] env[61923]: DEBUG nova.network.neutron [-] [instance: aba844c1-d488-4e95-a550-bd1cb7331fb2] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 999.073080] env[61923]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-fadcc168-57af-48f1-9258-b5ea75822f7f {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.082982] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51c5bd37-e718-47b2-89b3-d80d797cb650 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.106767] env[61923]: DEBUG nova.compute.manager [req-2fa8695b-97e1-41f1-9af2-3851e0b44f38 req-c60419d0-56dd-437a-b4d8-51dea6b23a14 service nova] [instance: aba844c1-d488-4e95-a550-bd1cb7331fb2] Detach interface failed, port_id=6326df12-72bc-44d6-b660-4abdb5ca5cfb, reason: Instance aba844c1-d488-4e95-a550-bd1cb7331fb2 could not be found. {{(pid=61923) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 999.552103] env[61923]: INFO nova.compute.manager [-] [instance: aba844c1-d488-4e95-a550-bd1cb7331fb2] Took 1.43 seconds to deallocate network for instance. [ 999.931485] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f35f5375-5b80-4931-bbe4-6a85e1930b7a {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.938506] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-840e3a4a-26ef-4475-ac35-7926d56cd97a {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.968437] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d06cf87-7ee0-494f-9556-563904a7a447 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.975101] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e28cdaa-1bcc-429f-92c4-c07350712635 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.988010] env[61923]: DEBUG nova.compute.provider_tree [None req-fdfd1e98-519f-4911-9320-f9cd337334e8 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1000.058076] env[61923]: DEBUG oslo_concurrency.lockutils [None req-441b5d3c-129b-46cf-84bc-aff8092ca7a7 tempest-ServersV294TestFqdnHostnames-833840938 tempest-ServersV294TestFqdnHostnames-833840938-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1000.490933] env[61923]: DEBUG nova.scheduler.client.report [None req-fdfd1e98-519f-4911-9320-f9cd337334e8 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1000.995430] env[61923]: DEBUG oslo_concurrency.lockutils [None req-fdfd1e98-519f-4911-9320-f9cd337334e8 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.120s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1000.995983] env[61923]: DEBUG nova.compute.manager [None req-fdfd1e98-519f-4911-9320-f9cd337334e8 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: a2f97947-6506-40b3-a24a-4ac03b88c5f5] Start building networks asynchronously for instance. {{(pid=61923) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1000.998521] env[61923]: DEBUG oslo_concurrency.lockutils [None req-441b5d3c-129b-46cf-84bc-aff8092ca7a7 tempest-ServersV294TestFqdnHostnames-833840938 tempest-ServersV294TestFqdnHostnames-833840938-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.941s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1000.998747] env[61923]: DEBUG nova.objects.instance [None req-441b5d3c-129b-46cf-84bc-aff8092ca7a7 tempest-ServersV294TestFqdnHostnames-833840938 tempest-ServersV294TestFqdnHostnames-833840938-project-member] Lazy-loading 'resources' on Instance uuid aba844c1-d488-4e95-a550-bd1cb7331fb2 {{(pid=61923) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1001.502276] env[61923]: DEBUG nova.compute.utils [None req-fdfd1e98-519f-4911-9320-f9cd337334e8 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Using /dev/sd instead of None {{(pid=61923) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1001.503706] env[61923]: DEBUG nova.compute.manager [None req-fdfd1e98-519f-4911-9320-f9cd337334e8 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: a2f97947-6506-40b3-a24a-4ac03b88c5f5] Allocating IP information in the background. {{(pid=61923) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1001.503872] env[61923]: DEBUG nova.network.neutron [None req-fdfd1e98-519f-4911-9320-f9cd337334e8 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: a2f97947-6506-40b3-a24a-4ac03b88c5f5] allocate_for_instance() {{(pid=61923) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1001.551188] env[61923]: DEBUG nova.policy [None req-fdfd1e98-519f-4911-9320-f9cd337334e8 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '69be89006ff141b686261d643b3b9a73', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '780d3541d9604417b977bb62390c4299', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61923) authorize /opt/stack/nova/nova/policy.py:201}} [ 1001.554009] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7878f87d-ec2a-45d7-90bc-ef2b62c2e597 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.561312] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4eaeb937-6323-4707-a51e-7e6b2ea9798c {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.592085] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-605ffbdf-da22-400f-8251-fdc950bc9f45 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.599420] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b022696-cc63-4fb0-bca6-91934ee35520 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.613103] env[61923]: DEBUG nova.compute.provider_tree [None req-441b5d3c-129b-46cf-84bc-aff8092ca7a7 tempest-ServersV294TestFqdnHostnames-833840938 tempest-ServersV294TestFqdnHostnames-833840938-project-member] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1001.794753] env[61923]: DEBUG nova.network.neutron [None req-fdfd1e98-519f-4911-9320-f9cd337334e8 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: a2f97947-6506-40b3-a24a-4ac03b88c5f5] Successfully created port: 1aa4cc84-2b2c-4ca9-9b99-e2c94daacf4a {{(pid=61923) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1002.006876] env[61923]: DEBUG nova.compute.manager [None req-fdfd1e98-519f-4911-9320-f9cd337334e8 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: a2f97947-6506-40b3-a24a-4ac03b88c5f5] Start building block device mappings for instance. {{(pid=61923) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1002.116323] env[61923]: DEBUG nova.scheduler.client.report [None req-441b5d3c-129b-46cf-84bc-aff8092ca7a7 tempest-ServersV294TestFqdnHostnames-833840938 tempest-ServersV294TestFqdnHostnames-833840938-project-member] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1002.621251] env[61923]: DEBUG oslo_concurrency.lockutils [None req-441b5d3c-129b-46cf-84bc-aff8092ca7a7 tempest-ServersV294TestFqdnHostnames-833840938 tempest-ServersV294TestFqdnHostnames-833840938-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.623s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1002.645940] env[61923]: INFO nova.scheduler.client.report [None req-441b5d3c-129b-46cf-84bc-aff8092ca7a7 tempest-ServersV294TestFqdnHostnames-833840938 tempest-ServersV294TestFqdnHostnames-833840938-project-member] Deleted allocations for instance aba844c1-d488-4e95-a550-bd1cb7331fb2 [ 1003.016047] env[61923]: DEBUG nova.compute.manager [None req-fdfd1e98-519f-4911-9320-f9cd337334e8 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: a2f97947-6506-40b3-a24a-4ac03b88c5f5] Start spawning the instance on the hypervisor. {{(pid=61923) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1003.040896] env[61923]: DEBUG nova.virt.hardware [None req-fdfd1e98-519f-4911-9320-f9cd337334e8 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-29T20:07:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-29T20:07:35Z,direct_url=,disk_format='vmdk',id=0f153f63-ae0a-45b1-b7f5-7f9b673c947f,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='4e7b5e3b3c3443c8bd5c70f8a15739f5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-29T20:07:36Z,virtual_size=,visibility=), allow threads: False {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1003.041571] env[61923]: DEBUG nova.virt.hardware [None req-fdfd1e98-519f-4911-9320-f9cd337334e8 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Flavor limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1003.041571] env[61923]: DEBUG nova.virt.hardware [None req-fdfd1e98-519f-4911-9320-f9cd337334e8 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Image limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1003.041571] env[61923]: DEBUG nova.virt.hardware [None req-fdfd1e98-519f-4911-9320-f9cd337334e8 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Flavor pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1003.041732] env[61923]: DEBUG nova.virt.hardware [None req-fdfd1e98-519f-4911-9320-f9cd337334e8 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Image pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1003.041732] env[61923]: DEBUG nova.virt.hardware [None req-fdfd1e98-519f-4911-9320-f9cd337334e8 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1003.041946] env[61923]: DEBUG nova.virt.hardware [None req-fdfd1e98-519f-4911-9320-f9cd337334e8 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1003.042133] env[61923]: DEBUG nova.virt.hardware [None req-fdfd1e98-519f-4911-9320-f9cd337334e8 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1003.042289] env[61923]: DEBUG nova.virt.hardware [None req-fdfd1e98-519f-4911-9320-f9cd337334e8 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Got 1 possible topologies {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1003.042450] env[61923]: DEBUG nova.virt.hardware [None req-fdfd1e98-519f-4911-9320-f9cd337334e8 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1003.042622] env[61923]: DEBUG nova.virt.hardware [None req-fdfd1e98-519f-4911-9320-f9cd337334e8 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1003.043492] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4149c9e-a1d4-4a93-89fd-cb0da1d6bc4f {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.051384] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c781c89f-af47-47a4-aafe-ca900157addc {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.154845] env[61923]: DEBUG oslo_concurrency.lockutils [None req-441b5d3c-129b-46cf-84bc-aff8092ca7a7 tempest-ServersV294TestFqdnHostnames-833840938 tempest-ServersV294TestFqdnHostnames-833840938-project-member] Lock "aba844c1-d488-4e95-a550-bd1cb7331fb2" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.143s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1003.185947] env[61923]: DEBUG nova.compute.manager [req-a912ae88-b212-406d-8103-d0969de07ef0 req-a5e74f38-d719-4326-8c8b-c8e323f96365 service nova] [instance: a2f97947-6506-40b3-a24a-4ac03b88c5f5] Received event network-vif-plugged-1aa4cc84-2b2c-4ca9-9b99-e2c94daacf4a {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1003.185947] env[61923]: DEBUG oslo_concurrency.lockutils [req-a912ae88-b212-406d-8103-d0969de07ef0 req-a5e74f38-d719-4326-8c8b-c8e323f96365 service nova] Acquiring lock "a2f97947-6506-40b3-a24a-4ac03b88c5f5-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1003.185947] env[61923]: DEBUG oslo_concurrency.lockutils [req-a912ae88-b212-406d-8103-d0969de07ef0 req-a5e74f38-d719-4326-8c8b-c8e323f96365 service nova] Lock "a2f97947-6506-40b3-a24a-4ac03b88c5f5-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1003.186336] env[61923]: DEBUG oslo_concurrency.lockutils [req-a912ae88-b212-406d-8103-d0969de07ef0 req-a5e74f38-d719-4326-8c8b-c8e323f96365 service nova] Lock "a2f97947-6506-40b3-a24a-4ac03b88c5f5-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1003.186336] env[61923]: DEBUG nova.compute.manager [req-a912ae88-b212-406d-8103-d0969de07ef0 req-a5e74f38-d719-4326-8c8b-c8e323f96365 service nova] [instance: a2f97947-6506-40b3-a24a-4ac03b88c5f5] No waiting events found dispatching network-vif-plugged-1aa4cc84-2b2c-4ca9-9b99-e2c94daacf4a {{(pid=61923) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1003.186429] env[61923]: WARNING nova.compute.manager [req-a912ae88-b212-406d-8103-d0969de07ef0 req-a5e74f38-d719-4326-8c8b-c8e323f96365 service nova] [instance: a2f97947-6506-40b3-a24a-4ac03b88c5f5] Received unexpected event network-vif-plugged-1aa4cc84-2b2c-4ca9-9b99-e2c94daacf4a for instance with vm_state building and task_state spawning. [ 1003.757291] env[61923]: DEBUG nova.network.neutron [None req-fdfd1e98-519f-4911-9320-f9cd337334e8 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: a2f97947-6506-40b3-a24a-4ac03b88c5f5] Successfully updated port: 1aa4cc84-2b2c-4ca9-9b99-e2c94daacf4a {{(pid=61923) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1003.791197] env[61923]: DEBUG nova.compute.manager [req-51bc3418-66c7-4263-9ef9-3945b8ea0906 req-fc893dde-a2b2-45cd-a9d4-55fee00fd057 service nova] [instance: a2f97947-6506-40b3-a24a-4ac03b88c5f5] Received event network-changed-1aa4cc84-2b2c-4ca9-9b99-e2c94daacf4a {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1003.791197] env[61923]: DEBUG nova.compute.manager [req-51bc3418-66c7-4263-9ef9-3945b8ea0906 req-fc893dde-a2b2-45cd-a9d4-55fee00fd057 service nova] [instance: a2f97947-6506-40b3-a24a-4ac03b88c5f5] Refreshing instance network info cache due to event network-changed-1aa4cc84-2b2c-4ca9-9b99-e2c94daacf4a. {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1003.791197] env[61923]: DEBUG oslo_concurrency.lockutils [req-51bc3418-66c7-4263-9ef9-3945b8ea0906 req-fc893dde-a2b2-45cd-a9d4-55fee00fd057 service nova] Acquiring lock "refresh_cache-a2f97947-6506-40b3-a24a-4ac03b88c5f5" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1003.791197] env[61923]: DEBUG oslo_concurrency.lockutils [req-51bc3418-66c7-4263-9ef9-3945b8ea0906 req-fc893dde-a2b2-45cd-a9d4-55fee00fd057 service nova] Acquired lock "refresh_cache-a2f97947-6506-40b3-a24a-4ac03b88c5f5" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1003.791197] env[61923]: DEBUG nova.network.neutron [req-51bc3418-66c7-4263-9ef9-3945b8ea0906 req-fc893dde-a2b2-45cd-a9d4-55fee00fd057 service nova] [instance: a2f97947-6506-40b3-a24a-4ac03b88c5f5] Refreshing network info cache for port 1aa4cc84-2b2c-4ca9-9b99-e2c94daacf4a {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1004.259053] env[61923]: DEBUG oslo_concurrency.lockutils [None req-fdfd1e98-519f-4911-9320-f9cd337334e8 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Acquiring lock "refresh_cache-a2f97947-6506-40b3-a24a-4ac03b88c5f5" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1004.327377] env[61923]: DEBUG nova.network.neutron [req-51bc3418-66c7-4263-9ef9-3945b8ea0906 req-fc893dde-a2b2-45cd-a9d4-55fee00fd057 service nova] [instance: a2f97947-6506-40b3-a24a-4ac03b88c5f5] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1004.399187] env[61923]: DEBUG nova.network.neutron [req-51bc3418-66c7-4263-9ef9-3945b8ea0906 req-fc893dde-a2b2-45cd-a9d4-55fee00fd057 service nova] [instance: a2f97947-6506-40b3-a24a-4ac03b88c5f5] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1004.901654] env[61923]: DEBUG oslo_concurrency.lockutils [req-51bc3418-66c7-4263-9ef9-3945b8ea0906 req-fc893dde-a2b2-45cd-a9d4-55fee00fd057 service nova] Releasing lock "refresh_cache-a2f97947-6506-40b3-a24a-4ac03b88c5f5" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1004.902436] env[61923]: DEBUG oslo_concurrency.lockutils [None req-fdfd1e98-519f-4911-9320-f9cd337334e8 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Acquired lock "refresh_cache-a2f97947-6506-40b3-a24a-4ac03b88c5f5" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1004.902436] env[61923]: DEBUG nova.network.neutron [None req-fdfd1e98-519f-4911-9320-f9cd337334e8 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: a2f97947-6506-40b3-a24a-4ac03b88c5f5] Building network info cache for instance {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1005.433291] env[61923]: DEBUG nova.network.neutron [None req-fdfd1e98-519f-4911-9320-f9cd337334e8 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: a2f97947-6506-40b3-a24a-4ac03b88c5f5] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1005.565390] env[61923]: DEBUG nova.network.neutron [None req-fdfd1e98-519f-4911-9320-f9cd337334e8 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: a2f97947-6506-40b3-a24a-4ac03b88c5f5] Updating instance_info_cache with network_info: [{"id": "1aa4cc84-2b2c-4ca9-9b99-e2c94daacf4a", "address": "fa:16:3e:a1:5b:41", "network": {"id": "9533e79d-fde2-4c10-86a0-86a36845a8cf", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1736743137-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "780d3541d9604417b977bb62390c4299", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6d650b26-c3e7-4de7-98db-5e4b816d123a", "external-id": "nsx-vlan-transportzone-757", "segmentation_id": 757, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1aa4cc84-2b", "ovs_interfaceid": "1aa4cc84-2b2c-4ca9-9b99-e2c94daacf4a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1006.068022] env[61923]: DEBUG oslo_concurrency.lockutils [None req-fdfd1e98-519f-4911-9320-f9cd337334e8 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Releasing lock "refresh_cache-a2f97947-6506-40b3-a24a-4ac03b88c5f5" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1006.068367] env[61923]: DEBUG nova.compute.manager [None req-fdfd1e98-519f-4911-9320-f9cd337334e8 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: a2f97947-6506-40b3-a24a-4ac03b88c5f5] Instance network_info: |[{"id": "1aa4cc84-2b2c-4ca9-9b99-e2c94daacf4a", "address": "fa:16:3e:a1:5b:41", "network": {"id": "9533e79d-fde2-4c10-86a0-86a36845a8cf", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1736743137-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "780d3541d9604417b977bb62390c4299", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6d650b26-c3e7-4de7-98db-5e4b816d123a", "external-id": "nsx-vlan-transportzone-757", "segmentation_id": 757, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1aa4cc84-2b", "ovs_interfaceid": "1aa4cc84-2b2c-4ca9-9b99-e2c94daacf4a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61923) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1006.068822] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-fdfd1e98-519f-4911-9320-f9cd337334e8 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: a2f97947-6506-40b3-a24a-4ac03b88c5f5] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:a1:5b:41', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '6d650b26-c3e7-4de7-98db-5e4b816d123a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '1aa4cc84-2b2c-4ca9-9b99-e2c94daacf4a', 'vif_model': 'vmxnet3'}] {{(pid=61923) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1006.076792] env[61923]: DEBUG oslo.service.loopingcall [None req-fdfd1e98-519f-4911-9320-f9cd337334e8 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61923) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1006.077015] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a2f97947-6506-40b3-a24a-4ac03b88c5f5] Creating VM on the ESX host {{(pid=61923) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1006.077253] env[61923]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-0b4e2b94-cf9d-4020-9d6f-5fb25f4153ab {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.097069] env[61923]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1006.097069] env[61923]: value = "task-1377924" [ 1006.097069] env[61923]: _type = "Task" [ 1006.097069] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1006.104532] env[61923]: DEBUG oslo_vmware.api [-] Task: {'id': task-1377924, 'name': CreateVM_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1006.616593] env[61923]: DEBUG oslo_vmware.api [-] Task: {'id': task-1377924, 'name': CreateVM_Task, 'duration_secs': 0.320807} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1006.616813] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a2f97947-6506-40b3-a24a-4ac03b88c5f5] Created VM on the ESX host {{(pid=61923) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1006.618671] env[61923]: DEBUG oslo_concurrency.lockutils [None req-fdfd1e98-519f-4911-9320-f9cd337334e8 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1006.619228] env[61923]: DEBUG oslo_concurrency.lockutils [None req-fdfd1e98-519f-4911-9320-f9cd337334e8 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1006.619564] env[61923]: DEBUG oslo_concurrency.lockutils [None req-fdfd1e98-519f-4911-9320-f9cd337334e8 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1006.619879] env[61923]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5c7b9e8f-9d68-4564-b050-fd8d03ae965b {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.624848] env[61923]: DEBUG oslo_vmware.api [None req-fdfd1e98-519f-4911-9320-f9cd337334e8 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Waiting for the task: (returnval){ [ 1006.624848] env[61923]: value = "session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52c0ccb1-755b-f3a4-a296-17c28fe0e0f7" [ 1006.624848] env[61923]: _type = "Task" [ 1006.624848] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1006.635184] env[61923]: DEBUG oslo_vmware.api [None req-fdfd1e98-519f-4911-9320-f9cd337334e8 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52c0ccb1-755b-f3a4-a296-17c28fe0e0f7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1007.135050] env[61923]: DEBUG oslo_vmware.api [None req-fdfd1e98-519f-4911-9320-f9cd337334e8 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52c0ccb1-755b-f3a4-a296-17c28fe0e0f7, 'name': SearchDatastore_Task, 'duration_secs': 0.012251} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1007.135373] env[61923]: DEBUG oslo_concurrency.lockutils [None req-fdfd1e98-519f-4911-9320-f9cd337334e8 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1007.135611] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-fdfd1e98-519f-4911-9320-f9cd337334e8 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: a2f97947-6506-40b3-a24a-4ac03b88c5f5] Processing image 0f153f63-ae0a-45b1-b7f5-7f9b673c947f {{(pid=61923) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1007.135845] env[61923]: DEBUG oslo_concurrency.lockutils [None req-fdfd1e98-519f-4911-9320-f9cd337334e8 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1007.135995] env[61923]: DEBUG oslo_concurrency.lockutils [None req-fdfd1e98-519f-4911-9320-f9cd337334e8 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1007.136187] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-fdfd1e98-519f-4911-9320-f9cd337334e8 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61923) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1007.136450] env[61923]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-eb1fc458-45cc-4cbd-9870-1da6ec6f8dbf {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.145365] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-fdfd1e98-519f-4911-9320-f9cd337334e8 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61923) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1007.145572] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-fdfd1e98-519f-4911-9320-f9cd337334e8 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61923) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1007.146354] env[61923]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0d5e08e6-333b-468d-b169-c26494641122 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.151146] env[61923]: DEBUG oslo_vmware.api [None req-fdfd1e98-519f-4911-9320-f9cd337334e8 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Waiting for the task: (returnval){ [ 1007.151146] env[61923]: value = "session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]5235536b-4a18-9d10-4d13-2033b5c3ab05" [ 1007.151146] env[61923]: _type = "Task" [ 1007.151146] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1007.159517] env[61923]: DEBUG oslo_vmware.api [None req-fdfd1e98-519f-4911-9320-f9cd337334e8 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]5235536b-4a18-9d10-4d13-2033b5c3ab05, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1007.661373] env[61923]: DEBUG oslo_vmware.api [None req-fdfd1e98-519f-4911-9320-f9cd337334e8 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]5235536b-4a18-9d10-4d13-2033b5c3ab05, 'name': SearchDatastore_Task, 'duration_secs': 0.008336} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1007.662278] env[61923]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9d839cae-ffa8-4e6d-a5d1-ac3c6096e7f1 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.667443] env[61923]: DEBUG oslo_vmware.api [None req-fdfd1e98-519f-4911-9320-f9cd337334e8 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Waiting for the task: (returnval){ [ 1007.667443] env[61923]: value = "session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]522d0b5f-8cb7-5de5-a16e-3863d29fed2f" [ 1007.667443] env[61923]: _type = "Task" [ 1007.667443] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1007.674935] env[61923]: DEBUG oslo_vmware.api [None req-fdfd1e98-519f-4911-9320-f9cd337334e8 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]522d0b5f-8cb7-5de5-a16e-3863d29fed2f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1008.039156] env[61923]: DEBUG oslo_concurrency.lockutils [None req-e2375eee-5ef0-47c4-8309-7404fe8cdfd1 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Acquiring lock "0c3788fc-6cce-4806-b4f8-84f84489fb93" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1008.039400] env[61923]: DEBUG oslo_concurrency.lockutils [None req-e2375eee-5ef0-47c4-8309-7404fe8cdfd1 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Lock "0c3788fc-6cce-4806-b4f8-84f84489fb93" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1008.179099] env[61923]: DEBUG oslo_vmware.api [None req-fdfd1e98-519f-4911-9320-f9cd337334e8 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]522d0b5f-8cb7-5de5-a16e-3863d29fed2f, 'name': SearchDatastore_Task, 'duration_secs': 0.009554} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1008.179377] env[61923]: DEBUG oslo_concurrency.lockutils [None req-fdfd1e98-519f-4911-9320-f9cd337334e8 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1008.179659] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-fdfd1e98-519f-4911-9320-f9cd337334e8 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk to [datastore1] a2f97947-6506-40b3-a24a-4ac03b88c5f5/a2f97947-6506-40b3-a24a-4ac03b88c5f5.vmdk {{(pid=61923) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1008.179958] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-bc0c117a-5249-4e96-9864-a05edf3a0c12 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.187712] env[61923]: DEBUG oslo_vmware.api [None req-fdfd1e98-519f-4911-9320-f9cd337334e8 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Waiting for the task: (returnval){ [ 1008.187712] env[61923]: value = "task-1377925" [ 1008.187712] env[61923]: _type = "Task" [ 1008.187712] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1008.195166] env[61923]: DEBUG oslo_vmware.api [None req-fdfd1e98-519f-4911-9320-f9cd337334e8 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Task: {'id': task-1377925, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1008.542347] env[61923]: DEBUG nova.compute.manager [None req-e2375eee-5ef0-47c4-8309-7404fe8cdfd1 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: 0c3788fc-6cce-4806-b4f8-84f84489fb93] Starting instance... {{(pid=61923) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1008.698021] env[61923]: DEBUG oslo_vmware.api [None req-fdfd1e98-519f-4911-9320-f9cd337334e8 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Task: {'id': task-1377925, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.451151} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1008.698021] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-fdfd1e98-519f-4911-9320-f9cd337334e8 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk to [datastore1] a2f97947-6506-40b3-a24a-4ac03b88c5f5/a2f97947-6506-40b3-a24a-4ac03b88c5f5.vmdk {{(pid=61923) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1008.698382] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-fdfd1e98-519f-4911-9320-f9cd337334e8 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: a2f97947-6506-40b3-a24a-4ac03b88c5f5] Extending root virtual disk to 1048576 {{(pid=61923) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1008.698714] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-f4e508a2-7414-4fb3-bfcd-53f78bacc2c3 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.705447] env[61923]: DEBUG oslo_vmware.api [None req-fdfd1e98-519f-4911-9320-f9cd337334e8 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Waiting for the task: (returnval){ [ 1008.705447] env[61923]: value = "task-1377926" [ 1008.705447] env[61923]: _type = "Task" [ 1008.705447] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1008.714125] env[61923]: DEBUG oslo_vmware.api [None req-fdfd1e98-519f-4911-9320-f9cd337334e8 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Task: {'id': task-1377926, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1009.064747] env[61923]: DEBUG oslo_concurrency.lockutils [None req-e2375eee-5ef0-47c4-8309-7404fe8cdfd1 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1009.065048] env[61923]: DEBUG oslo_concurrency.lockutils [None req-e2375eee-5ef0-47c4-8309-7404fe8cdfd1 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1009.066479] env[61923]: INFO nova.compute.claims [None req-e2375eee-5ef0-47c4-8309-7404fe8cdfd1 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: 0c3788fc-6cce-4806-b4f8-84f84489fb93] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1009.215427] env[61923]: DEBUG oslo_vmware.api [None req-fdfd1e98-519f-4911-9320-f9cd337334e8 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Task: {'id': task-1377926, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.069204} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1009.215940] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-fdfd1e98-519f-4911-9320-f9cd337334e8 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: a2f97947-6506-40b3-a24a-4ac03b88c5f5] Extended root virtual disk {{(pid=61923) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1009.216759] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ffe4afe-ccfa-46c6-97e3-201705394858 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.237701] env[61923]: DEBUG nova.virt.vmwareapi.volumeops [None req-fdfd1e98-519f-4911-9320-f9cd337334e8 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: a2f97947-6506-40b3-a24a-4ac03b88c5f5] Reconfiguring VM instance instance-0000005e to attach disk [datastore1] a2f97947-6506-40b3-a24a-4ac03b88c5f5/a2f97947-6506-40b3-a24a-4ac03b88c5f5.vmdk or device None with type sparse {{(pid=61923) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1009.237975] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f99150a2-9f88-4d3c-a8b4-ace4275d070a {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.258227] env[61923]: DEBUG oslo_vmware.api [None req-fdfd1e98-519f-4911-9320-f9cd337334e8 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Waiting for the task: (returnval){ [ 1009.258227] env[61923]: value = "task-1377927" [ 1009.258227] env[61923]: _type = "Task" [ 1009.258227] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1009.265462] env[61923]: DEBUG oslo_vmware.api [None req-fdfd1e98-519f-4911-9320-f9cd337334e8 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Task: {'id': task-1377927, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1009.768300] env[61923]: DEBUG oslo_vmware.api [None req-fdfd1e98-519f-4911-9320-f9cd337334e8 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Task: {'id': task-1377927, 'name': ReconfigVM_Task, 'duration_secs': 0.281107} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1009.768608] env[61923]: DEBUG nova.virt.vmwareapi.volumeops [None req-fdfd1e98-519f-4911-9320-f9cd337334e8 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: a2f97947-6506-40b3-a24a-4ac03b88c5f5] Reconfigured VM instance instance-0000005e to attach disk [datastore1] a2f97947-6506-40b3-a24a-4ac03b88c5f5/a2f97947-6506-40b3-a24a-4ac03b88c5f5.vmdk or device None with type sparse {{(pid=61923) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1009.769199] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b8b064d3-4261-456c-afb3-1e27236cddb5 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.775961] env[61923]: DEBUG oslo_vmware.api [None req-fdfd1e98-519f-4911-9320-f9cd337334e8 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Waiting for the task: (returnval){ [ 1009.775961] env[61923]: value = "task-1377928" [ 1009.775961] env[61923]: _type = "Task" [ 1009.775961] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1009.783396] env[61923]: DEBUG oslo_vmware.api [None req-fdfd1e98-519f-4911-9320-f9cd337334e8 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Task: {'id': task-1377928, 'name': Rename_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1010.125949] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce03e678-7020-49bc-b449-bce39048d443 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.133622] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae6d1cdc-c46b-4bfd-a2b9-88d0b5dad6a7 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.162732] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7460b30a-d602-4624-9405-8f89150d9257 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.169408] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88b4b5ce-ef90-4fcb-bbcf-0daf9da56564 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.183116] env[61923]: DEBUG nova.compute.provider_tree [None req-e2375eee-5ef0-47c4-8309-7404fe8cdfd1 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1010.286986] env[61923]: DEBUG oslo_vmware.api [None req-fdfd1e98-519f-4911-9320-f9cd337334e8 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Task: {'id': task-1377928, 'name': Rename_Task} progress is 99%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1010.686763] env[61923]: DEBUG nova.scheduler.client.report [None req-e2375eee-5ef0-47c4-8309-7404fe8cdfd1 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1010.786420] env[61923]: DEBUG oslo_vmware.api [None req-fdfd1e98-519f-4911-9320-f9cd337334e8 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Task: {'id': task-1377928, 'name': Rename_Task} progress is 99%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1011.190917] env[61923]: DEBUG oslo_concurrency.lockutils [None req-e2375eee-5ef0-47c4-8309-7404fe8cdfd1 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.126s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1011.191500] env[61923]: DEBUG nova.compute.manager [None req-e2375eee-5ef0-47c4-8309-7404fe8cdfd1 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: 0c3788fc-6cce-4806-b4f8-84f84489fb93] Start building networks asynchronously for instance. {{(pid=61923) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1011.287288] env[61923]: DEBUG oslo_vmware.api [None req-fdfd1e98-519f-4911-9320-f9cd337334e8 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Task: {'id': task-1377928, 'name': Rename_Task, 'duration_secs': 1.137776} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1011.287852] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-fdfd1e98-519f-4911-9320-f9cd337334e8 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: a2f97947-6506-40b3-a24a-4ac03b88c5f5] Powering on the VM {{(pid=61923) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1011.288111] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-135ff89f-7086-4eb1-9565-92565223ccc0 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.294395] env[61923]: DEBUG oslo_vmware.api [None req-fdfd1e98-519f-4911-9320-f9cd337334e8 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Waiting for the task: (returnval){ [ 1011.294395] env[61923]: value = "task-1377929" [ 1011.294395] env[61923]: _type = "Task" [ 1011.294395] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1011.301794] env[61923]: DEBUG oslo_vmware.api [None req-fdfd1e98-519f-4911-9320-f9cd337334e8 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Task: {'id': task-1377929, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1011.696239] env[61923]: DEBUG nova.compute.utils [None req-e2375eee-5ef0-47c4-8309-7404fe8cdfd1 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Using /dev/sd instead of None {{(pid=61923) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1011.697705] env[61923]: DEBUG nova.compute.manager [None req-e2375eee-5ef0-47c4-8309-7404fe8cdfd1 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: 0c3788fc-6cce-4806-b4f8-84f84489fb93] Allocating IP information in the background. {{(pid=61923) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1011.697854] env[61923]: DEBUG nova.network.neutron [None req-e2375eee-5ef0-47c4-8309-7404fe8cdfd1 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: 0c3788fc-6cce-4806-b4f8-84f84489fb93] allocate_for_instance() {{(pid=61923) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1011.737964] env[61923]: DEBUG nova.policy [None req-e2375eee-5ef0-47c4-8309-7404fe8cdfd1 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6087a385c42e4383aa26a7f586cca87f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5a4db4f8aa194a219e80c538cc715d43', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61923) authorize /opt/stack/nova/nova/policy.py:201}} [ 1011.804024] env[61923]: DEBUG oslo_vmware.api [None req-fdfd1e98-519f-4911-9320-f9cd337334e8 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Task: {'id': task-1377929, 'name': PowerOnVM_Task, 'duration_secs': 0.437517} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1011.804367] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-fdfd1e98-519f-4911-9320-f9cd337334e8 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: a2f97947-6506-40b3-a24a-4ac03b88c5f5] Powered on the VM {{(pid=61923) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1011.804535] env[61923]: INFO nova.compute.manager [None req-fdfd1e98-519f-4911-9320-f9cd337334e8 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: a2f97947-6506-40b3-a24a-4ac03b88c5f5] Took 8.79 seconds to spawn the instance on the hypervisor. [ 1011.804718] env[61923]: DEBUG nova.compute.manager [None req-fdfd1e98-519f-4911-9320-f9cd337334e8 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: a2f97947-6506-40b3-a24a-4ac03b88c5f5] Checking state {{(pid=61923) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1011.805498] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31262d14-7f05-4d10-b77b-ec8bd8182501 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.978809] env[61923]: DEBUG nova.network.neutron [None req-e2375eee-5ef0-47c4-8309-7404fe8cdfd1 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: 0c3788fc-6cce-4806-b4f8-84f84489fb93] Successfully created port: 504f15e6-4eee-4266-817f-d7935b5d0609 {{(pid=61923) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1012.201391] env[61923]: DEBUG nova.compute.manager [None req-e2375eee-5ef0-47c4-8309-7404fe8cdfd1 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: 0c3788fc-6cce-4806-b4f8-84f84489fb93] Start building block device mappings for instance. {{(pid=61923) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1012.321727] env[61923]: INFO nova.compute.manager [None req-fdfd1e98-519f-4911-9320-f9cd337334e8 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: a2f97947-6506-40b3-a24a-4ac03b88c5f5] Took 13.46 seconds to build instance. [ 1012.823969] env[61923]: DEBUG oslo_concurrency.lockutils [None req-fdfd1e98-519f-4911-9320-f9cd337334e8 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Lock "a2f97947-6506-40b3-a24a-4ac03b88c5f5" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.973s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1013.212145] env[61923]: DEBUG nova.compute.manager [None req-e2375eee-5ef0-47c4-8309-7404fe8cdfd1 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: 0c3788fc-6cce-4806-b4f8-84f84489fb93] Start spawning the instance on the hypervisor. {{(pid=61923) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1013.232896] env[61923]: DEBUG nova.virt.hardware [None req-e2375eee-5ef0-47c4-8309-7404fe8cdfd1 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-29T20:07:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-29T20:07:35Z,direct_url=,disk_format='vmdk',id=0f153f63-ae0a-45b1-b7f5-7f9b673c947f,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='4e7b5e3b3c3443c8bd5c70f8a15739f5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-29T20:07:36Z,virtual_size=,visibility=), allow threads: False {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1013.233178] env[61923]: DEBUG nova.virt.hardware [None req-e2375eee-5ef0-47c4-8309-7404fe8cdfd1 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Flavor limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1013.233345] env[61923]: DEBUG nova.virt.hardware [None req-e2375eee-5ef0-47c4-8309-7404fe8cdfd1 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Image limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1013.233528] env[61923]: DEBUG nova.virt.hardware [None req-e2375eee-5ef0-47c4-8309-7404fe8cdfd1 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Flavor pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1013.233674] env[61923]: DEBUG nova.virt.hardware [None req-e2375eee-5ef0-47c4-8309-7404fe8cdfd1 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Image pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1013.233820] env[61923]: DEBUG nova.virt.hardware [None req-e2375eee-5ef0-47c4-8309-7404fe8cdfd1 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1013.234129] env[61923]: DEBUG nova.virt.hardware [None req-e2375eee-5ef0-47c4-8309-7404fe8cdfd1 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1013.234341] env[61923]: DEBUG nova.virt.hardware [None req-e2375eee-5ef0-47c4-8309-7404fe8cdfd1 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1013.234520] env[61923]: DEBUG nova.virt.hardware [None req-e2375eee-5ef0-47c4-8309-7404fe8cdfd1 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Got 1 possible topologies {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1013.234688] env[61923]: DEBUG nova.virt.hardware [None req-e2375eee-5ef0-47c4-8309-7404fe8cdfd1 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1013.234862] env[61923]: DEBUG nova.virt.hardware [None req-e2375eee-5ef0-47c4-8309-7404fe8cdfd1 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1013.235818] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb351fea-7856-4f90-a1ee-0fdea8800afb {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.245918] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6932c47-0524-40cf-a386-e2825926d9f4 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.347361] env[61923]: DEBUG nova.compute.manager [req-a3f632c8-f8f8-45f7-a4fc-cb7fa936c701 req-a335f0b4-78c9-4d30-a795-86d80a259eba service nova] [instance: 0c3788fc-6cce-4806-b4f8-84f84489fb93] Received event network-vif-plugged-504f15e6-4eee-4266-817f-d7935b5d0609 {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1013.347634] env[61923]: DEBUG oslo_concurrency.lockutils [req-a3f632c8-f8f8-45f7-a4fc-cb7fa936c701 req-a335f0b4-78c9-4d30-a795-86d80a259eba service nova] Acquiring lock "0c3788fc-6cce-4806-b4f8-84f84489fb93-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1013.348045] env[61923]: DEBUG oslo_concurrency.lockutils [req-a3f632c8-f8f8-45f7-a4fc-cb7fa936c701 req-a335f0b4-78c9-4d30-a795-86d80a259eba service nova] Lock "0c3788fc-6cce-4806-b4f8-84f84489fb93-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1013.348278] env[61923]: DEBUG oslo_concurrency.lockutils [req-a3f632c8-f8f8-45f7-a4fc-cb7fa936c701 req-a335f0b4-78c9-4d30-a795-86d80a259eba service nova] Lock "0c3788fc-6cce-4806-b4f8-84f84489fb93-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1013.348498] env[61923]: DEBUG nova.compute.manager [req-a3f632c8-f8f8-45f7-a4fc-cb7fa936c701 req-a335f0b4-78c9-4d30-a795-86d80a259eba service nova] [instance: 0c3788fc-6cce-4806-b4f8-84f84489fb93] No waiting events found dispatching network-vif-plugged-504f15e6-4eee-4266-817f-d7935b5d0609 {{(pid=61923) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1013.348685] env[61923]: WARNING nova.compute.manager [req-a3f632c8-f8f8-45f7-a4fc-cb7fa936c701 req-a335f0b4-78c9-4d30-a795-86d80a259eba service nova] [instance: 0c3788fc-6cce-4806-b4f8-84f84489fb93] Received unexpected event network-vif-plugged-504f15e6-4eee-4266-817f-d7935b5d0609 for instance with vm_state building and task_state spawning. [ 1013.435796] env[61923]: DEBUG nova.network.neutron [None req-e2375eee-5ef0-47c4-8309-7404fe8cdfd1 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: 0c3788fc-6cce-4806-b4f8-84f84489fb93] Successfully updated port: 504f15e6-4eee-4266-817f-d7935b5d0609 {{(pid=61923) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1013.542970] env[61923]: DEBUG oslo_concurrency.lockutils [None req-ca0212d9-f46f-4fe8-98d2-915ce325cd19 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Acquiring lock "a2f97947-6506-40b3-a24a-4ac03b88c5f5" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1013.543270] env[61923]: DEBUG oslo_concurrency.lockutils [None req-ca0212d9-f46f-4fe8-98d2-915ce325cd19 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Lock "a2f97947-6506-40b3-a24a-4ac03b88c5f5" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1013.543455] env[61923]: INFO nova.compute.manager [None req-ca0212d9-f46f-4fe8-98d2-915ce325cd19 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: a2f97947-6506-40b3-a24a-4ac03b88c5f5] Shelving [ 1013.938126] env[61923]: DEBUG oslo_concurrency.lockutils [None req-e2375eee-5ef0-47c4-8309-7404fe8cdfd1 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Acquiring lock "refresh_cache-0c3788fc-6cce-4806-b4f8-84f84489fb93" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1013.938438] env[61923]: DEBUG oslo_concurrency.lockutils [None req-e2375eee-5ef0-47c4-8309-7404fe8cdfd1 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Acquired lock "refresh_cache-0c3788fc-6cce-4806-b4f8-84f84489fb93" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1013.938488] env[61923]: DEBUG nova.network.neutron [None req-e2375eee-5ef0-47c4-8309-7404fe8cdfd1 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: 0c3788fc-6cce-4806-b4f8-84f84489fb93] Building network info cache for instance {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1014.050090] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-ca0212d9-f46f-4fe8-98d2-915ce325cd19 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: a2f97947-6506-40b3-a24a-4ac03b88c5f5] Powering off the VM {{(pid=61923) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1014.050358] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-afd8238b-f693-465d-997e-35c3e3a1dbb9 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.057869] env[61923]: DEBUG oslo_vmware.api [None req-ca0212d9-f46f-4fe8-98d2-915ce325cd19 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Waiting for the task: (returnval){ [ 1014.057869] env[61923]: value = "task-1377930" [ 1014.057869] env[61923]: _type = "Task" [ 1014.057869] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1014.066020] env[61923]: DEBUG oslo_vmware.api [None req-ca0212d9-f46f-4fe8-98d2-915ce325cd19 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Task: {'id': task-1377930, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.469609] env[61923]: DEBUG nova.network.neutron [None req-e2375eee-5ef0-47c4-8309-7404fe8cdfd1 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: 0c3788fc-6cce-4806-b4f8-84f84489fb93] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1014.569867] env[61923]: DEBUG oslo_vmware.api [None req-ca0212d9-f46f-4fe8-98d2-915ce325cd19 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Task: {'id': task-1377930, 'name': PowerOffVM_Task, 'duration_secs': 0.174455} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1014.570197] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-ca0212d9-f46f-4fe8-98d2-915ce325cd19 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: a2f97947-6506-40b3-a24a-4ac03b88c5f5] Powered off the VM {{(pid=61923) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1014.571043] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-359ecdfd-5ede-4c45-af93-0348d4259d21 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.590184] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d36433ce-217e-47e1-b87c-8d13a9bd5a72 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.595098] env[61923]: DEBUG nova.network.neutron [None req-e2375eee-5ef0-47c4-8309-7404fe8cdfd1 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: 0c3788fc-6cce-4806-b4f8-84f84489fb93] Updating instance_info_cache with network_info: [{"id": "504f15e6-4eee-4266-817f-d7935b5d0609", "address": "fa:16:3e:37:68:89", "network": {"id": "77aec6ca-d1be-4b39-9d51-08853a1183a1", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1813085414-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5a4db4f8aa194a219e80c538cc715d43", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "56b944d8-803d-43f2-945d-0f334ee4ea1c", "external-id": "nsx-vlan-transportzone-799", "segmentation_id": 799, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap504f15e6-4e", "ovs_interfaceid": "504f15e6-4eee-4266-817f-d7935b5d0609", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1015.101586] env[61923]: DEBUG oslo_concurrency.lockutils [None req-e2375eee-5ef0-47c4-8309-7404fe8cdfd1 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Releasing lock "refresh_cache-0c3788fc-6cce-4806-b4f8-84f84489fb93" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1015.101920] env[61923]: DEBUG nova.compute.manager [None req-e2375eee-5ef0-47c4-8309-7404fe8cdfd1 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: 0c3788fc-6cce-4806-b4f8-84f84489fb93] Instance network_info: |[{"id": "504f15e6-4eee-4266-817f-d7935b5d0609", "address": "fa:16:3e:37:68:89", "network": {"id": "77aec6ca-d1be-4b39-9d51-08853a1183a1", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1813085414-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5a4db4f8aa194a219e80c538cc715d43", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "56b944d8-803d-43f2-945d-0f334ee4ea1c", "external-id": "nsx-vlan-transportzone-799", "segmentation_id": 799, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap504f15e6-4e", "ovs_interfaceid": "504f15e6-4eee-4266-817f-d7935b5d0609", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61923) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1015.102862] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-ca0212d9-f46f-4fe8-98d2-915ce325cd19 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: a2f97947-6506-40b3-a24a-4ac03b88c5f5] Creating Snapshot of the VM instance {{(pid=61923) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 1015.103277] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-e2375eee-5ef0-47c4-8309-7404fe8cdfd1 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: 0c3788fc-6cce-4806-b4f8-84f84489fb93] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:37:68:89', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '56b944d8-803d-43f2-945d-0f334ee4ea1c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '504f15e6-4eee-4266-817f-d7935b5d0609', 'vif_model': 'vmxnet3'}] {{(pid=61923) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1015.110548] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-e2375eee-5ef0-47c4-8309-7404fe8cdfd1 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Creating folder: Project (5a4db4f8aa194a219e80c538cc715d43). Parent ref: group-v292629. {{(pid=61923) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1015.110801] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-aee5cefb-0a35-4180-8679-7ce806a69e05 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.113109] env[61923]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a5dad711-18c0-4d98-af83-1013e3eb259c {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.120619] env[61923]: DEBUG oslo_vmware.api [None req-ca0212d9-f46f-4fe8-98d2-915ce325cd19 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Waiting for the task: (returnval){ [ 1015.120619] env[61923]: value = "task-1377932" [ 1015.120619] env[61923]: _type = "Task" [ 1015.120619] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1015.124558] env[61923]: INFO nova.virt.vmwareapi.vm_util [None req-e2375eee-5ef0-47c4-8309-7404fe8cdfd1 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Created folder: Project (5a4db4f8aa194a219e80c538cc715d43) in parent group-v292629. [ 1015.124741] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-e2375eee-5ef0-47c4-8309-7404fe8cdfd1 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Creating folder: Instances. Parent ref: group-v292741. {{(pid=61923) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1015.125274] env[61923]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-b3d75791-d677-4ce7-be7a-0d79d1e6f8c7 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.130364] env[61923]: DEBUG oslo_vmware.api [None req-ca0212d9-f46f-4fe8-98d2-915ce325cd19 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Task: {'id': task-1377932, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1015.135100] env[61923]: INFO nova.virt.vmwareapi.vm_util [None req-e2375eee-5ef0-47c4-8309-7404fe8cdfd1 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Created folder: Instances in parent group-v292741. [ 1015.135379] env[61923]: DEBUG oslo.service.loopingcall [None req-e2375eee-5ef0-47c4-8309-7404fe8cdfd1 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61923) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1015.135611] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0c3788fc-6cce-4806-b4f8-84f84489fb93] Creating VM on the ESX host {{(pid=61923) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1015.135859] env[61923]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-7e8d6e72-0774-4249-804b-4d5eb473466d {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.155692] env[61923]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1015.155692] env[61923]: value = "task-1377934" [ 1015.155692] env[61923]: _type = "Task" [ 1015.155692] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1015.162760] env[61923]: DEBUG oslo_vmware.api [-] Task: {'id': task-1377934, 'name': CreateVM_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1015.308674] env[61923]: DEBUG oslo_concurrency.lockutils [None req-9ecd4890-934b-4cbe-9edf-89f5103075d2 tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Acquiring lock "b779d183-89ae-4e4d-ae99-e514e145ed43" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1015.309026] env[61923]: DEBUG oslo_concurrency.lockutils [None req-9ecd4890-934b-4cbe-9edf-89f5103075d2 tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Lock "b779d183-89ae-4e4d-ae99-e514e145ed43" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1015.309322] env[61923]: DEBUG oslo_concurrency.lockutils [None req-9ecd4890-934b-4cbe-9edf-89f5103075d2 tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Acquiring lock "b779d183-89ae-4e4d-ae99-e514e145ed43-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1015.309548] env[61923]: DEBUG oslo_concurrency.lockutils [None req-9ecd4890-934b-4cbe-9edf-89f5103075d2 tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Lock "b779d183-89ae-4e4d-ae99-e514e145ed43-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1015.309748] env[61923]: DEBUG oslo_concurrency.lockutils [None req-9ecd4890-934b-4cbe-9edf-89f5103075d2 tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Lock "b779d183-89ae-4e4d-ae99-e514e145ed43-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1015.312236] env[61923]: INFO nova.compute.manager [None req-9ecd4890-934b-4cbe-9edf-89f5103075d2 tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] [instance: b779d183-89ae-4e4d-ae99-e514e145ed43] Terminating instance [ 1015.314192] env[61923]: DEBUG nova.compute.manager [None req-9ecd4890-934b-4cbe-9edf-89f5103075d2 tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] [instance: b779d183-89ae-4e4d-ae99-e514e145ed43] Start destroying the instance on the hypervisor. {{(pid=61923) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1015.314408] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-9ecd4890-934b-4cbe-9edf-89f5103075d2 tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] [instance: b779d183-89ae-4e4d-ae99-e514e145ed43] Destroying instance {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1015.315330] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a023a06-6c4f-4008-92d1-003ccd9afb2f {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.323692] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-9ecd4890-934b-4cbe-9edf-89f5103075d2 tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] [instance: b779d183-89ae-4e4d-ae99-e514e145ed43] Powering off the VM {{(pid=61923) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1015.324062] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2ab03ddf-959f-4e1a-beff-dd9d153c04a3 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.330487] env[61923]: DEBUG oslo_vmware.api [None req-9ecd4890-934b-4cbe-9edf-89f5103075d2 tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Waiting for the task: (returnval){ [ 1015.330487] env[61923]: value = "task-1377935" [ 1015.330487] env[61923]: _type = "Task" [ 1015.330487] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1015.339966] env[61923]: DEBUG oslo_vmware.api [None req-9ecd4890-934b-4cbe-9edf-89f5103075d2 tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Task: {'id': task-1377935, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1015.372069] env[61923]: DEBUG nova.compute.manager [req-4648cad1-2899-4a76-8d83-0afa7fe12be8 req-14ea5125-1b8c-46b9-9ce4-4f4c62b3ebd3 service nova] [instance: 0c3788fc-6cce-4806-b4f8-84f84489fb93] Received event network-changed-504f15e6-4eee-4266-817f-d7935b5d0609 {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1015.372309] env[61923]: DEBUG nova.compute.manager [req-4648cad1-2899-4a76-8d83-0afa7fe12be8 req-14ea5125-1b8c-46b9-9ce4-4f4c62b3ebd3 service nova] [instance: 0c3788fc-6cce-4806-b4f8-84f84489fb93] Refreshing instance network info cache due to event network-changed-504f15e6-4eee-4266-817f-d7935b5d0609. {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1015.372577] env[61923]: DEBUG oslo_concurrency.lockutils [req-4648cad1-2899-4a76-8d83-0afa7fe12be8 req-14ea5125-1b8c-46b9-9ce4-4f4c62b3ebd3 service nova] Acquiring lock "refresh_cache-0c3788fc-6cce-4806-b4f8-84f84489fb93" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1015.372729] env[61923]: DEBUG oslo_concurrency.lockutils [req-4648cad1-2899-4a76-8d83-0afa7fe12be8 req-14ea5125-1b8c-46b9-9ce4-4f4c62b3ebd3 service nova] Acquired lock "refresh_cache-0c3788fc-6cce-4806-b4f8-84f84489fb93" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1015.372909] env[61923]: DEBUG nova.network.neutron [req-4648cad1-2899-4a76-8d83-0afa7fe12be8 req-14ea5125-1b8c-46b9-9ce4-4f4c62b3ebd3 service nova] [instance: 0c3788fc-6cce-4806-b4f8-84f84489fb93] Refreshing network info cache for port 504f15e6-4eee-4266-817f-d7935b5d0609 {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1015.630961] env[61923]: DEBUG oslo_vmware.api [None req-ca0212d9-f46f-4fe8-98d2-915ce325cd19 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Task: {'id': task-1377932, 'name': CreateSnapshot_Task, 'duration_secs': 0.376704} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1015.631272] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-ca0212d9-f46f-4fe8-98d2-915ce325cd19 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: a2f97947-6506-40b3-a24a-4ac03b88c5f5] Created Snapshot of the VM instance {{(pid=61923) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 1015.631975] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7936838d-a8e6-4821-aa2e-49e3d9a05aa7 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.663294] env[61923]: DEBUG oslo_vmware.api [-] Task: {'id': task-1377934, 'name': CreateVM_Task, 'duration_secs': 0.266219} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1015.663461] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0c3788fc-6cce-4806-b4f8-84f84489fb93] Created VM on the ESX host {{(pid=61923) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1015.664657] env[61923]: DEBUG oslo_concurrency.lockutils [None req-e2375eee-5ef0-47c4-8309-7404fe8cdfd1 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1015.664657] env[61923]: DEBUG oslo_concurrency.lockutils [None req-e2375eee-5ef0-47c4-8309-7404fe8cdfd1 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1015.664657] env[61923]: DEBUG oslo_concurrency.lockutils [None req-e2375eee-5ef0-47c4-8309-7404fe8cdfd1 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1015.664880] env[61923]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e20067c6-83fc-4c86-a4bc-ffad0c6dd477 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.668657] env[61923]: DEBUG oslo_vmware.api [None req-e2375eee-5ef0-47c4-8309-7404fe8cdfd1 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Waiting for the task: (returnval){ [ 1015.668657] env[61923]: value = "session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52d21ad2-750e-4a14-1083-ff6552461d63" [ 1015.668657] env[61923]: _type = "Task" [ 1015.668657] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1015.676292] env[61923]: DEBUG oslo_vmware.api [None req-e2375eee-5ef0-47c4-8309-7404fe8cdfd1 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52d21ad2-750e-4a14-1083-ff6552461d63, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1015.839729] env[61923]: DEBUG oslo_vmware.api [None req-9ecd4890-934b-4cbe-9edf-89f5103075d2 tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Task: {'id': task-1377935, 'name': PowerOffVM_Task, 'duration_secs': 0.190622} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1015.840036] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-9ecd4890-934b-4cbe-9edf-89f5103075d2 tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] [instance: b779d183-89ae-4e4d-ae99-e514e145ed43] Powered off the VM {{(pid=61923) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1015.840236] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-9ecd4890-934b-4cbe-9edf-89f5103075d2 tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] [instance: b779d183-89ae-4e4d-ae99-e514e145ed43] Unregistering the VM {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1015.840484] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-7447ffcc-c622-49d3-a4e9-8e3c2d541524 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.901397] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-9ecd4890-934b-4cbe-9edf-89f5103075d2 tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] [instance: b779d183-89ae-4e4d-ae99-e514e145ed43] Unregistered the VM {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1015.901397] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-9ecd4890-934b-4cbe-9edf-89f5103075d2 tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] [instance: b779d183-89ae-4e4d-ae99-e514e145ed43] Deleting contents of the VM from datastore datastore2 {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1015.901626] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-9ecd4890-934b-4cbe-9edf-89f5103075d2 tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Deleting the datastore file [datastore2] b779d183-89ae-4e4d-ae99-e514e145ed43 {{(pid=61923) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1015.901835] env[61923]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-354ac9fe-4ff8-4ef4-9c8c-5e7fd029718c {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.907353] env[61923]: DEBUG oslo_vmware.api [None req-9ecd4890-934b-4cbe-9edf-89f5103075d2 tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Waiting for the task: (returnval){ [ 1015.907353] env[61923]: value = "task-1377937" [ 1015.907353] env[61923]: _type = "Task" [ 1015.907353] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1015.916490] env[61923]: DEBUG oslo_vmware.api [None req-9ecd4890-934b-4cbe-9edf-89f5103075d2 tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Task: {'id': task-1377937, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1016.063588] env[61923]: DEBUG nova.network.neutron [req-4648cad1-2899-4a76-8d83-0afa7fe12be8 req-14ea5125-1b8c-46b9-9ce4-4f4c62b3ebd3 service nova] [instance: 0c3788fc-6cce-4806-b4f8-84f84489fb93] Updated VIF entry in instance network info cache for port 504f15e6-4eee-4266-817f-d7935b5d0609. {{(pid=61923) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1016.063986] env[61923]: DEBUG nova.network.neutron [req-4648cad1-2899-4a76-8d83-0afa7fe12be8 req-14ea5125-1b8c-46b9-9ce4-4f4c62b3ebd3 service nova] [instance: 0c3788fc-6cce-4806-b4f8-84f84489fb93] Updating instance_info_cache with network_info: [{"id": "504f15e6-4eee-4266-817f-d7935b5d0609", "address": "fa:16:3e:37:68:89", "network": {"id": "77aec6ca-d1be-4b39-9d51-08853a1183a1", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1813085414-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5a4db4f8aa194a219e80c538cc715d43", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "56b944d8-803d-43f2-945d-0f334ee4ea1c", "external-id": "nsx-vlan-transportzone-799", "segmentation_id": 799, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap504f15e6-4e", "ovs_interfaceid": "504f15e6-4eee-4266-817f-d7935b5d0609", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1016.148734] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-ca0212d9-f46f-4fe8-98d2-915ce325cd19 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: a2f97947-6506-40b3-a24a-4ac03b88c5f5] Creating linked-clone VM from snapshot {{(pid=61923) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 1016.149106] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-c52b3ed8-21e6-44fc-ac42-49f92aeb3d8e {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.157473] env[61923]: DEBUG oslo_vmware.api [None req-ca0212d9-f46f-4fe8-98d2-915ce325cd19 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Waiting for the task: (returnval){ [ 1016.157473] env[61923]: value = "task-1377938" [ 1016.157473] env[61923]: _type = "Task" [ 1016.157473] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1016.164774] env[61923]: DEBUG oslo_vmware.api [None req-ca0212d9-f46f-4fe8-98d2-915ce325cd19 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Task: {'id': task-1377938, 'name': CloneVM_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1016.176236] env[61923]: DEBUG oslo_vmware.api [None req-e2375eee-5ef0-47c4-8309-7404fe8cdfd1 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52d21ad2-750e-4a14-1083-ff6552461d63, 'name': SearchDatastore_Task, 'duration_secs': 0.009034} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1016.176501] env[61923]: DEBUG oslo_concurrency.lockutils [None req-e2375eee-5ef0-47c4-8309-7404fe8cdfd1 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1016.176724] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-e2375eee-5ef0-47c4-8309-7404fe8cdfd1 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: 0c3788fc-6cce-4806-b4f8-84f84489fb93] Processing image 0f153f63-ae0a-45b1-b7f5-7f9b673c947f {{(pid=61923) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1016.176951] env[61923]: DEBUG oslo_concurrency.lockutils [None req-e2375eee-5ef0-47c4-8309-7404fe8cdfd1 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1016.177112] env[61923]: DEBUG oslo_concurrency.lockutils [None req-e2375eee-5ef0-47c4-8309-7404fe8cdfd1 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1016.177290] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-e2375eee-5ef0-47c4-8309-7404fe8cdfd1 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61923) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1016.177513] env[61923]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-958d870f-42cb-42e8-8b2e-191965ac2443 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.184380] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-e2375eee-5ef0-47c4-8309-7404fe8cdfd1 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61923) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1016.184551] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-e2375eee-5ef0-47c4-8309-7404fe8cdfd1 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61923) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1016.185192] env[61923]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-32d119b6-c938-41ca-a254-731a5dd05cf8 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.189503] env[61923]: DEBUG oslo_vmware.api [None req-e2375eee-5ef0-47c4-8309-7404fe8cdfd1 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Waiting for the task: (returnval){ [ 1016.189503] env[61923]: value = "session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]5247e7fe-dacd-c850-e864-02de9beeaecd" [ 1016.189503] env[61923]: _type = "Task" [ 1016.189503] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1016.196314] env[61923]: DEBUG oslo_vmware.api [None req-e2375eee-5ef0-47c4-8309-7404fe8cdfd1 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]5247e7fe-dacd-c850-e864-02de9beeaecd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1016.416989] env[61923]: DEBUG oslo_vmware.api [None req-9ecd4890-934b-4cbe-9edf-89f5103075d2 tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Task: {'id': task-1377937, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.144181} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1016.417203] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-9ecd4890-934b-4cbe-9edf-89f5103075d2 tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Deleted the datastore file {{(pid=61923) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1016.417379] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-9ecd4890-934b-4cbe-9edf-89f5103075d2 tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] [instance: b779d183-89ae-4e4d-ae99-e514e145ed43] Deleted contents of the VM from datastore datastore2 {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1016.417560] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-9ecd4890-934b-4cbe-9edf-89f5103075d2 tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] [instance: b779d183-89ae-4e4d-ae99-e514e145ed43] Instance destroyed {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1016.417734] env[61923]: INFO nova.compute.manager [None req-9ecd4890-934b-4cbe-9edf-89f5103075d2 tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] [instance: b779d183-89ae-4e4d-ae99-e514e145ed43] Took 1.10 seconds to destroy the instance on the hypervisor. [ 1016.417978] env[61923]: DEBUG oslo.service.loopingcall [None req-9ecd4890-934b-4cbe-9edf-89f5103075d2 tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61923) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1016.418186] env[61923]: DEBUG nova.compute.manager [-] [instance: b779d183-89ae-4e4d-ae99-e514e145ed43] Deallocating network for instance {{(pid=61923) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1016.418281] env[61923]: DEBUG nova.network.neutron [-] [instance: b779d183-89ae-4e4d-ae99-e514e145ed43] deallocate_for_instance() {{(pid=61923) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1016.566984] env[61923]: DEBUG oslo_concurrency.lockutils [req-4648cad1-2899-4a76-8d83-0afa7fe12be8 req-14ea5125-1b8c-46b9-9ce4-4f4c62b3ebd3 service nova] Releasing lock "refresh_cache-0c3788fc-6cce-4806-b4f8-84f84489fb93" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1016.666983] env[61923]: DEBUG oslo_vmware.api [None req-ca0212d9-f46f-4fe8-98d2-915ce325cd19 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Task: {'id': task-1377938, 'name': CloneVM_Task} progress is 94%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1016.699635] env[61923]: DEBUG oslo_vmware.api [None req-e2375eee-5ef0-47c4-8309-7404fe8cdfd1 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]5247e7fe-dacd-c850-e864-02de9beeaecd, 'name': SearchDatastore_Task, 'duration_secs': 0.008905} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1016.700516] env[61923]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1177154c-92bd-4a04-8c99-e80c3a178fd8 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.705857] env[61923]: DEBUG oslo_vmware.api [None req-e2375eee-5ef0-47c4-8309-7404fe8cdfd1 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Waiting for the task: (returnval){ [ 1016.705857] env[61923]: value = "session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52c203c9-6933-15f6-56f4-ebec7229746b" [ 1016.705857] env[61923]: _type = "Task" [ 1016.705857] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1016.713549] env[61923]: DEBUG oslo_vmware.api [None req-e2375eee-5ef0-47c4-8309-7404fe8cdfd1 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52c203c9-6933-15f6-56f4-ebec7229746b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1017.168793] env[61923]: DEBUG oslo_vmware.api [None req-ca0212d9-f46f-4fe8-98d2-915ce325cd19 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Task: {'id': task-1377938, 'name': CloneVM_Task} progress is 95%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1017.214990] env[61923]: DEBUG oslo_vmware.api [None req-e2375eee-5ef0-47c4-8309-7404fe8cdfd1 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52c203c9-6933-15f6-56f4-ebec7229746b, 'name': SearchDatastore_Task, 'duration_secs': 0.009003} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1017.215290] env[61923]: DEBUG oslo_concurrency.lockutils [None req-e2375eee-5ef0-47c4-8309-7404fe8cdfd1 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1017.215552] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-e2375eee-5ef0-47c4-8309-7404fe8cdfd1 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk to [datastore1] 0c3788fc-6cce-4806-b4f8-84f84489fb93/0c3788fc-6cce-4806-b4f8-84f84489fb93.vmdk {{(pid=61923) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1017.215808] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-410f701e-8fff-40cf-8baa-20bc3d8cd303 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.224496] env[61923]: DEBUG oslo_vmware.api [None req-e2375eee-5ef0-47c4-8309-7404fe8cdfd1 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Waiting for the task: (returnval){ [ 1017.224496] env[61923]: value = "task-1377939" [ 1017.224496] env[61923]: _type = "Task" [ 1017.224496] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1017.231519] env[61923]: DEBUG oslo_vmware.api [None req-e2375eee-5ef0-47c4-8309-7404fe8cdfd1 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Task: {'id': task-1377939, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1017.357174] env[61923]: DEBUG nova.network.neutron [-] [instance: b779d183-89ae-4e4d-ae99-e514e145ed43] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1017.399988] env[61923]: DEBUG nova.compute.manager [req-4bdabaa3-1ca5-4cf0-96e3-246c9158a0ab req-d57f0d5c-54b5-4da0-8c09-84e74c4e9d83 service nova] [instance: b779d183-89ae-4e4d-ae99-e514e145ed43] Received event network-vif-deleted-d49a3010-c84a-4bef-beed-9ca4aa7d5dc2 {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1017.671630] env[61923]: DEBUG oslo_vmware.api [None req-ca0212d9-f46f-4fe8-98d2-915ce325cd19 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Task: {'id': task-1377938, 'name': CloneVM_Task, 'duration_secs': 1.346379} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1017.671996] env[61923]: INFO nova.virt.vmwareapi.vmops [None req-ca0212d9-f46f-4fe8-98d2-915ce325cd19 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: a2f97947-6506-40b3-a24a-4ac03b88c5f5] Created linked-clone VM from snapshot [ 1017.672995] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a169690-299a-4ec5-af2f-2cb6f98ec5f2 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.683819] env[61923]: DEBUG nova.virt.vmwareapi.images [None req-ca0212d9-f46f-4fe8-98d2-915ce325cd19 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: a2f97947-6506-40b3-a24a-4ac03b88c5f5] Uploading image a57425c8-c766-4a16-9bbd-41d100681a0e {{(pid=61923) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 1017.709578] env[61923]: DEBUG oslo_vmware.rw_handles [None req-ca0212d9-f46f-4fe8-98d2-915ce325cd19 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 1017.709578] env[61923]: value = "vm-292745" [ 1017.709578] env[61923]: _type = "VirtualMachine" [ 1017.709578] env[61923]: }. {{(pid=61923) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 1017.709882] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-22ed362f-2646-420e-ba1a-12dcba9f482c {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.716895] env[61923]: DEBUG oslo_vmware.rw_handles [None req-ca0212d9-f46f-4fe8-98d2-915ce325cd19 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Lease: (returnval){ [ 1017.716895] env[61923]: value = "session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52f48c2c-476d-fdc7-30b9-08618b2aa1a0" [ 1017.716895] env[61923]: _type = "HttpNfcLease" [ 1017.716895] env[61923]: } obtained for exporting VM: (result){ [ 1017.716895] env[61923]: value = "vm-292745" [ 1017.716895] env[61923]: _type = "VirtualMachine" [ 1017.716895] env[61923]: }. {{(pid=61923) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 1017.717207] env[61923]: DEBUG oslo_vmware.api [None req-ca0212d9-f46f-4fe8-98d2-915ce325cd19 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Waiting for the lease: (returnval){ [ 1017.717207] env[61923]: value = "session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52f48c2c-476d-fdc7-30b9-08618b2aa1a0" [ 1017.717207] env[61923]: _type = "HttpNfcLease" [ 1017.717207] env[61923]: } to be ready. {{(pid=61923) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1017.723051] env[61923]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1017.723051] env[61923]: value = "session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52f48c2c-476d-fdc7-30b9-08618b2aa1a0" [ 1017.723051] env[61923]: _type = "HttpNfcLease" [ 1017.723051] env[61923]: } is initializing. {{(pid=61923) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1017.733364] env[61923]: DEBUG oslo_vmware.api [None req-e2375eee-5ef0-47c4-8309-7404fe8cdfd1 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Task: {'id': task-1377939, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.505931} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1017.733576] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-e2375eee-5ef0-47c4-8309-7404fe8cdfd1 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk to [datastore1] 0c3788fc-6cce-4806-b4f8-84f84489fb93/0c3788fc-6cce-4806-b4f8-84f84489fb93.vmdk {{(pid=61923) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1017.733797] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-e2375eee-5ef0-47c4-8309-7404fe8cdfd1 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: 0c3788fc-6cce-4806-b4f8-84f84489fb93] Extending root virtual disk to 1048576 {{(pid=61923) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1017.734133] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-8aee113e-308a-46c8-9090-4cfb54ebcbf4 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.739984] env[61923]: DEBUG oslo_vmware.api [None req-e2375eee-5ef0-47c4-8309-7404fe8cdfd1 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Waiting for the task: (returnval){ [ 1017.739984] env[61923]: value = "task-1377941" [ 1017.739984] env[61923]: _type = "Task" [ 1017.739984] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1017.747735] env[61923]: DEBUG oslo_vmware.api [None req-e2375eee-5ef0-47c4-8309-7404fe8cdfd1 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Task: {'id': task-1377941, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1017.860614] env[61923]: INFO nova.compute.manager [-] [instance: b779d183-89ae-4e4d-ae99-e514e145ed43] Took 1.44 seconds to deallocate network for instance. [ 1018.225452] env[61923]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1018.225452] env[61923]: value = "session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52f48c2c-476d-fdc7-30b9-08618b2aa1a0" [ 1018.225452] env[61923]: _type = "HttpNfcLease" [ 1018.225452] env[61923]: } is ready. {{(pid=61923) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1018.225896] env[61923]: DEBUG oslo_vmware.rw_handles [None req-ca0212d9-f46f-4fe8-98d2-915ce325cd19 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1018.225896] env[61923]: value = "session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52f48c2c-476d-fdc7-30b9-08618b2aa1a0" [ 1018.225896] env[61923]: _type = "HttpNfcLease" [ 1018.225896] env[61923]: }. {{(pid=61923) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 1018.226500] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-697c8b27-0f09-459f-b3ea-945ea9ff853b {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.233572] env[61923]: DEBUG oslo_vmware.rw_handles [None req-ca0212d9-f46f-4fe8-98d2-915ce325cd19 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/526a7206-e23a-8499-a5df-733b976b75a4/disk-0.vmdk from lease info. {{(pid=61923) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1018.233748] env[61923]: DEBUG oslo_vmware.rw_handles [None req-ca0212d9-f46f-4fe8-98d2-915ce325cd19 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Opening URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/526a7206-e23a-8499-a5df-733b976b75a4/disk-0.vmdk for reading. {{(pid=61923) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 1018.297252] env[61923]: DEBUG oslo_vmware.api [None req-e2375eee-5ef0-47c4-8309-7404fe8cdfd1 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Task: {'id': task-1377941, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.059678} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1018.297519] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-e2375eee-5ef0-47c4-8309-7404fe8cdfd1 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: 0c3788fc-6cce-4806-b4f8-84f84489fb93] Extended root virtual disk {{(pid=61923) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1018.298279] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f73db30-f92c-4c78-91f1-c088fcb9d673 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.319644] env[61923]: DEBUG nova.virt.vmwareapi.volumeops [None req-e2375eee-5ef0-47c4-8309-7404fe8cdfd1 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: 0c3788fc-6cce-4806-b4f8-84f84489fb93] Reconfiguring VM instance instance-0000005f to attach disk [datastore1] 0c3788fc-6cce-4806-b4f8-84f84489fb93/0c3788fc-6cce-4806-b4f8-84f84489fb93.vmdk or device None with type sparse {{(pid=61923) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1018.319885] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-89eef93a-2053-44c4-95d9-6f1d5c53bfc9 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.341481] env[61923]: DEBUG oslo_vmware.api [None req-e2375eee-5ef0-47c4-8309-7404fe8cdfd1 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Waiting for the task: (returnval){ [ 1018.341481] env[61923]: value = "task-1377942" [ 1018.341481] env[61923]: _type = "Task" [ 1018.341481] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1018.348979] env[61923]: DEBUG oslo_vmware.api [None req-e2375eee-5ef0-47c4-8309-7404fe8cdfd1 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Task: {'id': task-1377942, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1018.367090] env[61923]: DEBUG oslo_concurrency.lockutils [None req-9ecd4890-934b-4cbe-9edf-89f5103075d2 tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1018.367350] env[61923]: DEBUG oslo_concurrency.lockutils [None req-9ecd4890-934b-4cbe-9edf-89f5103075d2 tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1018.367571] env[61923]: DEBUG nova.objects.instance [None req-9ecd4890-934b-4cbe-9edf-89f5103075d2 tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Lazy-loading 'resources' on Instance uuid b779d183-89ae-4e4d-ae99-e514e145ed43 {{(pid=61923) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1018.417741] env[61923]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-15c73174-9656-4458-94eb-6d93df1d5bac {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.852117] env[61923]: DEBUG oslo_vmware.api [None req-e2375eee-5ef0-47c4-8309-7404fe8cdfd1 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Task: {'id': task-1377942, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1018.929778] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c7f1417-75ad-43d8-9a62-fe5a8ae69dc0 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.937377] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39887331-9437-48b6-84c4-27d94efa528c {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.969828] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc556117-4b50-417e-9b55-cebf96f2a9eb {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.977324] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5692a73-0741-4c6e-b4b0-913c8697cd0d {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.991086] env[61923]: DEBUG nova.compute.provider_tree [None req-9ecd4890-934b-4cbe-9edf-89f5103075d2 tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1019.352259] env[61923]: DEBUG oslo_vmware.api [None req-e2375eee-5ef0-47c4-8309-7404fe8cdfd1 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Task: {'id': task-1377942, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1019.496243] env[61923]: DEBUG nova.scheduler.client.report [None req-9ecd4890-934b-4cbe-9edf-89f5103075d2 tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1019.852847] env[61923]: DEBUG oslo_vmware.api [None req-e2375eee-5ef0-47c4-8309-7404fe8cdfd1 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Task: {'id': task-1377942, 'name': ReconfigVM_Task, 'duration_secs': 1.35028} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1019.853212] env[61923]: DEBUG nova.virt.vmwareapi.volumeops [None req-e2375eee-5ef0-47c4-8309-7404fe8cdfd1 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: 0c3788fc-6cce-4806-b4f8-84f84489fb93] Reconfigured VM instance instance-0000005f to attach disk [datastore1] 0c3788fc-6cce-4806-b4f8-84f84489fb93/0c3788fc-6cce-4806-b4f8-84f84489fb93.vmdk or device None with type sparse {{(pid=61923) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1019.853918] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-d2db1380-305d-4e6e-81ec-98cac82e14df {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.860609] env[61923]: DEBUG oslo_vmware.api [None req-e2375eee-5ef0-47c4-8309-7404fe8cdfd1 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Waiting for the task: (returnval){ [ 1019.860609] env[61923]: value = "task-1377943" [ 1019.860609] env[61923]: _type = "Task" [ 1019.860609] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1019.868284] env[61923]: DEBUG oslo_vmware.api [None req-e2375eee-5ef0-47c4-8309-7404fe8cdfd1 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Task: {'id': task-1377943, 'name': Rename_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1020.002649] env[61923]: DEBUG oslo_concurrency.lockutils [None req-9ecd4890-934b-4cbe-9edf-89f5103075d2 tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.635s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1020.022990] env[61923]: INFO nova.scheduler.client.report [None req-9ecd4890-934b-4cbe-9edf-89f5103075d2 tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Deleted allocations for instance b779d183-89ae-4e4d-ae99-e514e145ed43 [ 1020.370773] env[61923]: DEBUG oslo_vmware.api [None req-e2375eee-5ef0-47c4-8309-7404fe8cdfd1 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Task: {'id': task-1377943, 'name': Rename_Task, 'duration_secs': 0.186767} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1020.371128] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-e2375eee-5ef0-47c4-8309-7404fe8cdfd1 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: 0c3788fc-6cce-4806-b4f8-84f84489fb93] Powering on the VM {{(pid=61923) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1020.371392] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-25966d91-5b7b-4c70-a2ab-19db0d576c0e {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.377840] env[61923]: DEBUG oslo_vmware.api [None req-e2375eee-5ef0-47c4-8309-7404fe8cdfd1 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Waiting for the task: (returnval){ [ 1020.377840] env[61923]: value = "task-1377944" [ 1020.377840] env[61923]: _type = "Task" [ 1020.377840] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1020.385494] env[61923]: DEBUG oslo_vmware.api [None req-e2375eee-5ef0-47c4-8309-7404fe8cdfd1 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Task: {'id': task-1377944, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1020.531603] env[61923]: DEBUG oslo_concurrency.lockutils [None req-9ecd4890-934b-4cbe-9edf-89f5103075d2 tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Lock "b779d183-89ae-4e4d-ae99-e514e145ed43" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.222s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1020.891587] env[61923]: DEBUG oslo_vmware.api [None req-e2375eee-5ef0-47c4-8309-7404fe8cdfd1 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Task: {'id': task-1377944, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1021.392956] env[61923]: DEBUG oslo_vmware.api [None req-e2375eee-5ef0-47c4-8309-7404fe8cdfd1 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Task: {'id': task-1377944, 'name': PowerOnVM_Task, 'duration_secs': 0.912161} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1021.393414] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-e2375eee-5ef0-47c4-8309-7404fe8cdfd1 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: 0c3788fc-6cce-4806-b4f8-84f84489fb93] Powered on the VM {{(pid=61923) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1021.393624] env[61923]: INFO nova.compute.manager [None req-e2375eee-5ef0-47c4-8309-7404fe8cdfd1 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: 0c3788fc-6cce-4806-b4f8-84f84489fb93] Took 8.18 seconds to spawn the instance on the hypervisor. [ 1021.393828] env[61923]: DEBUG nova.compute.manager [None req-e2375eee-5ef0-47c4-8309-7404fe8cdfd1 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: 0c3788fc-6cce-4806-b4f8-84f84489fb93] Checking state {{(pid=61923) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1021.396466] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e36259c-636c-4210-808d-91cb79b91f67 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.914978] env[61923]: INFO nova.compute.manager [None req-e2375eee-5ef0-47c4-8309-7404fe8cdfd1 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: 0c3788fc-6cce-4806-b4f8-84f84489fb93] Took 12.87 seconds to build instance. [ 1022.417429] env[61923]: DEBUG oslo_concurrency.lockutils [None req-e2375eee-5ef0-47c4-8309-7404fe8cdfd1 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Lock "0c3788fc-6cce-4806-b4f8-84f84489fb93" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.378s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1022.462860] env[61923]: DEBUG oslo_concurrency.lockutils [None req-d17d2b02-87ac-45f0-981e-237fcddcf45c tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Acquiring lock "35bdbd45-5838-4b09-a232-7f4a02322ad7" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1022.463114] env[61923]: DEBUG oslo_concurrency.lockutils [None req-d17d2b02-87ac-45f0-981e-237fcddcf45c tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Lock "35bdbd45-5838-4b09-a232-7f4a02322ad7" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1022.500254] env[61923]: INFO nova.compute.manager [None req-c35e31ab-0af5-4818-ac4e-4115156ea899 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: 0c3788fc-6cce-4806-b4f8-84f84489fb93] Rebuilding instance [ 1022.539652] env[61923]: DEBUG nova.compute.manager [None req-c35e31ab-0af5-4818-ac4e-4115156ea899 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: 0c3788fc-6cce-4806-b4f8-84f84489fb93] Checking state {{(pid=61923) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1022.540664] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-672add87-cbba-4cfd-9f1d-2d2bd914747d {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.965021] env[61923]: DEBUG nova.compute.manager [None req-d17d2b02-87ac-45f0-981e-237fcddcf45c tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] [instance: 35bdbd45-5838-4b09-a232-7f4a02322ad7] Starting instance... {{(pid=61923) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1023.051206] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-c35e31ab-0af5-4818-ac4e-4115156ea899 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: 0c3788fc-6cce-4806-b4f8-84f84489fb93] Powering off the VM {{(pid=61923) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1023.051540] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-36224cf0-f530-4d90-beca-5514a3afccef {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.059139] env[61923]: DEBUG oslo_vmware.api [None req-c35e31ab-0af5-4818-ac4e-4115156ea899 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Waiting for the task: (returnval){ [ 1023.059139] env[61923]: value = "task-1377945" [ 1023.059139] env[61923]: _type = "Task" [ 1023.059139] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1023.067629] env[61923]: DEBUG oslo_vmware.api [None req-c35e31ab-0af5-4818-ac4e-4115156ea899 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Task: {'id': task-1377945, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1023.492241] env[61923]: DEBUG oslo_concurrency.lockutils [None req-d17d2b02-87ac-45f0-981e-237fcddcf45c tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1023.492627] env[61923]: DEBUG oslo_concurrency.lockutils [None req-d17d2b02-87ac-45f0-981e-237fcddcf45c tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.001s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1023.494753] env[61923]: INFO nova.compute.claims [None req-d17d2b02-87ac-45f0-981e-237fcddcf45c tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] [instance: 35bdbd45-5838-4b09-a232-7f4a02322ad7] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1023.571370] env[61923]: DEBUG oslo_vmware.api [None req-c35e31ab-0af5-4818-ac4e-4115156ea899 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Task: {'id': task-1377945, 'name': PowerOffVM_Task, 'duration_secs': 0.204977} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1023.571690] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-c35e31ab-0af5-4818-ac4e-4115156ea899 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: 0c3788fc-6cce-4806-b4f8-84f84489fb93] Powered off the VM {{(pid=61923) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1023.571923] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-c35e31ab-0af5-4818-ac4e-4115156ea899 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: 0c3788fc-6cce-4806-b4f8-84f84489fb93] Destroying instance {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1023.572787] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1785b77e-35b3-4fd8-8ac4-fb478762e1ed {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.579587] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-c35e31ab-0af5-4818-ac4e-4115156ea899 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: 0c3788fc-6cce-4806-b4f8-84f84489fb93] Unregistering the VM {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1023.579847] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9b486dc4-a306-4537-9d9f-d3a228245594 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.657497] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-c35e31ab-0af5-4818-ac4e-4115156ea899 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: 0c3788fc-6cce-4806-b4f8-84f84489fb93] Unregistered the VM {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1023.657988] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-c35e31ab-0af5-4818-ac4e-4115156ea899 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: 0c3788fc-6cce-4806-b4f8-84f84489fb93] Deleting contents of the VM from datastore datastore1 {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1023.658358] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-c35e31ab-0af5-4818-ac4e-4115156ea899 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Deleting the datastore file [datastore1] 0c3788fc-6cce-4806-b4f8-84f84489fb93 {{(pid=61923) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1023.658755] env[61923]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-60d896e8-d08f-4a9c-a0b8-ad9d2e5f42d5 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.667423] env[61923]: DEBUG oslo_vmware.api [None req-c35e31ab-0af5-4818-ac4e-4115156ea899 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Waiting for the task: (returnval){ [ 1023.667423] env[61923]: value = "task-1377947" [ 1023.667423] env[61923]: _type = "Task" [ 1023.667423] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1023.679631] env[61923]: DEBUG oslo_vmware.api [None req-c35e31ab-0af5-4818-ac4e-4115156ea899 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Task: {'id': task-1377947, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1024.178303] env[61923]: DEBUG oslo_vmware.api [None req-c35e31ab-0af5-4818-ac4e-4115156ea899 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Task: {'id': task-1377947, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.187426} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1024.178599] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-c35e31ab-0af5-4818-ac4e-4115156ea899 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Deleted the datastore file {{(pid=61923) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1024.178798] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-c35e31ab-0af5-4818-ac4e-4115156ea899 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: 0c3788fc-6cce-4806-b4f8-84f84489fb93] Deleted contents of the VM from datastore datastore1 {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1024.178991] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-c35e31ab-0af5-4818-ac4e-4115156ea899 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: 0c3788fc-6cce-4806-b4f8-84f84489fb93] Instance destroyed {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1024.557787] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77ea67df-dae0-4ab1-b266-33943d6901a7 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.565877] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4677c515-7650-425e-86f2-ce565ff8218a {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.596386] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9c7ffea-edd3-4b8a-97be-b14de1d7943a {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.604302] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d2b3451-e66a-416d-85f4-ae267bf3ace0 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.617717] env[61923]: DEBUG nova.compute.provider_tree [None req-d17d2b02-87ac-45f0-981e-237fcddcf45c tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1025.120850] env[61923]: DEBUG nova.scheduler.client.report [None req-d17d2b02-87ac-45f0-981e-237fcddcf45c tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1025.208944] env[61923]: DEBUG nova.virt.hardware [None req-c35e31ab-0af5-4818-ac4e-4115156ea899 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-29T20:07:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-29T20:07:35Z,direct_url=,disk_format='vmdk',id=0f153f63-ae0a-45b1-b7f5-7f9b673c947f,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='4e7b5e3b3c3443c8bd5c70f8a15739f5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-29T20:07:36Z,virtual_size=,visibility=), allow threads: False {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1025.209234] env[61923]: DEBUG nova.virt.hardware [None req-c35e31ab-0af5-4818-ac4e-4115156ea899 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Flavor limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1025.209409] env[61923]: DEBUG nova.virt.hardware [None req-c35e31ab-0af5-4818-ac4e-4115156ea899 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Image limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1025.209600] env[61923]: DEBUG nova.virt.hardware [None req-c35e31ab-0af5-4818-ac4e-4115156ea899 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Flavor pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1025.209766] env[61923]: DEBUG nova.virt.hardware [None req-c35e31ab-0af5-4818-ac4e-4115156ea899 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Image pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1025.209921] env[61923]: DEBUG nova.virt.hardware [None req-c35e31ab-0af5-4818-ac4e-4115156ea899 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1025.210216] env[61923]: DEBUG nova.virt.hardware [None req-c35e31ab-0af5-4818-ac4e-4115156ea899 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1025.210408] env[61923]: DEBUG nova.virt.hardware [None req-c35e31ab-0af5-4818-ac4e-4115156ea899 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1025.210582] env[61923]: DEBUG nova.virt.hardware [None req-c35e31ab-0af5-4818-ac4e-4115156ea899 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Got 1 possible topologies {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1025.210746] env[61923]: DEBUG nova.virt.hardware [None req-c35e31ab-0af5-4818-ac4e-4115156ea899 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1025.210917] env[61923]: DEBUG nova.virt.hardware [None req-c35e31ab-0af5-4818-ac4e-4115156ea899 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1025.211857] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-308c3fe1-a343-4baf-ab4b-e33ca0685e5e {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.220332] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de729bc3-8596-409b-86cf-78beb56b1430 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.234457] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-c35e31ab-0af5-4818-ac4e-4115156ea899 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: 0c3788fc-6cce-4806-b4f8-84f84489fb93] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:37:68:89', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '56b944d8-803d-43f2-945d-0f334ee4ea1c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '504f15e6-4eee-4266-817f-d7935b5d0609', 'vif_model': 'vmxnet3'}] {{(pid=61923) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1025.242165] env[61923]: DEBUG oslo.service.loopingcall [None req-c35e31ab-0af5-4818-ac4e-4115156ea899 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61923) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1025.242444] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0c3788fc-6cce-4806-b4f8-84f84489fb93] Creating VM on the ESX host {{(pid=61923) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1025.242672] env[61923]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-be11d483-a854-41c6-8eb5-e25479f07649 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.262337] env[61923]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1025.262337] env[61923]: value = "task-1377948" [ 1025.262337] env[61923]: _type = "Task" [ 1025.262337] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1025.270140] env[61923]: DEBUG oslo_vmware.api [-] Task: {'id': task-1377948, 'name': CreateVM_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.626445] env[61923]: DEBUG oslo_concurrency.lockutils [None req-d17d2b02-87ac-45f0-981e-237fcddcf45c tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.134s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1025.627037] env[61923]: DEBUG nova.compute.manager [None req-d17d2b02-87ac-45f0-981e-237fcddcf45c tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] [instance: 35bdbd45-5838-4b09-a232-7f4a02322ad7] Start building networks asynchronously for instance. {{(pid=61923) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1025.771867] env[61923]: DEBUG oslo_vmware.api [-] Task: {'id': task-1377948, 'name': CreateVM_Task, 'duration_secs': 0.428133} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1025.772071] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0c3788fc-6cce-4806-b4f8-84f84489fb93] Created VM on the ESX host {{(pid=61923) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1025.772689] env[61923]: DEBUG oslo_concurrency.lockutils [None req-c35e31ab-0af5-4818-ac4e-4115156ea899 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1025.772881] env[61923]: DEBUG oslo_concurrency.lockutils [None req-c35e31ab-0af5-4818-ac4e-4115156ea899 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1025.773238] env[61923]: DEBUG oslo_concurrency.lockutils [None req-c35e31ab-0af5-4818-ac4e-4115156ea899 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1025.773516] env[61923]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-42761372-7597-4d01-9644-340891251026 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.777951] env[61923]: DEBUG oslo_vmware.api [None req-c35e31ab-0af5-4818-ac4e-4115156ea899 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Waiting for the task: (returnval){ [ 1025.777951] env[61923]: value = "session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52c3e21a-ba07-1d27-c67c-d45c9e9b07d8" [ 1025.777951] env[61923]: _type = "Task" [ 1025.777951] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1025.785353] env[61923]: DEBUG oslo_vmware.api [None req-c35e31ab-0af5-4818-ac4e-4115156ea899 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52c3e21a-ba07-1d27-c67c-d45c9e9b07d8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1026.132561] env[61923]: DEBUG nova.compute.utils [None req-d17d2b02-87ac-45f0-981e-237fcddcf45c tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Using /dev/sd instead of None {{(pid=61923) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1026.134233] env[61923]: DEBUG nova.compute.manager [None req-d17d2b02-87ac-45f0-981e-237fcddcf45c tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] [instance: 35bdbd45-5838-4b09-a232-7f4a02322ad7] Allocating IP information in the background. {{(pid=61923) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1026.134463] env[61923]: DEBUG nova.network.neutron [None req-d17d2b02-87ac-45f0-981e-237fcddcf45c tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] [instance: 35bdbd45-5838-4b09-a232-7f4a02322ad7] allocate_for_instance() {{(pid=61923) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1026.173729] env[61923]: DEBUG nova.policy [None req-d17d2b02-87ac-45f0-981e-237fcddcf45c tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ced04976ec034e09a0ad122827292901', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a13d827933fa4597984afdb91dbbdd39', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61923) authorize /opt/stack/nova/nova/policy.py:201}} [ 1026.289734] env[61923]: DEBUG oslo_vmware.api [None req-c35e31ab-0af5-4818-ac4e-4115156ea899 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52c3e21a-ba07-1d27-c67c-d45c9e9b07d8, 'name': SearchDatastore_Task, 'duration_secs': 0.010144} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1026.291063] env[61923]: DEBUG oslo_concurrency.lockutils [None req-c35e31ab-0af5-4818-ac4e-4115156ea899 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1026.291063] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-c35e31ab-0af5-4818-ac4e-4115156ea899 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: 0c3788fc-6cce-4806-b4f8-84f84489fb93] Processing image 0f153f63-ae0a-45b1-b7f5-7f9b673c947f {{(pid=61923) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1026.291063] env[61923]: DEBUG oslo_concurrency.lockutils [None req-c35e31ab-0af5-4818-ac4e-4115156ea899 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1026.291063] env[61923]: DEBUG oslo_concurrency.lockutils [None req-c35e31ab-0af5-4818-ac4e-4115156ea899 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1026.291418] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-c35e31ab-0af5-4818-ac4e-4115156ea899 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61923) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1026.291550] env[61923]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-7fc09877-8e3f-4ccd-82ca-71908dc48f24 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.299676] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-c35e31ab-0af5-4818-ac4e-4115156ea899 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61923) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1026.299933] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-c35e31ab-0af5-4818-ac4e-4115156ea899 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61923) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1026.300704] env[61923]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b4614ca2-a053-44cf-bf35-14f5fd3be65b {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.305755] env[61923]: DEBUG oslo_vmware.api [None req-c35e31ab-0af5-4818-ac4e-4115156ea899 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Waiting for the task: (returnval){ [ 1026.305755] env[61923]: value = "session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]5255db4a-e73c-f4ae-be45-1baae92f29f6" [ 1026.305755] env[61923]: _type = "Task" [ 1026.305755] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1026.313902] env[61923]: DEBUG oslo_vmware.api [None req-c35e31ab-0af5-4818-ac4e-4115156ea899 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]5255db4a-e73c-f4ae-be45-1baae92f29f6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1026.437899] env[61923]: DEBUG nova.network.neutron [None req-d17d2b02-87ac-45f0-981e-237fcddcf45c tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] [instance: 35bdbd45-5838-4b09-a232-7f4a02322ad7] Successfully created port: 42e3ee29-941c-43d0-a7f3-81fc7755f64b {{(pid=61923) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1026.637822] env[61923]: DEBUG nova.compute.manager [None req-d17d2b02-87ac-45f0-981e-237fcddcf45c tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] [instance: 35bdbd45-5838-4b09-a232-7f4a02322ad7] Start building block device mappings for instance. {{(pid=61923) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1026.818724] env[61923]: DEBUG oslo_vmware.api [None req-c35e31ab-0af5-4818-ac4e-4115156ea899 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]5255db4a-e73c-f4ae-be45-1baae92f29f6, 'name': SearchDatastore_Task, 'duration_secs': 0.008411} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1026.819528] env[61923]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f8683466-cf73-4cb2-a886-d3081d345be3 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.824444] env[61923]: DEBUG oslo_vmware.api [None req-c35e31ab-0af5-4818-ac4e-4115156ea899 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Waiting for the task: (returnval){ [ 1026.824444] env[61923]: value = "session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]5214db7a-2337-c01d-c50f-9a9988b83a2d" [ 1026.824444] env[61923]: _type = "Task" [ 1026.824444] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1026.833205] env[61923]: DEBUG oslo_vmware.api [None req-c35e31ab-0af5-4818-ac4e-4115156ea899 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]5214db7a-2337-c01d-c50f-9a9988b83a2d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1026.835117] env[61923]: DEBUG oslo_vmware.rw_handles [None req-ca0212d9-f46f-4fe8-98d2-915ce325cd19 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/526a7206-e23a-8499-a5df-733b976b75a4/disk-0.vmdk. {{(pid=61923) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1026.835867] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be5c88b2-a5f9-4e33-ae90-5ba003c02b07 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.841047] env[61923]: DEBUG oslo_vmware.rw_handles [None req-ca0212d9-f46f-4fe8-98d2-915ce325cd19 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/526a7206-e23a-8499-a5df-733b976b75a4/disk-0.vmdk is in state: ready. {{(pid=61923) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1026.841238] env[61923]: ERROR oslo_vmware.rw_handles [None req-ca0212d9-f46f-4fe8-98d2-915ce325cd19 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Aborting lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/526a7206-e23a-8499-a5df-733b976b75a4/disk-0.vmdk due to incomplete transfer. [ 1026.841454] env[61923]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-3c0e8673-936b-4499-91f2-0b89f36ae081 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.847263] env[61923]: DEBUG oslo_vmware.rw_handles [None req-ca0212d9-f46f-4fe8-98d2-915ce325cd19 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Closed VMDK read handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/526a7206-e23a-8499-a5df-733b976b75a4/disk-0.vmdk. {{(pid=61923) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 1026.847676] env[61923]: DEBUG nova.virt.vmwareapi.images [None req-ca0212d9-f46f-4fe8-98d2-915ce325cd19 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: a2f97947-6506-40b3-a24a-4ac03b88c5f5] Uploaded image a57425c8-c766-4a16-9bbd-41d100681a0e to the Glance image server {{(pid=61923) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 1026.849809] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-ca0212d9-f46f-4fe8-98d2-915ce325cd19 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: a2f97947-6506-40b3-a24a-4ac03b88c5f5] Destroying the VM {{(pid=61923) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 1026.850100] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-09eebb8a-bb08-41b9-84e0-49c385e43a5e {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.855698] env[61923]: DEBUG oslo_vmware.api [None req-ca0212d9-f46f-4fe8-98d2-915ce325cd19 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Waiting for the task: (returnval){ [ 1026.855698] env[61923]: value = "task-1377949" [ 1026.855698] env[61923]: _type = "Task" [ 1026.855698] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1026.863017] env[61923]: DEBUG oslo_vmware.api [None req-ca0212d9-f46f-4fe8-98d2-915ce325cd19 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Task: {'id': task-1377949, 'name': Destroy_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1027.335958] env[61923]: DEBUG oslo_vmware.api [None req-c35e31ab-0af5-4818-ac4e-4115156ea899 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]5214db7a-2337-c01d-c50f-9a9988b83a2d, 'name': SearchDatastore_Task, 'duration_secs': 0.009809} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1027.336245] env[61923]: DEBUG oslo_concurrency.lockutils [None req-c35e31ab-0af5-4818-ac4e-4115156ea899 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1027.336506] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-c35e31ab-0af5-4818-ac4e-4115156ea899 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk to [datastore2] 0c3788fc-6cce-4806-b4f8-84f84489fb93/0c3788fc-6cce-4806-b4f8-84f84489fb93.vmdk {{(pid=61923) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1027.336751] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c991405e-a869-4500-ae6d-e97e311ce93f {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.342770] env[61923]: DEBUG oslo_vmware.api [None req-c35e31ab-0af5-4818-ac4e-4115156ea899 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Waiting for the task: (returnval){ [ 1027.342770] env[61923]: value = "task-1377950" [ 1027.342770] env[61923]: _type = "Task" [ 1027.342770] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1027.349864] env[61923]: DEBUG oslo_vmware.api [None req-c35e31ab-0af5-4818-ac4e-4115156ea899 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Task: {'id': task-1377950, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1027.363225] env[61923]: DEBUG oslo_vmware.api [None req-ca0212d9-f46f-4fe8-98d2-915ce325cd19 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Task: {'id': task-1377949, 'name': Destroy_Task, 'duration_secs': 0.289545} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1027.363467] env[61923]: INFO nova.virt.vmwareapi.vm_util [None req-ca0212d9-f46f-4fe8-98d2-915ce325cd19 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: a2f97947-6506-40b3-a24a-4ac03b88c5f5] Destroyed the VM [ 1027.363701] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-ca0212d9-f46f-4fe8-98d2-915ce325cd19 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: a2f97947-6506-40b3-a24a-4ac03b88c5f5] Deleting Snapshot of the VM instance {{(pid=61923) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 1027.363918] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-f943a712-09cf-46ab-bddc-13c832896e43 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.369615] env[61923]: DEBUG oslo_vmware.api [None req-ca0212d9-f46f-4fe8-98d2-915ce325cd19 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Waiting for the task: (returnval){ [ 1027.369615] env[61923]: value = "task-1377951" [ 1027.369615] env[61923]: _type = "Task" [ 1027.369615] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1027.376557] env[61923]: DEBUG oslo_vmware.api [None req-ca0212d9-f46f-4fe8-98d2-915ce325cd19 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Task: {'id': task-1377951, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1027.648183] env[61923]: DEBUG nova.compute.manager [None req-d17d2b02-87ac-45f0-981e-237fcddcf45c tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] [instance: 35bdbd45-5838-4b09-a232-7f4a02322ad7] Start spawning the instance on the hypervisor. {{(pid=61923) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1027.675357] env[61923]: DEBUG nova.virt.hardware [None req-d17d2b02-87ac-45f0-981e-237fcddcf45c tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-29T20:07:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-29T20:07:35Z,direct_url=,disk_format='vmdk',id=0f153f63-ae0a-45b1-b7f5-7f9b673c947f,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='4e7b5e3b3c3443c8bd5c70f8a15739f5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-29T20:07:36Z,virtual_size=,visibility=), allow threads: False {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1027.675708] env[61923]: DEBUG nova.virt.hardware [None req-d17d2b02-87ac-45f0-981e-237fcddcf45c tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Flavor limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1027.675916] env[61923]: DEBUG nova.virt.hardware [None req-d17d2b02-87ac-45f0-981e-237fcddcf45c tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Image limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1027.676189] env[61923]: DEBUG nova.virt.hardware [None req-d17d2b02-87ac-45f0-981e-237fcddcf45c tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Flavor pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1027.676384] env[61923]: DEBUG nova.virt.hardware [None req-d17d2b02-87ac-45f0-981e-237fcddcf45c tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Image pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1027.676606] env[61923]: DEBUG nova.virt.hardware [None req-d17d2b02-87ac-45f0-981e-237fcddcf45c tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1027.676873] env[61923]: DEBUG nova.virt.hardware [None req-d17d2b02-87ac-45f0-981e-237fcddcf45c tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1027.677089] env[61923]: DEBUG nova.virt.hardware [None req-d17d2b02-87ac-45f0-981e-237fcddcf45c tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1027.677335] env[61923]: DEBUG nova.virt.hardware [None req-d17d2b02-87ac-45f0-981e-237fcddcf45c tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Got 1 possible topologies {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1027.677717] env[61923]: DEBUG nova.virt.hardware [None req-d17d2b02-87ac-45f0-981e-237fcddcf45c tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1027.677914] env[61923]: DEBUG nova.virt.hardware [None req-d17d2b02-87ac-45f0-981e-237fcddcf45c tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1027.678766] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a68f2aa-3011-4fc3-bd42-dd0ba10c3b40 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.687979] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c7c2d4e-7aaf-44d7-a0c3-e92e6b73ceca {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.852065] env[61923]: DEBUG oslo_vmware.api [None req-c35e31ab-0af5-4818-ac4e-4115156ea899 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Task: {'id': task-1377950, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.452871} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1027.852358] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-c35e31ab-0af5-4818-ac4e-4115156ea899 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk to [datastore2] 0c3788fc-6cce-4806-b4f8-84f84489fb93/0c3788fc-6cce-4806-b4f8-84f84489fb93.vmdk {{(pid=61923) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1027.852573] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-c35e31ab-0af5-4818-ac4e-4115156ea899 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: 0c3788fc-6cce-4806-b4f8-84f84489fb93] Extending root virtual disk to 1048576 {{(pid=61923) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1027.852819] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-4cd3910e-63fd-4ed1-8e56-2d0e032386f9 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.858967] env[61923]: DEBUG oslo_vmware.api [None req-c35e31ab-0af5-4818-ac4e-4115156ea899 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Waiting for the task: (returnval){ [ 1027.858967] env[61923]: value = "task-1377952" [ 1027.858967] env[61923]: _type = "Task" [ 1027.858967] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1027.861213] env[61923]: DEBUG nova.compute.manager [req-e27351be-26f8-4f8c-b73d-4bb055ccd180 req-6e18b6c6-9083-458b-ab50-1a9952798f86 service nova] [instance: 35bdbd45-5838-4b09-a232-7f4a02322ad7] Received event network-vif-plugged-42e3ee29-941c-43d0-a7f3-81fc7755f64b {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1027.861392] env[61923]: DEBUG oslo_concurrency.lockutils [req-e27351be-26f8-4f8c-b73d-4bb055ccd180 req-6e18b6c6-9083-458b-ab50-1a9952798f86 service nova] Acquiring lock "35bdbd45-5838-4b09-a232-7f4a02322ad7-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1027.861598] env[61923]: DEBUG oslo_concurrency.lockutils [req-e27351be-26f8-4f8c-b73d-4bb055ccd180 req-6e18b6c6-9083-458b-ab50-1a9952798f86 service nova] Lock "35bdbd45-5838-4b09-a232-7f4a02322ad7-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1027.861769] env[61923]: DEBUG oslo_concurrency.lockutils [req-e27351be-26f8-4f8c-b73d-4bb055ccd180 req-6e18b6c6-9083-458b-ab50-1a9952798f86 service nova] Lock "35bdbd45-5838-4b09-a232-7f4a02322ad7-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1027.861938] env[61923]: DEBUG nova.compute.manager [req-e27351be-26f8-4f8c-b73d-4bb055ccd180 req-6e18b6c6-9083-458b-ab50-1a9952798f86 service nova] [instance: 35bdbd45-5838-4b09-a232-7f4a02322ad7] No waiting events found dispatching network-vif-plugged-42e3ee29-941c-43d0-a7f3-81fc7755f64b {{(pid=61923) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1027.862129] env[61923]: WARNING nova.compute.manager [req-e27351be-26f8-4f8c-b73d-4bb055ccd180 req-6e18b6c6-9083-458b-ab50-1a9952798f86 service nova] [instance: 35bdbd45-5838-4b09-a232-7f4a02322ad7] Received unexpected event network-vif-plugged-42e3ee29-941c-43d0-a7f3-81fc7755f64b for instance with vm_state building and task_state spawning. [ 1027.870328] env[61923]: DEBUG oslo_vmware.api [None req-c35e31ab-0af5-4818-ac4e-4115156ea899 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Task: {'id': task-1377952, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1027.877398] env[61923]: DEBUG oslo_vmware.api [None req-ca0212d9-f46f-4fe8-98d2-915ce325cd19 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Task: {'id': task-1377951, 'name': RemoveSnapshot_Task, 'duration_secs': 0.395751} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1027.877634] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-ca0212d9-f46f-4fe8-98d2-915ce325cd19 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: a2f97947-6506-40b3-a24a-4ac03b88c5f5] Deleted Snapshot of the VM instance {{(pid=61923) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1027.877905] env[61923]: DEBUG nova.compute.manager [None req-ca0212d9-f46f-4fe8-98d2-915ce325cd19 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: a2f97947-6506-40b3-a24a-4ac03b88c5f5] Checking state {{(pid=61923) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1027.878650] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-742a1b5c-0553-4c05-a0f4-d449a5a53989 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.952462] env[61923]: DEBUG nova.network.neutron [None req-d17d2b02-87ac-45f0-981e-237fcddcf45c tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] [instance: 35bdbd45-5838-4b09-a232-7f4a02322ad7] Successfully updated port: 42e3ee29-941c-43d0-a7f3-81fc7755f64b {{(pid=61923) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1028.371035] env[61923]: DEBUG oslo_vmware.api [None req-c35e31ab-0af5-4818-ac4e-4115156ea899 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Task: {'id': task-1377952, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.072875} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1028.371035] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-c35e31ab-0af5-4818-ac4e-4115156ea899 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: 0c3788fc-6cce-4806-b4f8-84f84489fb93] Extended root virtual disk {{(pid=61923) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1028.371035] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-157abfa9-66ba-4616-85bc-c7accc3f22ca {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.392305] env[61923]: DEBUG nova.virt.vmwareapi.volumeops [None req-c35e31ab-0af5-4818-ac4e-4115156ea899 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: 0c3788fc-6cce-4806-b4f8-84f84489fb93] Reconfiguring VM instance instance-0000005f to attach disk [datastore2] 0c3788fc-6cce-4806-b4f8-84f84489fb93/0c3788fc-6cce-4806-b4f8-84f84489fb93.vmdk or device None with type sparse {{(pid=61923) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1028.394786] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f57e9b1b-b9c6-4841-a23b-6e429e08a57e {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.408589] env[61923]: INFO nova.compute.manager [None req-ca0212d9-f46f-4fe8-98d2-915ce325cd19 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: a2f97947-6506-40b3-a24a-4ac03b88c5f5] Shelve offloading [ 1028.410449] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-ca0212d9-f46f-4fe8-98d2-915ce325cd19 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: a2f97947-6506-40b3-a24a-4ac03b88c5f5] Powering off the VM {{(pid=61923) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1028.410743] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ea94b111-56d5-427f-bd78-cdd39062f052 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.414146] env[61923]: DEBUG oslo_vmware.api [None req-c35e31ab-0af5-4818-ac4e-4115156ea899 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Waiting for the task: (returnval){ [ 1028.414146] env[61923]: value = "task-1377953" [ 1028.414146] env[61923]: _type = "Task" [ 1028.414146] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1028.418941] env[61923]: DEBUG oslo_vmware.api [None req-ca0212d9-f46f-4fe8-98d2-915ce325cd19 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Waiting for the task: (returnval){ [ 1028.418941] env[61923]: value = "task-1377954" [ 1028.418941] env[61923]: _type = "Task" [ 1028.418941] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1028.426297] env[61923]: DEBUG oslo_vmware.api [None req-c35e31ab-0af5-4818-ac4e-4115156ea899 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Task: {'id': task-1377953, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1028.429594] env[61923]: DEBUG oslo_vmware.api [None req-ca0212d9-f46f-4fe8-98d2-915ce325cd19 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Task: {'id': task-1377954, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1028.454504] env[61923]: DEBUG oslo_concurrency.lockutils [None req-d17d2b02-87ac-45f0-981e-237fcddcf45c tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Acquiring lock "refresh_cache-35bdbd45-5838-4b09-a232-7f4a02322ad7" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1028.454504] env[61923]: DEBUG oslo_concurrency.lockutils [None req-d17d2b02-87ac-45f0-981e-237fcddcf45c tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Acquired lock "refresh_cache-35bdbd45-5838-4b09-a232-7f4a02322ad7" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1028.454670] env[61923]: DEBUG nova.network.neutron [None req-d17d2b02-87ac-45f0-981e-237fcddcf45c tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] [instance: 35bdbd45-5838-4b09-a232-7f4a02322ad7] Building network info cache for instance {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1028.926058] env[61923]: DEBUG oslo_vmware.api [None req-c35e31ab-0af5-4818-ac4e-4115156ea899 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Task: {'id': task-1377953, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1028.931286] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-ca0212d9-f46f-4fe8-98d2-915ce325cd19 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: a2f97947-6506-40b3-a24a-4ac03b88c5f5] VM already powered off {{(pid=61923) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 1028.931493] env[61923]: DEBUG nova.compute.manager [None req-ca0212d9-f46f-4fe8-98d2-915ce325cd19 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: a2f97947-6506-40b3-a24a-4ac03b88c5f5] Checking state {{(pid=61923) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1028.932239] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08be96e2-5045-453f-a282-0cfa6014a112 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.937628] env[61923]: DEBUG oslo_concurrency.lockutils [None req-ca0212d9-f46f-4fe8-98d2-915ce325cd19 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Acquiring lock "refresh_cache-a2f97947-6506-40b3-a24a-4ac03b88c5f5" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1028.937800] env[61923]: DEBUG oslo_concurrency.lockutils [None req-ca0212d9-f46f-4fe8-98d2-915ce325cd19 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Acquired lock "refresh_cache-a2f97947-6506-40b3-a24a-4ac03b88c5f5" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1028.937968] env[61923]: DEBUG nova.network.neutron [None req-ca0212d9-f46f-4fe8-98d2-915ce325cd19 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: a2f97947-6506-40b3-a24a-4ac03b88c5f5] Building network info cache for instance {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1028.988927] env[61923]: DEBUG nova.network.neutron [None req-d17d2b02-87ac-45f0-981e-237fcddcf45c tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] [instance: 35bdbd45-5838-4b09-a232-7f4a02322ad7] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1029.113932] env[61923]: DEBUG nova.network.neutron [None req-d17d2b02-87ac-45f0-981e-237fcddcf45c tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] [instance: 35bdbd45-5838-4b09-a232-7f4a02322ad7] Updating instance_info_cache with network_info: [{"id": "42e3ee29-941c-43d0-a7f3-81fc7755f64b", "address": "fa:16:3e:88:34:3b", "network": {"id": "d2905082-7ca4-4f92-a142-bb85211bb7c4", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-885218622-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a13d827933fa4597984afdb91dbbdd39", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "03ac2c9c-6ad2-4a85-bfab-c7e336df859a", "external-id": "nsx-vlan-transportzone-379", "segmentation_id": 379, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap42e3ee29-94", "ovs_interfaceid": "42e3ee29-941c-43d0-a7f3-81fc7755f64b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1029.425286] env[61923]: DEBUG oslo_vmware.api [None req-c35e31ab-0af5-4818-ac4e-4115156ea899 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Task: {'id': task-1377953, 'name': ReconfigVM_Task} progress is 99%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1029.613868] env[61923]: DEBUG nova.network.neutron [None req-ca0212d9-f46f-4fe8-98d2-915ce325cd19 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: a2f97947-6506-40b3-a24a-4ac03b88c5f5] Updating instance_info_cache with network_info: [{"id": "1aa4cc84-2b2c-4ca9-9b99-e2c94daacf4a", "address": "fa:16:3e:a1:5b:41", "network": {"id": "9533e79d-fde2-4c10-86a0-86a36845a8cf", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1736743137-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "780d3541d9604417b977bb62390c4299", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6d650b26-c3e7-4de7-98db-5e4b816d123a", "external-id": "nsx-vlan-transportzone-757", "segmentation_id": 757, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1aa4cc84-2b", "ovs_interfaceid": "1aa4cc84-2b2c-4ca9-9b99-e2c94daacf4a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1029.615821] env[61923]: DEBUG oslo_concurrency.lockutils [None req-d17d2b02-87ac-45f0-981e-237fcddcf45c tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Releasing lock "refresh_cache-35bdbd45-5838-4b09-a232-7f4a02322ad7" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1029.616108] env[61923]: DEBUG nova.compute.manager [None req-d17d2b02-87ac-45f0-981e-237fcddcf45c tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] [instance: 35bdbd45-5838-4b09-a232-7f4a02322ad7] Instance network_info: |[{"id": "42e3ee29-941c-43d0-a7f3-81fc7755f64b", "address": "fa:16:3e:88:34:3b", "network": {"id": "d2905082-7ca4-4f92-a142-bb85211bb7c4", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-885218622-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a13d827933fa4597984afdb91dbbdd39", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "03ac2c9c-6ad2-4a85-bfab-c7e336df859a", "external-id": "nsx-vlan-transportzone-379", "segmentation_id": 379, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap42e3ee29-94", "ovs_interfaceid": "42e3ee29-941c-43d0-a7f3-81fc7755f64b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61923) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1029.616507] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-d17d2b02-87ac-45f0-981e-237fcddcf45c tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] [instance: 35bdbd45-5838-4b09-a232-7f4a02322ad7] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:88:34:3b', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '03ac2c9c-6ad2-4a85-bfab-c7e336df859a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '42e3ee29-941c-43d0-a7f3-81fc7755f64b', 'vif_model': 'vmxnet3'}] {{(pid=61923) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1029.624065] env[61923]: DEBUG oslo.service.loopingcall [None req-d17d2b02-87ac-45f0-981e-237fcddcf45c tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61923) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1029.624866] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 35bdbd45-5838-4b09-a232-7f4a02322ad7] Creating VM on the ESX host {{(pid=61923) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1029.625101] env[61923]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-20f77e09-f4a7-4ede-bf55-6c4fa42f12ea {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.644172] env[61923]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1029.644172] env[61923]: value = "task-1377955" [ 1029.644172] env[61923]: _type = "Task" [ 1029.644172] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1029.652994] env[61923]: DEBUG oslo_vmware.api [-] Task: {'id': task-1377955, 'name': CreateVM_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1029.887332] env[61923]: DEBUG nova.compute.manager [req-5820984c-970c-48b1-8494-f5d031a19470 req-c5cb40df-c2f6-4bc3-b035-60ebaf2b3024 service nova] [instance: 35bdbd45-5838-4b09-a232-7f4a02322ad7] Received event network-changed-42e3ee29-941c-43d0-a7f3-81fc7755f64b {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1029.887589] env[61923]: DEBUG nova.compute.manager [req-5820984c-970c-48b1-8494-f5d031a19470 req-c5cb40df-c2f6-4bc3-b035-60ebaf2b3024 service nova] [instance: 35bdbd45-5838-4b09-a232-7f4a02322ad7] Refreshing instance network info cache due to event network-changed-42e3ee29-941c-43d0-a7f3-81fc7755f64b. {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1029.887767] env[61923]: DEBUG oslo_concurrency.lockutils [req-5820984c-970c-48b1-8494-f5d031a19470 req-c5cb40df-c2f6-4bc3-b035-60ebaf2b3024 service nova] Acquiring lock "refresh_cache-35bdbd45-5838-4b09-a232-7f4a02322ad7" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1029.887886] env[61923]: DEBUG oslo_concurrency.lockutils [req-5820984c-970c-48b1-8494-f5d031a19470 req-c5cb40df-c2f6-4bc3-b035-60ebaf2b3024 service nova] Acquired lock "refresh_cache-35bdbd45-5838-4b09-a232-7f4a02322ad7" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1029.888065] env[61923]: DEBUG nova.network.neutron [req-5820984c-970c-48b1-8494-f5d031a19470 req-c5cb40df-c2f6-4bc3-b035-60ebaf2b3024 service nova] [instance: 35bdbd45-5838-4b09-a232-7f4a02322ad7] Refreshing network info cache for port 42e3ee29-941c-43d0-a7f3-81fc7755f64b {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1029.925446] env[61923]: DEBUG oslo_vmware.api [None req-c35e31ab-0af5-4818-ac4e-4115156ea899 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Task: {'id': task-1377953, 'name': ReconfigVM_Task, 'duration_secs': 1.27394} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1029.925673] env[61923]: DEBUG nova.virt.vmwareapi.volumeops [None req-c35e31ab-0af5-4818-ac4e-4115156ea899 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: 0c3788fc-6cce-4806-b4f8-84f84489fb93] Reconfigured VM instance instance-0000005f to attach disk [datastore2] 0c3788fc-6cce-4806-b4f8-84f84489fb93/0c3788fc-6cce-4806-b4f8-84f84489fb93.vmdk or device None with type sparse {{(pid=61923) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1029.926334] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-0fbd3d40-b6d5-4dd4-9af5-2afb0e689380 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.931986] env[61923]: DEBUG oslo_vmware.api [None req-c35e31ab-0af5-4818-ac4e-4115156ea899 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Waiting for the task: (returnval){ [ 1029.931986] env[61923]: value = "task-1377956" [ 1029.931986] env[61923]: _type = "Task" [ 1029.931986] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1029.939412] env[61923]: DEBUG oslo_vmware.api [None req-c35e31ab-0af5-4818-ac4e-4115156ea899 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Task: {'id': task-1377956, 'name': Rename_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1030.116956] env[61923]: DEBUG oslo_concurrency.lockutils [None req-ca0212d9-f46f-4fe8-98d2-915ce325cd19 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Releasing lock "refresh_cache-a2f97947-6506-40b3-a24a-4ac03b88c5f5" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1030.153720] env[61923]: DEBUG oslo_vmware.api [-] Task: {'id': task-1377955, 'name': CreateVM_Task} progress is 99%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1030.366444] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-ca0212d9-f46f-4fe8-98d2-915ce325cd19 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: a2f97947-6506-40b3-a24a-4ac03b88c5f5] Destroying instance {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1030.367396] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-033b91e0-22ac-42d4-8254-534f51d13c05 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.374607] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-ca0212d9-f46f-4fe8-98d2-915ce325cd19 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: a2f97947-6506-40b3-a24a-4ac03b88c5f5] Unregistering the VM {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1030.374842] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-663321d6-c721-4bf6-bf20-e8c0625d8030 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.441267] env[61923]: DEBUG oslo_vmware.api [None req-c35e31ab-0af5-4818-ac4e-4115156ea899 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Task: {'id': task-1377956, 'name': Rename_Task, 'duration_secs': 0.145914} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1030.441540] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-c35e31ab-0af5-4818-ac4e-4115156ea899 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: 0c3788fc-6cce-4806-b4f8-84f84489fb93] Powering on the VM {{(pid=61923) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1030.441784] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e75409ee-9926-46e8-9a76-3d605e6a993f {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.448054] env[61923]: DEBUG oslo_vmware.api [None req-c35e31ab-0af5-4818-ac4e-4115156ea899 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Waiting for the task: (returnval){ [ 1030.448054] env[61923]: value = "task-1377958" [ 1030.448054] env[61923]: _type = "Task" [ 1030.448054] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1030.456998] env[61923]: DEBUG oslo_vmware.api [None req-c35e31ab-0af5-4818-ac4e-4115156ea899 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Task: {'id': task-1377958, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1030.578985] env[61923]: DEBUG nova.network.neutron [req-5820984c-970c-48b1-8494-f5d031a19470 req-c5cb40df-c2f6-4bc3-b035-60ebaf2b3024 service nova] [instance: 35bdbd45-5838-4b09-a232-7f4a02322ad7] Updated VIF entry in instance network info cache for port 42e3ee29-941c-43d0-a7f3-81fc7755f64b. {{(pid=61923) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1030.579374] env[61923]: DEBUG nova.network.neutron [req-5820984c-970c-48b1-8494-f5d031a19470 req-c5cb40df-c2f6-4bc3-b035-60ebaf2b3024 service nova] [instance: 35bdbd45-5838-4b09-a232-7f4a02322ad7] Updating instance_info_cache with network_info: [{"id": "42e3ee29-941c-43d0-a7f3-81fc7755f64b", "address": "fa:16:3e:88:34:3b", "network": {"id": "d2905082-7ca4-4f92-a142-bb85211bb7c4", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-885218622-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a13d827933fa4597984afdb91dbbdd39", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "03ac2c9c-6ad2-4a85-bfab-c7e336df859a", "external-id": "nsx-vlan-transportzone-379", "segmentation_id": 379, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap42e3ee29-94", "ovs_interfaceid": "42e3ee29-941c-43d0-a7f3-81fc7755f64b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1030.654828] env[61923]: DEBUG oslo_vmware.api [-] Task: {'id': task-1377955, 'name': CreateVM_Task} progress is 99%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1030.780986] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-ca0212d9-f46f-4fe8-98d2-915ce325cd19 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: a2f97947-6506-40b3-a24a-4ac03b88c5f5] Unregistered the VM {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1030.781262] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-ca0212d9-f46f-4fe8-98d2-915ce325cd19 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: a2f97947-6506-40b3-a24a-4ac03b88c5f5] Deleting contents of the VM from datastore datastore1 {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1030.781479] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-ca0212d9-f46f-4fe8-98d2-915ce325cd19 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Deleting the datastore file [datastore1] a2f97947-6506-40b3-a24a-4ac03b88c5f5 {{(pid=61923) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1030.781764] env[61923]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-12361ab6-44d9-4eb1-ba65-70a070029d75 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.787607] env[61923]: DEBUG oslo_vmware.api [None req-ca0212d9-f46f-4fe8-98d2-915ce325cd19 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Waiting for the task: (returnval){ [ 1030.787607] env[61923]: value = "task-1377959" [ 1030.787607] env[61923]: _type = "Task" [ 1030.787607] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1030.795292] env[61923]: DEBUG oslo_vmware.api [None req-ca0212d9-f46f-4fe8-98d2-915ce325cd19 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Task: {'id': task-1377959, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1030.958242] env[61923]: DEBUG oslo_vmware.api [None req-c35e31ab-0af5-4818-ac4e-4115156ea899 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Task: {'id': task-1377958, 'name': PowerOnVM_Task, 'duration_secs': 0.47902} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1030.958646] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-c35e31ab-0af5-4818-ac4e-4115156ea899 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: 0c3788fc-6cce-4806-b4f8-84f84489fb93] Powered on the VM {{(pid=61923) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1030.958784] env[61923]: DEBUG nova.compute.manager [None req-c35e31ab-0af5-4818-ac4e-4115156ea899 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: 0c3788fc-6cce-4806-b4f8-84f84489fb93] Checking state {{(pid=61923) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1030.959560] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3f6ea75-5aff-4288-bc0b-9dc9221c13dc {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.081778] env[61923]: DEBUG oslo_concurrency.lockutils [req-5820984c-970c-48b1-8494-f5d031a19470 req-c5cb40df-c2f6-4bc3-b035-60ebaf2b3024 service nova] Releasing lock "refresh_cache-35bdbd45-5838-4b09-a232-7f4a02322ad7" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1031.154643] env[61923]: DEBUG oslo_vmware.api [-] Task: {'id': task-1377955, 'name': CreateVM_Task, 'duration_secs': 1.3986} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1031.154897] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 35bdbd45-5838-4b09-a232-7f4a02322ad7] Created VM on the ESX host {{(pid=61923) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1031.155618] env[61923]: DEBUG oslo_concurrency.lockutils [None req-d17d2b02-87ac-45f0-981e-237fcddcf45c tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1031.155793] env[61923]: DEBUG oslo_concurrency.lockutils [None req-d17d2b02-87ac-45f0-981e-237fcddcf45c tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1031.156134] env[61923]: DEBUG oslo_concurrency.lockutils [None req-d17d2b02-87ac-45f0-981e-237fcddcf45c tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1031.156391] env[61923]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ade85c35-1bf9-4031-b88f-7b936ad58c90 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.160866] env[61923]: DEBUG oslo_vmware.api [None req-d17d2b02-87ac-45f0-981e-237fcddcf45c tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Waiting for the task: (returnval){ [ 1031.160866] env[61923]: value = "session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52ff01cc-44b7-7d50-eb14-cbebf0e582cb" [ 1031.160866] env[61923]: _type = "Task" [ 1031.160866] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1031.169654] env[61923]: DEBUG oslo_vmware.api [None req-d17d2b02-87ac-45f0-981e-237fcddcf45c tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52ff01cc-44b7-7d50-eb14-cbebf0e582cb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1031.297561] env[61923]: DEBUG oslo_vmware.api [None req-ca0212d9-f46f-4fe8-98d2-915ce325cd19 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Task: {'id': task-1377959, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.149142} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1031.297931] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-ca0212d9-f46f-4fe8-98d2-915ce325cd19 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Deleted the datastore file {{(pid=61923) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1031.298193] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-ca0212d9-f46f-4fe8-98d2-915ce325cd19 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: a2f97947-6506-40b3-a24a-4ac03b88c5f5] Deleted contents of the VM from datastore datastore1 {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1031.298420] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-ca0212d9-f46f-4fe8-98d2-915ce325cd19 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: a2f97947-6506-40b3-a24a-4ac03b88c5f5] Instance destroyed {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1031.320317] env[61923]: INFO nova.scheduler.client.report [None req-ca0212d9-f46f-4fe8-98d2-915ce325cd19 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Deleted allocations for instance a2f97947-6506-40b3-a24a-4ac03b88c5f5 [ 1031.477024] env[61923]: DEBUG oslo_concurrency.lockutils [None req-c35e31ab-0af5-4818-ac4e-4115156ea899 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1031.477024] env[61923]: DEBUG oslo_concurrency.lockutils [None req-c35e31ab-0af5-4818-ac4e-4115156ea899 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1031.477024] env[61923]: DEBUG nova.objects.instance [None req-c35e31ab-0af5-4818-ac4e-4115156ea899 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: 0c3788fc-6cce-4806-b4f8-84f84489fb93] Trying to apply a migration context that does not seem to be set for this instance {{(pid=61923) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 1031.671665] env[61923]: DEBUG oslo_vmware.api [None req-d17d2b02-87ac-45f0-981e-237fcddcf45c tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52ff01cc-44b7-7d50-eb14-cbebf0e582cb, 'name': SearchDatastore_Task, 'duration_secs': 0.009045} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1031.671970] env[61923]: DEBUG oslo_concurrency.lockutils [None req-d17d2b02-87ac-45f0-981e-237fcddcf45c tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1031.672230] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-d17d2b02-87ac-45f0-981e-237fcddcf45c tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] [instance: 35bdbd45-5838-4b09-a232-7f4a02322ad7] Processing image 0f153f63-ae0a-45b1-b7f5-7f9b673c947f {{(pid=61923) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1031.672471] env[61923]: DEBUG oslo_concurrency.lockutils [None req-d17d2b02-87ac-45f0-981e-237fcddcf45c tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1031.672619] env[61923]: DEBUG oslo_concurrency.lockutils [None req-d17d2b02-87ac-45f0-981e-237fcddcf45c tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1031.672798] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-d17d2b02-87ac-45f0-981e-237fcddcf45c tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61923) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1031.673073] env[61923]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b6edfca8-eb51-45e0-a85b-f8080fa19a6f {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.680602] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-d17d2b02-87ac-45f0-981e-237fcddcf45c tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61923) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1031.680765] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-d17d2b02-87ac-45f0-981e-237fcddcf45c tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61923) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1031.681468] env[61923]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e43d5c26-c8a5-490f-91f6-d1eb55e56f34 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.686794] env[61923]: DEBUG oslo_vmware.api [None req-d17d2b02-87ac-45f0-981e-237fcddcf45c tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Waiting for the task: (returnval){ [ 1031.686794] env[61923]: value = "session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52ca899e-ef37-dd0c-91ba-ac7c41b28eb1" [ 1031.686794] env[61923]: _type = "Task" [ 1031.686794] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1031.694602] env[61923]: DEBUG oslo_vmware.api [None req-d17d2b02-87ac-45f0-981e-237fcddcf45c tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52ca899e-ef37-dd0c-91ba-ac7c41b28eb1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1031.824562] env[61923]: DEBUG oslo_concurrency.lockutils [None req-ca0212d9-f46f-4fe8-98d2-915ce325cd19 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1031.905232] env[61923]: DEBUG oslo_concurrency.lockutils [None req-1c23d35b-fb18-4b83-9ba6-8e1d103da1b8 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Acquiring lock "0c3788fc-6cce-4806-b4f8-84f84489fb93" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1031.905558] env[61923]: DEBUG oslo_concurrency.lockutils [None req-1c23d35b-fb18-4b83-9ba6-8e1d103da1b8 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Lock "0c3788fc-6cce-4806-b4f8-84f84489fb93" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1031.905781] env[61923]: DEBUG oslo_concurrency.lockutils [None req-1c23d35b-fb18-4b83-9ba6-8e1d103da1b8 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Acquiring lock "0c3788fc-6cce-4806-b4f8-84f84489fb93-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1031.905973] env[61923]: DEBUG oslo_concurrency.lockutils [None req-1c23d35b-fb18-4b83-9ba6-8e1d103da1b8 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Lock "0c3788fc-6cce-4806-b4f8-84f84489fb93-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1031.906172] env[61923]: DEBUG oslo_concurrency.lockutils [None req-1c23d35b-fb18-4b83-9ba6-8e1d103da1b8 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Lock "0c3788fc-6cce-4806-b4f8-84f84489fb93-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1031.908408] env[61923]: INFO nova.compute.manager [None req-1c23d35b-fb18-4b83-9ba6-8e1d103da1b8 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: 0c3788fc-6cce-4806-b4f8-84f84489fb93] Terminating instance [ 1031.910571] env[61923]: DEBUG nova.compute.manager [None req-1c23d35b-fb18-4b83-9ba6-8e1d103da1b8 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: 0c3788fc-6cce-4806-b4f8-84f84489fb93] Start destroying the instance on the hypervisor. {{(pid=61923) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1031.910789] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-1c23d35b-fb18-4b83-9ba6-8e1d103da1b8 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: 0c3788fc-6cce-4806-b4f8-84f84489fb93] Destroying instance {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1031.911680] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8cd056e-0ece-4546-8a38-6b15b17e426f {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.918239] env[61923]: DEBUG nova.compute.manager [req-6af173c8-07bf-4568-876e-114d61df50f3 req-b883af55-3414-4323-9ec1-dde5c3483982 service nova] [instance: a2f97947-6506-40b3-a24a-4ac03b88c5f5] Received event network-vif-unplugged-1aa4cc84-2b2c-4ca9-9b99-e2c94daacf4a {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1031.918239] env[61923]: DEBUG oslo_concurrency.lockutils [req-6af173c8-07bf-4568-876e-114d61df50f3 req-b883af55-3414-4323-9ec1-dde5c3483982 service nova] Acquiring lock "a2f97947-6506-40b3-a24a-4ac03b88c5f5-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1031.918239] env[61923]: DEBUG oslo_concurrency.lockutils [req-6af173c8-07bf-4568-876e-114d61df50f3 req-b883af55-3414-4323-9ec1-dde5c3483982 service nova] Lock "a2f97947-6506-40b3-a24a-4ac03b88c5f5-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1031.918239] env[61923]: DEBUG oslo_concurrency.lockutils [req-6af173c8-07bf-4568-876e-114d61df50f3 req-b883af55-3414-4323-9ec1-dde5c3483982 service nova] Lock "a2f97947-6506-40b3-a24a-4ac03b88c5f5-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1031.918239] env[61923]: DEBUG nova.compute.manager [req-6af173c8-07bf-4568-876e-114d61df50f3 req-b883af55-3414-4323-9ec1-dde5c3483982 service nova] [instance: a2f97947-6506-40b3-a24a-4ac03b88c5f5] No waiting events found dispatching network-vif-unplugged-1aa4cc84-2b2c-4ca9-9b99-e2c94daacf4a {{(pid=61923) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1031.918453] env[61923]: WARNING nova.compute.manager [req-6af173c8-07bf-4568-876e-114d61df50f3 req-b883af55-3414-4323-9ec1-dde5c3483982 service nova] [instance: a2f97947-6506-40b3-a24a-4ac03b88c5f5] Received unexpected event network-vif-unplugged-1aa4cc84-2b2c-4ca9-9b99-e2c94daacf4a for instance with vm_state shelved_offloaded and task_state None. [ 1031.918453] env[61923]: DEBUG nova.compute.manager [req-6af173c8-07bf-4568-876e-114d61df50f3 req-b883af55-3414-4323-9ec1-dde5c3483982 service nova] [instance: a2f97947-6506-40b3-a24a-4ac03b88c5f5] Received event network-changed-1aa4cc84-2b2c-4ca9-9b99-e2c94daacf4a {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1031.918609] env[61923]: DEBUG nova.compute.manager [req-6af173c8-07bf-4568-876e-114d61df50f3 req-b883af55-3414-4323-9ec1-dde5c3483982 service nova] [instance: a2f97947-6506-40b3-a24a-4ac03b88c5f5] Refreshing instance network info cache due to event network-changed-1aa4cc84-2b2c-4ca9-9b99-e2c94daacf4a. {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1031.918866] env[61923]: DEBUG oslo_concurrency.lockutils [req-6af173c8-07bf-4568-876e-114d61df50f3 req-b883af55-3414-4323-9ec1-dde5c3483982 service nova] Acquiring lock "refresh_cache-a2f97947-6506-40b3-a24a-4ac03b88c5f5" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1031.919036] env[61923]: DEBUG oslo_concurrency.lockutils [req-6af173c8-07bf-4568-876e-114d61df50f3 req-b883af55-3414-4323-9ec1-dde5c3483982 service nova] Acquired lock "refresh_cache-a2f97947-6506-40b3-a24a-4ac03b88c5f5" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1031.919207] env[61923]: DEBUG nova.network.neutron [req-6af173c8-07bf-4568-876e-114d61df50f3 req-b883af55-3414-4323-9ec1-dde5c3483982 service nova] [instance: a2f97947-6506-40b3-a24a-4ac03b88c5f5] Refreshing network info cache for port 1aa4cc84-2b2c-4ca9-9b99-e2c94daacf4a {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1031.922823] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-1c23d35b-fb18-4b83-9ba6-8e1d103da1b8 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: 0c3788fc-6cce-4806-b4f8-84f84489fb93] Powering off the VM {{(pid=61923) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1031.923056] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e7db9762-0be4-427a-840a-4391ef1bb2ae {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.930316] env[61923]: DEBUG oslo_vmware.api [None req-1c23d35b-fb18-4b83-9ba6-8e1d103da1b8 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Waiting for the task: (returnval){ [ 1031.930316] env[61923]: value = "task-1377960" [ 1031.930316] env[61923]: _type = "Task" [ 1031.930316] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1031.938066] env[61923]: DEBUG oslo_vmware.api [None req-1c23d35b-fb18-4b83-9ba6-8e1d103da1b8 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Task: {'id': task-1377960, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1032.198113] env[61923]: DEBUG oslo_vmware.api [None req-d17d2b02-87ac-45f0-981e-237fcddcf45c tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52ca899e-ef37-dd0c-91ba-ac7c41b28eb1, 'name': SearchDatastore_Task, 'duration_secs': 0.008735} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1032.198720] env[61923]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-707fa195-158e-40db-890c-e14a9481fbb6 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.203737] env[61923]: DEBUG oslo_vmware.api [None req-d17d2b02-87ac-45f0-981e-237fcddcf45c tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Waiting for the task: (returnval){ [ 1032.203737] env[61923]: value = "session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]5258a2d4-1a63-216e-60fc-b898de32cb3e" [ 1032.203737] env[61923]: _type = "Task" [ 1032.203737] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1032.210807] env[61923]: DEBUG oslo_vmware.api [None req-d17d2b02-87ac-45f0-981e-237fcddcf45c tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]5258a2d4-1a63-216e-60fc-b898de32cb3e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1032.356390] env[61923]: DEBUG oslo_concurrency.lockutils [None req-84927854-5e83-4111-abce-383944a18027 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Acquiring lock "9aeac922-aac2-47fe-8c11-835bef75e55c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1032.356623] env[61923]: DEBUG oslo_concurrency.lockutils [None req-84927854-5e83-4111-abce-383944a18027 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Lock "9aeac922-aac2-47fe-8c11-835bef75e55c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1032.440167] env[61923]: DEBUG oslo_vmware.api [None req-1c23d35b-fb18-4b83-9ba6-8e1d103da1b8 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Task: {'id': task-1377960, 'name': PowerOffVM_Task, 'duration_secs': 0.181021} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1032.440472] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-1c23d35b-fb18-4b83-9ba6-8e1d103da1b8 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: 0c3788fc-6cce-4806-b4f8-84f84489fb93] Powered off the VM {{(pid=61923) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1032.440884] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-1c23d35b-fb18-4b83-9ba6-8e1d103da1b8 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: 0c3788fc-6cce-4806-b4f8-84f84489fb93] Unregistering the VM {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1032.440976] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-5fb64a3c-734b-4d79-9f81-196da306ad8e {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.485352] env[61923]: DEBUG oslo_concurrency.lockutils [None req-c35e31ab-0af5-4818-ac4e-4115156ea899 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.009s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1032.487414] env[61923]: DEBUG oslo_concurrency.lockutils [None req-ca0212d9-f46f-4fe8-98d2-915ce325cd19 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.663s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1032.487776] env[61923]: DEBUG nova.objects.instance [None req-ca0212d9-f46f-4fe8-98d2-915ce325cd19 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Lazy-loading 'resources' on Instance uuid a2f97947-6506-40b3-a24a-4ac03b88c5f5 {{(pid=61923) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1032.503172] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-1c23d35b-fb18-4b83-9ba6-8e1d103da1b8 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: 0c3788fc-6cce-4806-b4f8-84f84489fb93] Unregistered the VM {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1032.503420] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-1c23d35b-fb18-4b83-9ba6-8e1d103da1b8 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: 0c3788fc-6cce-4806-b4f8-84f84489fb93] Deleting contents of the VM from datastore datastore2 {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1032.503554] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-1c23d35b-fb18-4b83-9ba6-8e1d103da1b8 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Deleting the datastore file [datastore2] 0c3788fc-6cce-4806-b4f8-84f84489fb93 {{(pid=61923) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1032.503811] env[61923]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-1e7a6742-db8a-4e2c-a2d4-16766e9e8c21 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.511764] env[61923]: DEBUG oslo_vmware.api [None req-1c23d35b-fb18-4b83-9ba6-8e1d103da1b8 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Waiting for the task: (returnval){ [ 1032.511764] env[61923]: value = "task-1377962" [ 1032.511764] env[61923]: _type = "Task" [ 1032.511764] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1032.517701] env[61923]: DEBUG oslo_vmware.api [None req-1c23d35b-fb18-4b83-9ba6-8e1d103da1b8 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Task: {'id': task-1377962, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1032.632335] env[61923]: DEBUG nova.network.neutron [req-6af173c8-07bf-4568-876e-114d61df50f3 req-b883af55-3414-4323-9ec1-dde5c3483982 service nova] [instance: a2f97947-6506-40b3-a24a-4ac03b88c5f5] Updated VIF entry in instance network info cache for port 1aa4cc84-2b2c-4ca9-9b99-e2c94daacf4a. {{(pid=61923) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1032.632704] env[61923]: DEBUG nova.network.neutron [req-6af173c8-07bf-4568-876e-114d61df50f3 req-b883af55-3414-4323-9ec1-dde5c3483982 service nova] [instance: a2f97947-6506-40b3-a24a-4ac03b88c5f5] Updating instance_info_cache with network_info: [{"id": "1aa4cc84-2b2c-4ca9-9b99-e2c94daacf4a", "address": "fa:16:3e:a1:5b:41", "network": {"id": "9533e79d-fde2-4c10-86a0-86a36845a8cf", "bridge": null, "label": "tempest-DeleteServersTestJSON-1736743137-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "780d3541d9604417b977bb62390c4299", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tap1aa4cc84-2b", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1032.714383] env[61923]: DEBUG oslo_vmware.api [None req-d17d2b02-87ac-45f0-981e-237fcddcf45c tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]5258a2d4-1a63-216e-60fc-b898de32cb3e, 'name': SearchDatastore_Task, 'duration_secs': 0.008112} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1032.714649] env[61923]: DEBUG oslo_concurrency.lockutils [None req-d17d2b02-87ac-45f0-981e-237fcddcf45c tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1032.714905] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-d17d2b02-87ac-45f0-981e-237fcddcf45c tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk to [datastore2] 35bdbd45-5838-4b09-a232-7f4a02322ad7/35bdbd45-5838-4b09-a232-7f4a02322ad7.vmdk {{(pid=61923) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1032.715177] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c9a29d70-7b6b-4329-a2c8-3e7d2d562811 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.721347] env[61923]: DEBUG oslo_vmware.api [None req-d17d2b02-87ac-45f0-981e-237fcddcf45c tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Waiting for the task: (returnval){ [ 1032.721347] env[61923]: value = "task-1377963" [ 1032.721347] env[61923]: _type = "Task" [ 1032.721347] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1032.729178] env[61923]: DEBUG oslo_vmware.api [None req-d17d2b02-87ac-45f0-981e-237fcddcf45c tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Task: {'id': task-1377963, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1032.862229] env[61923]: DEBUG nova.compute.manager [None req-84927854-5e83-4111-abce-383944a18027 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: 9aeac922-aac2-47fe-8c11-835bef75e55c] Starting instance... {{(pid=61923) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1032.923433] env[61923]: DEBUG oslo_service.periodic_task [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61923) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1032.924084] env[61923]: DEBUG oslo_service.periodic_task [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61923) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1032.924402] env[61923]: DEBUG nova.compute.manager [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Starting heal instance info cache {{(pid=61923) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 1032.989978] env[61923]: DEBUG nova.objects.instance [None req-ca0212d9-f46f-4fe8-98d2-915ce325cd19 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Lazy-loading 'numa_topology' on Instance uuid a2f97947-6506-40b3-a24a-4ac03b88c5f5 {{(pid=61923) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1033.022259] env[61923]: DEBUG oslo_vmware.api [None req-1c23d35b-fb18-4b83-9ba6-8e1d103da1b8 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Task: {'id': task-1377962, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.123024} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1033.022583] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-1c23d35b-fb18-4b83-9ba6-8e1d103da1b8 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Deleted the datastore file {{(pid=61923) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1033.022826] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-1c23d35b-fb18-4b83-9ba6-8e1d103da1b8 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: 0c3788fc-6cce-4806-b4f8-84f84489fb93] Deleted contents of the VM from datastore datastore2 {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1033.023056] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-1c23d35b-fb18-4b83-9ba6-8e1d103da1b8 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: 0c3788fc-6cce-4806-b4f8-84f84489fb93] Instance destroyed {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1033.023244] env[61923]: INFO nova.compute.manager [None req-1c23d35b-fb18-4b83-9ba6-8e1d103da1b8 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: 0c3788fc-6cce-4806-b4f8-84f84489fb93] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1033.023578] env[61923]: DEBUG oslo.service.loopingcall [None req-1c23d35b-fb18-4b83-9ba6-8e1d103da1b8 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61923) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1033.023798] env[61923]: DEBUG nova.compute.manager [-] [instance: 0c3788fc-6cce-4806-b4f8-84f84489fb93] Deallocating network for instance {{(pid=61923) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1033.023896] env[61923]: DEBUG nova.network.neutron [-] [instance: 0c3788fc-6cce-4806-b4f8-84f84489fb93] deallocate_for_instance() {{(pid=61923) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1033.043167] env[61923]: DEBUG oslo_concurrency.lockutils [None req-ec355d09-be8b-4126-8e84-491179ed7cd9 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Acquiring lock "a2f97947-6506-40b3-a24a-4ac03b88c5f5" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1033.135689] env[61923]: DEBUG oslo_concurrency.lockutils [req-6af173c8-07bf-4568-876e-114d61df50f3 req-b883af55-3414-4323-9ec1-dde5c3483982 service nova] Releasing lock "refresh_cache-a2f97947-6506-40b3-a24a-4ac03b88c5f5" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1033.231576] env[61923]: DEBUG oslo_vmware.api [None req-d17d2b02-87ac-45f0-981e-237fcddcf45c tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Task: {'id': task-1377963, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.433753} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1033.231904] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-d17d2b02-87ac-45f0-981e-237fcddcf45c tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk to [datastore2] 35bdbd45-5838-4b09-a232-7f4a02322ad7/35bdbd45-5838-4b09-a232-7f4a02322ad7.vmdk {{(pid=61923) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1033.232050] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-d17d2b02-87ac-45f0-981e-237fcddcf45c tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] [instance: 35bdbd45-5838-4b09-a232-7f4a02322ad7] Extending root virtual disk to 1048576 {{(pid=61923) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1033.232376] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-4ac44b68-0455-4bff-a4ba-2a9aa7f9bc9f {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.238808] env[61923]: DEBUG oslo_vmware.api [None req-d17d2b02-87ac-45f0-981e-237fcddcf45c tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Waiting for the task: (returnval){ [ 1033.238808] env[61923]: value = "task-1377964" [ 1033.238808] env[61923]: _type = "Task" [ 1033.238808] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1033.246165] env[61923]: DEBUG oslo_vmware.api [None req-d17d2b02-87ac-45f0-981e-237fcddcf45c tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Task: {'id': task-1377964, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1033.385382] env[61923]: DEBUG oslo_concurrency.lockutils [None req-84927854-5e83-4111-abce-383944a18027 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1033.492164] env[61923]: DEBUG nova.objects.base [None req-ca0212d9-f46f-4fe8-98d2-915ce325cd19 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Object Instance lazy-loaded attributes: resources,numa_topology {{(pid=61923) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 1033.536727] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c17424cf-1ae3-497a-bc56-5e7736314676 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.544604] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da9cbdfb-6d18-4670-a5dd-0f24e8b951c9 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.573524] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7154b90e-6c5d-4994-a79e-5d0c9a653b80 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.580118] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-715df250-5b47-48a4-b64c-e02c9dd454e3 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.593028] env[61923]: DEBUG nova.compute.provider_tree [None req-ca0212d9-f46f-4fe8-98d2-915ce325cd19 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1033.748738] env[61923]: DEBUG oslo_vmware.api [None req-d17d2b02-87ac-45f0-981e-237fcddcf45c tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Task: {'id': task-1377964, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.068626} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1033.749031] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-d17d2b02-87ac-45f0-981e-237fcddcf45c tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] [instance: 35bdbd45-5838-4b09-a232-7f4a02322ad7] Extended root virtual disk {{(pid=61923) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1033.749775] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c0d62ba-7a06-4dbd-9c42-bff35f2ad8c6 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.771632] env[61923]: DEBUG nova.virt.vmwareapi.volumeops [None req-d17d2b02-87ac-45f0-981e-237fcddcf45c tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] [instance: 35bdbd45-5838-4b09-a232-7f4a02322ad7] Reconfiguring VM instance instance-00000060 to attach disk [datastore2] 35bdbd45-5838-4b09-a232-7f4a02322ad7/35bdbd45-5838-4b09-a232-7f4a02322ad7.vmdk or device None with type sparse {{(pid=61923) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1033.771835] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-31db0f0c-dd4e-44e0-bb7c-80230fec2bf5 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.787511] env[61923]: DEBUG nova.network.neutron [-] [instance: 0c3788fc-6cce-4806-b4f8-84f84489fb93] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1033.789976] env[61923]: DEBUG oslo_vmware.api [None req-d17d2b02-87ac-45f0-981e-237fcddcf45c tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Waiting for the task: (returnval){ [ 1033.789976] env[61923]: value = "task-1377965" [ 1033.789976] env[61923]: _type = "Task" [ 1033.789976] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1033.798288] env[61923]: DEBUG oslo_vmware.api [None req-d17d2b02-87ac-45f0-981e-237fcddcf45c tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Task: {'id': task-1377965, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1033.931536] env[61923]: DEBUG nova.compute.manager [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Didn't find any instances for network info cache update. {{(pid=61923) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10014}} [ 1033.931831] env[61923]: DEBUG oslo_service.periodic_task [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61923) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1033.932081] env[61923]: DEBUG oslo_service.periodic_task [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61923) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1033.932313] env[61923]: DEBUG oslo_service.periodic_task [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61923) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1033.932509] env[61923]: DEBUG oslo_service.periodic_task [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61923) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1033.932774] env[61923]: DEBUG oslo_service.periodic_task [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61923) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1033.932952] env[61923]: DEBUG oslo_service.periodic_task [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61923) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1033.933182] env[61923]: DEBUG nova.compute.manager [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61923) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 1033.933406] env[61923]: DEBUG oslo_service.periodic_task [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Running periodic task ComputeManager.update_available_resource {{(pid=61923) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1033.944030] env[61923]: DEBUG nova.compute.manager [req-3a0739a0-991b-43af-81fb-9be3273e1e77 req-91eeded4-c4a6-4c23-baed-78b76edf17a0 service nova] [instance: 0c3788fc-6cce-4806-b4f8-84f84489fb93] Received event network-vif-deleted-504f15e6-4eee-4266-817f-d7935b5d0609 {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1034.095923] env[61923]: DEBUG nova.scheduler.client.report [None req-ca0212d9-f46f-4fe8-98d2-915ce325cd19 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1034.290985] env[61923]: INFO nova.compute.manager [-] [instance: 0c3788fc-6cce-4806-b4f8-84f84489fb93] Took 1.27 seconds to deallocate network for instance. [ 1034.303325] env[61923]: DEBUG oslo_vmware.api [None req-d17d2b02-87ac-45f0-981e-237fcddcf45c tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Task: {'id': task-1377965, 'name': ReconfigVM_Task, 'duration_secs': 0.25955} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1034.303598] env[61923]: DEBUG nova.virt.vmwareapi.volumeops [None req-d17d2b02-87ac-45f0-981e-237fcddcf45c tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] [instance: 35bdbd45-5838-4b09-a232-7f4a02322ad7] Reconfigured VM instance instance-00000060 to attach disk [datastore2] 35bdbd45-5838-4b09-a232-7f4a02322ad7/35bdbd45-5838-4b09-a232-7f4a02322ad7.vmdk or device None with type sparse {{(pid=61923) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1034.304207] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-98c5dff8-88c4-4366-87e9-5d7196ad8b21 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.310137] env[61923]: DEBUG oslo_vmware.api [None req-d17d2b02-87ac-45f0-981e-237fcddcf45c tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Waiting for the task: (returnval){ [ 1034.310137] env[61923]: value = "task-1377966" [ 1034.310137] env[61923]: _type = "Task" [ 1034.310137] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1034.321782] env[61923]: DEBUG oslo_vmware.api [None req-d17d2b02-87ac-45f0-981e-237fcddcf45c tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Task: {'id': task-1377966, 'name': Rename_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1034.443600] env[61923]: DEBUG oslo_concurrency.lockutils [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1034.601112] env[61923]: DEBUG oslo_concurrency.lockutils [None req-ca0212d9-f46f-4fe8-98d2-915ce325cd19 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.114s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1034.603999] env[61923]: DEBUG oslo_concurrency.lockutils [None req-84927854-5e83-4111-abce-383944a18027 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.219s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1034.605439] env[61923]: INFO nova.compute.claims [None req-84927854-5e83-4111-abce-383944a18027 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: 9aeac922-aac2-47fe-8c11-835bef75e55c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1034.799874] env[61923]: DEBUG oslo_concurrency.lockutils [None req-1c23d35b-fb18-4b83-9ba6-8e1d103da1b8 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1034.820428] env[61923]: DEBUG oslo_vmware.api [None req-d17d2b02-87ac-45f0-981e-237fcddcf45c tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Task: {'id': task-1377966, 'name': Rename_Task} progress is 99%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1035.114126] env[61923]: DEBUG oslo_concurrency.lockutils [None req-ca0212d9-f46f-4fe8-98d2-915ce325cd19 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Lock "a2f97947-6506-40b3-a24a-4ac03b88c5f5" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 21.571s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1035.115307] env[61923]: DEBUG oslo_concurrency.lockutils [None req-ec355d09-be8b-4126-8e84-491179ed7cd9 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Lock "a2f97947-6506-40b3-a24a-4ac03b88c5f5" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 2.072s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1035.115307] env[61923]: DEBUG oslo_concurrency.lockutils [None req-ec355d09-be8b-4126-8e84-491179ed7cd9 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Acquiring lock "a2f97947-6506-40b3-a24a-4ac03b88c5f5-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1035.115434] env[61923]: DEBUG oslo_concurrency.lockutils [None req-ec355d09-be8b-4126-8e84-491179ed7cd9 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Lock "a2f97947-6506-40b3-a24a-4ac03b88c5f5-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1035.115574] env[61923]: DEBUG oslo_concurrency.lockutils [None req-ec355d09-be8b-4126-8e84-491179ed7cd9 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Lock "a2f97947-6506-40b3-a24a-4ac03b88c5f5-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1035.117117] env[61923]: INFO nova.compute.manager [None req-ec355d09-be8b-4126-8e84-491179ed7cd9 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: a2f97947-6506-40b3-a24a-4ac03b88c5f5] Terminating instance [ 1035.118635] env[61923]: DEBUG nova.compute.manager [None req-ec355d09-be8b-4126-8e84-491179ed7cd9 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: a2f97947-6506-40b3-a24a-4ac03b88c5f5] Start destroying the instance on the hypervisor. {{(pid=61923) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1035.118832] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-ec355d09-be8b-4126-8e84-491179ed7cd9 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: a2f97947-6506-40b3-a24a-4ac03b88c5f5] Destroying instance {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1035.119109] env[61923]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-88439f9d-6636-4c3f-b6c7-dec7a5ddbc24 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.129090] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e1f3286-cc38-4e79-a0d0-77bf93cade53 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.152764] env[61923]: WARNING nova.virt.vmwareapi.vmops [None req-ec355d09-be8b-4126-8e84-491179ed7cd9 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: a2f97947-6506-40b3-a24a-4ac03b88c5f5] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance a2f97947-6506-40b3-a24a-4ac03b88c5f5 could not be found. [ 1035.152961] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-ec355d09-be8b-4126-8e84-491179ed7cd9 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: a2f97947-6506-40b3-a24a-4ac03b88c5f5] Instance destroyed {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1035.153159] env[61923]: INFO nova.compute.manager [None req-ec355d09-be8b-4126-8e84-491179ed7cd9 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: a2f97947-6506-40b3-a24a-4ac03b88c5f5] Took 0.03 seconds to destroy the instance on the hypervisor. [ 1035.153398] env[61923]: DEBUG oslo.service.loopingcall [None req-ec355d09-be8b-4126-8e84-491179ed7cd9 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61923) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1035.153609] env[61923]: DEBUG nova.compute.manager [-] [instance: a2f97947-6506-40b3-a24a-4ac03b88c5f5] Deallocating network for instance {{(pid=61923) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1035.153699] env[61923]: DEBUG nova.network.neutron [-] [instance: a2f97947-6506-40b3-a24a-4ac03b88c5f5] deallocate_for_instance() {{(pid=61923) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1035.320774] env[61923]: DEBUG oslo_vmware.api [None req-d17d2b02-87ac-45f0-981e-237fcddcf45c tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Task: {'id': task-1377966, 'name': Rename_Task} progress is 99%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1035.670411] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7900fb19-b7cc-48ba-a56b-13c255a787d7 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.677737] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c8649af-153f-4129-bd0c-240ed797412d {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.707069] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d6d5377-5a3e-41e2-847b-c7efec1bef5f {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.714233] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d432818-9499-49cb-bb20-e803f2751fc7 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.726905] env[61923]: DEBUG nova.compute.provider_tree [None req-84927854-5e83-4111-abce-383944a18027 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1035.822225] env[61923]: DEBUG oslo_vmware.api [None req-d17d2b02-87ac-45f0-981e-237fcddcf45c tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Task: {'id': task-1377966, 'name': Rename_Task, 'duration_secs': 1.137594} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1035.822635] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-d17d2b02-87ac-45f0-981e-237fcddcf45c tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] [instance: 35bdbd45-5838-4b09-a232-7f4a02322ad7] Powering on the VM {{(pid=61923) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1035.822834] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-4206d6f0-c747-4490-9a44-64d4ead55652 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.829555] env[61923]: DEBUG oslo_vmware.api [None req-d17d2b02-87ac-45f0-981e-237fcddcf45c tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Waiting for the task: (returnval){ [ 1035.829555] env[61923]: value = "task-1377967" [ 1035.829555] env[61923]: _type = "Task" [ 1035.829555] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1035.838732] env[61923]: DEBUG oslo_vmware.api [None req-d17d2b02-87ac-45f0-981e-237fcddcf45c tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Task: {'id': task-1377967, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1035.873881] env[61923]: DEBUG nova.network.neutron [-] [instance: a2f97947-6506-40b3-a24a-4ac03b88c5f5] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1036.230766] env[61923]: DEBUG nova.scheduler.client.report [None req-84927854-5e83-4111-abce-383944a18027 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1036.338923] env[61923]: DEBUG oslo_vmware.api [None req-d17d2b02-87ac-45f0-981e-237fcddcf45c tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Task: {'id': task-1377967, 'name': PowerOnVM_Task, 'duration_secs': 0.426612} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1036.339355] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-d17d2b02-87ac-45f0-981e-237fcddcf45c tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] [instance: 35bdbd45-5838-4b09-a232-7f4a02322ad7] Powered on the VM {{(pid=61923) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1036.339462] env[61923]: INFO nova.compute.manager [None req-d17d2b02-87ac-45f0-981e-237fcddcf45c tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] [instance: 35bdbd45-5838-4b09-a232-7f4a02322ad7] Took 8.69 seconds to spawn the instance on the hypervisor. [ 1036.339621] env[61923]: DEBUG nova.compute.manager [None req-d17d2b02-87ac-45f0-981e-237fcddcf45c tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] [instance: 35bdbd45-5838-4b09-a232-7f4a02322ad7] Checking state {{(pid=61923) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1036.340489] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5495964c-57c4-4d6b-9829-db1a7013ae1b {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.376586] env[61923]: INFO nova.compute.manager [-] [instance: a2f97947-6506-40b3-a24a-4ac03b88c5f5] Took 1.22 seconds to deallocate network for instance. [ 1036.735786] env[61923]: DEBUG oslo_concurrency.lockutils [None req-84927854-5e83-4111-abce-383944a18027 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.132s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1036.736341] env[61923]: DEBUG nova.compute.manager [None req-84927854-5e83-4111-abce-383944a18027 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: 9aeac922-aac2-47fe-8c11-835bef75e55c] Start building networks asynchronously for instance. {{(pid=61923) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1036.738889] env[61923]: DEBUG oslo_concurrency.lockutils [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 2.301s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1036.739070] env[61923]: DEBUG oslo_concurrency.lockutils [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1036.739225] env[61923]: DEBUG nova.compute.resource_tracker [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61923) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1036.739877] env[61923]: DEBUG oslo_concurrency.lockutils [None req-1c23d35b-fb18-4b83-9ba6-8e1d103da1b8 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.940s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1036.740113] env[61923]: DEBUG nova.objects.instance [None req-1c23d35b-fb18-4b83-9ba6-8e1d103da1b8 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Lazy-loading 'resources' on Instance uuid 0c3788fc-6cce-4806-b4f8-84f84489fb93 {{(pid=61923) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1036.741719] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eeb9fe80-bda8-4203-88f0-354bf3f57e89 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.750421] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ddd2057-5f9e-4e92-bff9-244484bf84b9 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.764795] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fcd8ef34-f5c8-44fc-813c-66544acbffb9 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.772433] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aae01ce8-f078-4e67-84b9-29f567bae35b {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.801179] env[61923]: DEBUG nova.compute.resource_tracker [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181088MB free_disk=178GB free_vcpus=48 pci_devices=None {{(pid=61923) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1036.801366] env[61923]: DEBUG oslo_concurrency.lockutils [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1036.856722] env[61923]: INFO nova.compute.manager [None req-d17d2b02-87ac-45f0-981e-237fcddcf45c tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] [instance: 35bdbd45-5838-4b09-a232-7f4a02322ad7] Took 13.38 seconds to build instance. [ 1037.242741] env[61923]: DEBUG nova.compute.utils [None req-84927854-5e83-4111-abce-383944a18027 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Using /dev/sd instead of None {{(pid=61923) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1037.244147] env[61923]: DEBUG nova.compute.manager [None req-84927854-5e83-4111-abce-383944a18027 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: 9aeac922-aac2-47fe-8c11-835bef75e55c] Allocating IP information in the background. {{(pid=61923) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1037.244337] env[61923]: DEBUG nova.network.neutron [None req-84927854-5e83-4111-abce-383944a18027 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: 9aeac922-aac2-47fe-8c11-835bef75e55c] allocate_for_instance() {{(pid=61923) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1037.296260] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d434d3c-8b67-46d7-b5cd-51ec073e3591 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.299946] env[61923]: DEBUG nova.policy [None req-84927854-5e83-4111-abce-383944a18027 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6087a385c42e4383aa26a7f586cca87f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5a4db4f8aa194a219e80c538cc715d43', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61923) authorize /opt/stack/nova/nova/policy.py:201}} [ 1037.305809] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-328876bf-6f49-4386-9876-91f15fce312f {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.335442] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d825410-d5d5-40e0-a99c-a3e456bf22c6 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.342899] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a17b76fb-c14e-48c9-af20-caf2f6e3ac7c {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.355996] env[61923]: DEBUG nova.compute.provider_tree [None req-1c23d35b-fb18-4b83-9ba6-8e1d103da1b8 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1037.357507] env[61923]: DEBUG oslo_concurrency.lockutils [None req-d17d2b02-87ac-45f0-981e-237fcddcf45c tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Lock "35bdbd45-5838-4b09-a232-7f4a02322ad7" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.894s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1037.401852] env[61923]: DEBUG oslo_concurrency.lockutils [None req-ec355d09-be8b-4126-8e84-491179ed7cd9 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Lock "a2f97947-6506-40b3-a24a-4ac03b88c5f5" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 2.287s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1037.541646] env[61923]: DEBUG nova.network.neutron [None req-84927854-5e83-4111-abce-383944a18027 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: 9aeac922-aac2-47fe-8c11-835bef75e55c] Successfully created port: 7f812cb5-e162-4dc7-9683-39a4c4048623 {{(pid=61923) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1037.747521] env[61923]: DEBUG nova.compute.manager [None req-84927854-5e83-4111-abce-383944a18027 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: 9aeac922-aac2-47fe-8c11-835bef75e55c] Start building block device mappings for instance. {{(pid=61923) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1037.858982] env[61923]: DEBUG nova.scheduler.client.report [None req-1c23d35b-fb18-4b83-9ba6-8e1d103da1b8 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1037.908665] env[61923]: DEBUG nova.compute.manager [req-0455a57b-722a-47f3-8947-bbae65ea31b1 req-f81fabeb-73f7-4396-ad15-981a6e1ae6f4 service nova] [instance: 35bdbd45-5838-4b09-a232-7f4a02322ad7] Received event network-changed-42e3ee29-941c-43d0-a7f3-81fc7755f64b {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1037.908870] env[61923]: DEBUG nova.compute.manager [req-0455a57b-722a-47f3-8947-bbae65ea31b1 req-f81fabeb-73f7-4396-ad15-981a6e1ae6f4 service nova] [instance: 35bdbd45-5838-4b09-a232-7f4a02322ad7] Refreshing instance network info cache due to event network-changed-42e3ee29-941c-43d0-a7f3-81fc7755f64b. {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1037.909102] env[61923]: DEBUG oslo_concurrency.lockutils [req-0455a57b-722a-47f3-8947-bbae65ea31b1 req-f81fabeb-73f7-4396-ad15-981a6e1ae6f4 service nova] Acquiring lock "refresh_cache-35bdbd45-5838-4b09-a232-7f4a02322ad7" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1037.909248] env[61923]: DEBUG oslo_concurrency.lockutils [req-0455a57b-722a-47f3-8947-bbae65ea31b1 req-f81fabeb-73f7-4396-ad15-981a6e1ae6f4 service nova] Acquired lock "refresh_cache-35bdbd45-5838-4b09-a232-7f4a02322ad7" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1037.909430] env[61923]: DEBUG nova.network.neutron [req-0455a57b-722a-47f3-8947-bbae65ea31b1 req-f81fabeb-73f7-4396-ad15-981a6e1ae6f4 service nova] [instance: 35bdbd45-5838-4b09-a232-7f4a02322ad7] Refreshing network info cache for port 42e3ee29-941c-43d0-a7f3-81fc7755f64b {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1038.363960] env[61923]: DEBUG oslo_concurrency.lockutils [None req-1c23d35b-fb18-4b83-9ba6-8e1d103da1b8 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.624s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1038.366137] env[61923]: DEBUG oslo_concurrency.lockutils [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 1.565s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1038.392685] env[61923]: INFO nova.scheduler.client.report [None req-1c23d35b-fb18-4b83-9ba6-8e1d103da1b8 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Deleted allocations for instance 0c3788fc-6cce-4806-b4f8-84f84489fb93 [ 1038.629100] env[61923]: DEBUG nova.network.neutron [req-0455a57b-722a-47f3-8947-bbae65ea31b1 req-f81fabeb-73f7-4396-ad15-981a6e1ae6f4 service nova] [instance: 35bdbd45-5838-4b09-a232-7f4a02322ad7] Updated VIF entry in instance network info cache for port 42e3ee29-941c-43d0-a7f3-81fc7755f64b. {{(pid=61923) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1038.629492] env[61923]: DEBUG nova.network.neutron [req-0455a57b-722a-47f3-8947-bbae65ea31b1 req-f81fabeb-73f7-4396-ad15-981a6e1ae6f4 service nova] [instance: 35bdbd45-5838-4b09-a232-7f4a02322ad7] Updating instance_info_cache with network_info: [{"id": "42e3ee29-941c-43d0-a7f3-81fc7755f64b", "address": "fa:16:3e:88:34:3b", "network": {"id": "d2905082-7ca4-4f92-a142-bb85211bb7c4", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-885218622-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.159", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a13d827933fa4597984afdb91dbbdd39", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "03ac2c9c-6ad2-4a85-bfab-c7e336df859a", "external-id": "nsx-vlan-transportzone-379", "segmentation_id": 379, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap42e3ee29-94", "ovs_interfaceid": "42e3ee29-941c-43d0-a7f3-81fc7755f64b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1038.756386] env[61923]: DEBUG nova.compute.manager [None req-84927854-5e83-4111-abce-383944a18027 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: 9aeac922-aac2-47fe-8c11-835bef75e55c] Start spawning the instance on the hypervisor. {{(pid=61923) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1038.777669] env[61923]: DEBUG nova.virt.hardware [None req-84927854-5e83-4111-abce-383944a18027 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-29T20:07:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-29T20:07:35Z,direct_url=,disk_format='vmdk',id=0f153f63-ae0a-45b1-b7f5-7f9b673c947f,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='4e7b5e3b3c3443c8bd5c70f8a15739f5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-29T20:07:36Z,virtual_size=,visibility=), allow threads: False {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1038.777948] env[61923]: DEBUG nova.virt.hardware [None req-84927854-5e83-4111-abce-383944a18027 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Flavor limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1038.778131] env[61923]: DEBUG nova.virt.hardware [None req-84927854-5e83-4111-abce-383944a18027 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Image limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1038.778340] env[61923]: DEBUG nova.virt.hardware [None req-84927854-5e83-4111-abce-383944a18027 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Flavor pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1038.778496] env[61923]: DEBUG nova.virt.hardware [None req-84927854-5e83-4111-abce-383944a18027 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Image pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1038.778646] env[61923]: DEBUG nova.virt.hardware [None req-84927854-5e83-4111-abce-383944a18027 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1038.778873] env[61923]: DEBUG nova.virt.hardware [None req-84927854-5e83-4111-abce-383944a18027 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1038.779045] env[61923]: DEBUG nova.virt.hardware [None req-84927854-5e83-4111-abce-383944a18027 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1038.779234] env[61923]: DEBUG nova.virt.hardware [None req-84927854-5e83-4111-abce-383944a18027 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Got 1 possible topologies {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1038.779403] env[61923]: DEBUG nova.virt.hardware [None req-84927854-5e83-4111-abce-383944a18027 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1038.779576] env[61923]: DEBUG nova.virt.hardware [None req-84927854-5e83-4111-abce-383944a18027 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1038.780517] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c463e64-28c4-46ea-86cb-b9687e4e7074 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.789403] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc15d442-aa07-41f1-aba7-d1a704b42ea8 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.900870] env[61923]: DEBUG oslo_concurrency.lockutils [None req-1c23d35b-fb18-4b83-9ba6-8e1d103da1b8 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Lock "0c3788fc-6cce-4806-b4f8-84f84489fb93" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.995s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1038.917806] env[61923]: DEBUG nova.compute.manager [req-87f2b7d6-8b2b-4ced-aba6-ff61f717a821 req-37c68cb7-fb15-43d8-971f-8897a7c56450 service nova] [instance: 9aeac922-aac2-47fe-8c11-835bef75e55c] Received event network-vif-plugged-7f812cb5-e162-4dc7-9683-39a4c4048623 {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1038.918066] env[61923]: DEBUG oslo_concurrency.lockutils [req-87f2b7d6-8b2b-4ced-aba6-ff61f717a821 req-37c68cb7-fb15-43d8-971f-8897a7c56450 service nova] Acquiring lock "9aeac922-aac2-47fe-8c11-835bef75e55c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1038.918376] env[61923]: DEBUG oslo_concurrency.lockutils [req-87f2b7d6-8b2b-4ced-aba6-ff61f717a821 req-37c68cb7-fb15-43d8-971f-8897a7c56450 service nova] Lock "9aeac922-aac2-47fe-8c11-835bef75e55c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1038.918575] env[61923]: DEBUG oslo_concurrency.lockutils [req-87f2b7d6-8b2b-4ced-aba6-ff61f717a821 req-37c68cb7-fb15-43d8-971f-8897a7c56450 service nova] Lock "9aeac922-aac2-47fe-8c11-835bef75e55c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1038.918744] env[61923]: DEBUG nova.compute.manager [req-87f2b7d6-8b2b-4ced-aba6-ff61f717a821 req-37c68cb7-fb15-43d8-971f-8897a7c56450 service nova] [instance: 9aeac922-aac2-47fe-8c11-835bef75e55c] No waiting events found dispatching network-vif-plugged-7f812cb5-e162-4dc7-9683-39a4c4048623 {{(pid=61923) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1038.919074] env[61923]: WARNING nova.compute.manager [req-87f2b7d6-8b2b-4ced-aba6-ff61f717a821 req-37c68cb7-fb15-43d8-971f-8897a7c56450 service nova] [instance: 9aeac922-aac2-47fe-8c11-835bef75e55c] Received unexpected event network-vif-plugged-7f812cb5-e162-4dc7-9683-39a4c4048623 for instance with vm_state building and task_state spawning. [ 1038.999784] env[61923]: DEBUG oslo_concurrency.lockutils [None req-b47c5602-b8b6-48e8-b951-0e67592f330e tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Acquiring lock "eeb51744-e397-40f8-8e1a-813fa23ebe95" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1039.000287] env[61923]: DEBUG oslo_concurrency.lockutils [None req-b47c5602-b8b6-48e8-b951-0e67592f330e tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Lock "eeb51744-e397-40f8-8e1a-813fa23ebe95" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1039.132077] env[61923]: DEBUG oslo_concurrency.lockutils [req-0455a57b-722a-47f3-8947-bbae65ea31b1 req-f81fabeb-73f7-4396-ad15-981a6e1ae6f4 service nova] Releasing lock "refresh_cache-35bdbd45-5838-4b09-a232-7f4a02322ad7" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1039.394305] env[61923]: DEBUG nova.compute.resource_tracker [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Instance 35bdbd45-5838-4b09-a232-7f4a02322ad7 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61923) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1039.394683] env[61923]: DEBUG nova.compute.resource_tracker [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Instance 9aeac922-aac2-47fe-8c11-835bef75e55c actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61923) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1039.472192] env[61923]: DEBUG nova.network.neutron [None req-84927854-5e83-4111-abce-383944a18027 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: 9aeac922-aac2-47fe-8c11-835bef75e55c] Successfully updated port: 7f812cb5-e162-4dc7-9683-39a4c4048623 {{(pid=61923) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1039.493796] env[61923]: DEBUG nova.compute.manager [req-950de443-feb7-4588-84dd-1ed24a0819d2 req-48d6ee00-2104-4bd5-bbe3-d24c27769665 service nova] [instance: 9aeac922-aac2-47fe-8c11-835bef75e55c] Received event network-changed-7f812cb5-e162-4dc7-9683-39a4c4048623 {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1039.494217] env[61923]: DEBUG nova.compute.manager [req-950de443-feb7-4588-84dd-1ed24a0819d2 req-48d6ee00-2104-4bd5-bbe3-d24c27769665 service nova] [instance: 9aeac922-aac2-47fe-8c11-835bef75e55c] Refreshing instance network info cache due to event network-changed-7f812cb5-e162-4dc7-9683-39a4c4048623. {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1039.494334] env[61923]: DEBUG oslo_concurrency.lockutils [req-950de443-feb7-4588-84dd-1ed24a0819d2 req-48d6ee00-2104-4bd5-bbe3-d24c27769665 service nova] Acquiring lock "refresh_cache-9aeac922-aac2-47fe-8c11-835bef75e55c" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1039.494397] env[61923]: DEBUG oslo_concurrency.lockutils [req-950de443-feb7-4588-84dd-1ed24a0819d2 req-48d6ee00-2104-4bd5-bbe3-d24c27769665 service nova] Acquired lock "refresh_cache-9aeac922-aac2-47fe-8c11-835bef75e55c" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1039.494606] env[61923]: DEBUG nova.network.neutron [req-950de443-feb7-4588-84dd-1ed24a0819d2 req-48d6ee00-2104-4bd5-bbe3-d24c27769665 service nova] [instance: 9aeac922-aac2-47fe-8c11-835bef75e55c] Refreshing network info cache for port 7f812cb5-e162-4dc7-9683-39a4c4048623 {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1039.503094] env[61923]: DEBUG nova.compute.manager [None req-b47c5602-b8b6-48e8-b951-0e67592f330e tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: eeb51744-e397-40f8-8e1a-813fa23ebe95] Starting instance... {{(pid=61923) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1039.897576] env[61923]: DEBUG nova.compute.resource_tracker [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Instance eeb51744-e397-40f8-8e1a-813fa23ebe95 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61923) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 1039.897881] env[61923]: DEBUG nova.compute.resource_tracker [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Total usable vcpus: 48, total allocated vcpus: 2 {{(pid=61923) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1039.897978] env[61923]: DEBUG nova.compute.resource_tracker [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=896MB phys_disk=200GB used_disk=2GB total_vcpus=48 used_vcpus=2 pci_stats=[] {{(pid=61923) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1039.951175] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8d6769f-a454-422e-80dd-b955999b4a9b {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.959923] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5bbdee3-68fa-444a-9ed9-7c493e7d2ce6 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.990531] env[61923]: DEBUG oslo_concurrency.lockutils [None req-84927854-5e83-4111-abce-383944a18027 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Acquiring lock "refresh_cache-9aeac922-aac2-47fe-8c11-835bef75e55c" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1039.991563] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-326a751d-b9f1-42c6-a27b-294dd94da901 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.001680] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d05ff1a0-3152-44b2-8dfb-99529f9263bf {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.017302] env[61923]: DEBUG nova.compute.provider_tree [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1040.027761] env[61923]: DEBUG nova.network.neutron [req-950de443-feb7-4588-84dd-1ed24a0819d2 req-48d6ee00-2104-4bd5-bbe3-d24c27769665 service nova] [instance: 9aeac922-aac2-47fe-8c11-835bef75e55c] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1040.030624] env[61923]: DEBUG oslo_concurrency.lockutils [None req-b47c5602-b8b6-48e8-b951-0e67592f330e tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1040.109550] env[61923]: DEBUG nova.network.neutron [req-950de443-feb7-4588-84dd-1ed24a0819d2 req-48d6ee00-2104-4bd5-bbe3-d24c27769665 service nova] [instance: 9aeac922-aac2-47fe-8c11-835bef75e55c] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1040.521903] env[61923]: DEBUG nova.scheduler.client.report [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1040.612667] env[61923]: DEBUG oslo_concurrency.lockutils [req-950de443-feb7-4588-84dd-1ed24a0819d2 req-48d6ee00-2104-4bd5-bbe3-d24c27769665 service nova] Releasing lock "refresh_cache-9aeac922-aac2-47fe-8c11-835bef75e55c" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1040.613013] env[61923]: DEBUG oslo_concurrency.lockutils [None req-84927854-5e83-4111-abce-383944a18027 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Acquired lock "refresh_cache-9aeac922-aac2-47fe-8c11-835bef75e55c" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1040.613181] env[61923]: DEBUG nova.network.neutron [None req-84927854-5e83-4111-abce-383944a18027 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: 9aeac922-aac2-47fe-8c11-835bef75e55c] Building network info cache for instance {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1041.028630] env[61923]: DEBUG nova.compute.resource_tracker [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61923) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1041.028840] env[61923]: DEBUG oslo_concurrency.lockutils [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.663s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1041.029134] env[61923]: DEBUG oslo_concurrency.lockutils [None req-b47c5602-b8b6-48e8-b951-0e67592f330e tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.999s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1041.030723] env[61923]: INFO nova.compute.claims [None req-b47c5602-b8b6-48e8-b951-0e67592f330e tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: eeb51744-e397-40f8-8e1a-813fa23ebe95] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1041.142360] env[61923]: DEBUG nova.network.neutron [None req-84927854-5e83-4111-abce-383944a18027 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: 9aeac922-aac2-47fe-8c11-835bef75e55c] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1041.262737] env[61923]: DEBUG nova.network.neutron [None req-84927854-5e83-4111-abce-383944a18027 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: 9aeac922-aac2-47fe-8c11-835bef75e55c] Updating instance_info_cache with network_info: [{"id": "7f812cb5-e162-4dc7-9683-39a4c4048623", "address": "fa:16:3e:00:7d:78", "network": {"id": "77aec6ca-d1be-4b39-9d51-08853a1183a1", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1813085414-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5a4db4f8aa194a219e80c538cc715d43", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "56b944d8-803d-43f2-945d-0f334ee4ea1c", "external-id": "nsx-vlan-transportzone-799", "segmentation_id": 799, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7f812cb5-e1", "ovs_interfaceid": "7f812cb5-e162-4dc7-9683-39a4c4048623", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1041.765146] env[61923]: DEBUG oslo_concurrency.lockutils [None req-84927854-5e83-4111-abce-383944a18027 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Releasing lock "refresh_cache-9aeac922-aac2-47fe-8c11-835bef75e55c" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1041.765504] env[61923]: DEBUG nova.compute.manager [None req-84927854-5e83-4111-abce-383944a18027 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: 9aeac922-aac2-47fe-8c11-835bef75e55c] Instance network_info: |[{"id": "7f812cb5-e162-4dc7-9683-39a4c4048623", "address": "fa:16:3e:00:7d:78", "network": {"id": "77aec6ca-d1be-4b39-9d51-08853a1183a1", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1813085414-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5a4db4f8aa194a219e80c538cc715d43", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "56b944d8-803d-43f2-945d-0f334ee4ea1c", "external-id": "nsx-vlan-transportzone-799", "segmentation_id": 799, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7f812cb5-e1", "ovs_interfaceid": "7f812cb5-e162-4dc7-9683-39a4c4048623", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61923) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1041.765869] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-84927854-5e83-4111-abce-383944a18027 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: 9aeac922-aac2-47fe-8c11-835bef75e55c] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:00:7d:78', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '56b944d8-803d-43f2-945d-0f334ee4ea1c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '7f812cb5-e162-4dc7-9683-39a4c4048623', 'vif_model': 'vmxnet3'}] {{(pid=61923) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1041.773215] env[61923]: DEBUG oslo.service.loopingcall [None req-84927854-5e83-4111-abce-383944a18027 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61923) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1041.773420] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9aeac922-aac2-47fe-8c11-835bef75e55c] Creating VM on the ESX host {{(pid=61923) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1041.773643] env[61923]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-6b5d596e-a4fd-499c-8ba8-66e6b5ee3ebb {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.793111] env[61923]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1041.793111] env[61923]: value = "task-1377968" [ 1041.793111] env[61923]: _type = "Task" [ 1041.793111] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1041.801734] env[61923]: DEBUG oslo_vmware.api [-] Task: {'id': task-1377968, 'name': CreateVM_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1042.084033] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79f3f982-257c-48ea-90f8-4174996e6446 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.091887] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b26b008d-b54c-476b-962c-57e5c86de332 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.120631] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5a7d46a-5d47-4060-8872-418f7bde2734 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.127372] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50f6a435-7c54-4147-bcc5-365a75112041 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.140227] env[61923]: DEBUG nova.compute.provider_tree [None req-b47c5602-b8b6-48e8-b951-0e67592f330e tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1042.303596] env[61923]: DEBUG oslo_vmware.api [-] Task: {'id': task-1377968, 'name': CreateVM_Task} progress is 25%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1042.644065] env[61923]: DEBUG nova.scheduler.client.report [None req-b47c5602-b8b6-48e8-b951-0e67592f330e tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1042.802678] env[61923]: DEBUG oslo_vmware.api [-] Task: {'id': task-1377968, 'name': CreateVM_Task, 'duration_secs': 0.765438} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1042.803157] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9aeac922-aac2-47fe-8c11-835bef75e55c] Created VM on the ESX host {{(pid=61923) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1042.803522] env[61923]: DEBUG oslo_concurrency.lockutils [None req-84927854-5e83-4111-abce-383944a18027 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1042.803691] env[61923]: DEBUG oslo_concurrency.lockutils [None req-84927854-5e83-4111-abce-383944a18027 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1042.804020] env[61923]: DEBUG oslo_concurrency.lockutils [None req-84927854-5e83-4111-abce-383944a18027 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1042.804262] env[61923]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-66249134-d4b2-4d19-b3a7-b36c81d4a949 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.808330] env[61923]: DEBUG oslo_vmware.api [None req-84927854-5e83-4111-abce-383944a18027 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Waiting for the task: (returnval){ [ 1042.808330] env[61923]: value = "session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52033d54-2228-d2f6-0a28-c495e1ed4a7b" [ 1042.808330] env[61923]: _type = "Task" [ 1042.808330] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1042.815229] env[61923]: DEBUG oslo_vmware.api [None req-84927854-5e83-4111-abce-383944a18027 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52033d54-2228-d2f6-0a28-c495e1ed4a7b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1043.149059] env[61923]: DEBUG oslo_concurrency.lockutils [None req-b47c5602-b8b6-48e8-b951-0e67592f330e tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.120s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1043.149623] env[61923]: DEBUG nova.compute.manager [None req-b47c5602-b8b6-48e8-b951-0e67592f330e tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: eeb51744-e397-40f8-8e1a-813fa23ebe95] Start building networks asynchronously for instance. {{(pid=61923) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1043.319385] env[61923]: DEBUG oslo_vmware.api [None req-84927854-5e83-4111-abce-383944a18027 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52033d54-2228-d2f6-0a28-c495e1ed4a7b, 'name': SearchDatastore_Task, 'duration_secs': 0.010767} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1043.319686] env[61923]: DEBUG oslo_concurrency.lockutils [None req-84927854-5e83-4111-abce-383944a18027 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1043.319933] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-84927854-5e83-4111-abce-383944a18027 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: 9aeac922-aac2-47fe-8c11-835bef75e55c] Processing image 0f153f63-ae0a-45b1-b7f5-7f9b673c947f {{(pid=61923) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1043.320181] env[61923]: DEBUG oslo_concurrency.lockutils [None req-84927854-5e83-4111-abce-383944a18027 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1043.320355] env[61923]: DEBUG oslo_concurrency.lockutils [None req-84927854-5e83-4111-abce-383944a18027 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1043.320545] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-84927854-5e83-4111-abce-383944a18027 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61923) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1043.320804] env[61923]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-5e100975-476c-4437-b5bd-8ca96377bd53 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.328252] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-84927854-5e83-4111-abce-383944a18027 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61923) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1043.328423] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-84927854-5e83-4111-abce-383944a18027 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61923) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1043.329094] env[61923]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-505c6c34-9a81-4885-982c-d2f0a5634e74 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.333688] env[61923]: DEBUG oslo_vmware.api [None req-84927854-5e83-4111-abce-383944a18027 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Waiting for the task: (returnval){ [ 1043.333688] env[61923]: value = "session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]527be0b4-a6ad-4203-f06a-db545db367d5" [ 1043.333688] env[61923]: _type = "Task" [ 1043.333688] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1043.340569] env[61923]: DEBUG oslo_vmware.api [None req-84927854-5e83-4111-abce-383944a18027 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]527be0b4-a6ad-4203-f06a-db545db367d5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1043.654433] env[61923]: DEBUG nova.compute.utils [None req-b47c5602-b8b6-48e8-b951-0e67592f330e tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Using /dev/sd instead of None {{(pid=61923) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1043.655962] env[61923]: DEBUG nova.compute.manager [None req-b47c5602-b8b6-48e8-b951-0e67592f330e tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: eeb51744-e397-40f8-8e1a-813fa23ebe95] Allocating IP information in the background. {{(pid=61923) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1043.656131] env[61923]: DEBUG nova.network.neutron [None req-b47c5602-b8b6-48e8-b951-0e67592f330e tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: eeb51744-e397-40f8-8e1a-813fa23ebe95] allocate_for_instance() {{(pid=61923) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1043.694561] env[61923]: DEBUG nova.policy [None req-b47c5602-b8b6-48e8-b951-0e67592f330e tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '69be89006ff141b686261d643b3b9a73', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '780d3541d9604417b977bb62390c4299', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61923) authorize /opt/stack/nova/nova/policy.py:201}} [ 1043.843809] env[61923]: DEBUG oslo_vmware.api [None req-84927854-5e83-4111-abce-383944a18027 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]527be0b4-a6ad-4203-f06a-db545db367d5, 'name': SearchDatastore_Task, 'duration_secs': 0.007109} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1043.844589] env[61923]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-283f802a-7023-47c8-ab38-e9098d633fe9 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.850179] env[61923]: DEBUG oslo_vmware.api [None req-84927854-5e83-4111-abce-383944a18027 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Waiting for the task: (returnval){ [ 1043.850179] env[61923]: value = "session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52bb76a2-ef97-4d9d-c619-7ab6afad8888" [ 1043.850179] env[61923]: _type = "Task" [ 1043.850179] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1043.857928] env[61923]: DEBUG oslo_vmware.api [None req-84927854-5e83-4111-abce-383944a18027 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52bb76a2-ef97-4d9d-c619-7ab6afad8888, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1044.136738] env[61923]: DEBUG nova.network.neutron [None req-b47c5602-b8b6-48e8-b951-0e67592f330e tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: eeb51744-e397-40f8-8e1a-813fa23ebe95] Successfully created port: 494148c6-b0ca-4198-bd04-465adbaffd47 {{(pid=61923) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1044.159017] env[61923]: DEBUG nova.compute.manager [None req-b47c5602-b8b6-48e8-b951-0e67592f330e tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: eeb51744-e397-40f8-8e1a-813fa23ebe95] Start building block device mappings for instance. {{(pid=61923) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1044.359867] env[61923]: DEBUG oslo_vmware.api [None req-84927854-5e83-4111-abce-383944a18027 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52bb76a2-ef97-4d9d-c619-7ab6afad8888, 'name': SearchDatastore_Task, 'duration_secs': 0.00933} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1044.360170] env[61923]: DEBUG oslo_concurrency.lockutils [None req-84927854-5e83-4111-abce-383944a18027 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1044.360465] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-84927854-5e83-4111-abce-383944a18027 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk to [datastore1] 9aeac922-aac2-47fe-8c11-835bef75e55c/9aeac922-aac2-47fe-8c11-835bef75e55c.vmdk {{(pid=61923) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1044.360722] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-1b076494-e6d5-4fad-bc9a-4bd5d67ee94f {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.367165] env[61923]: DEBUG oslo_vmware.api [None req-84927854-5e83-4111-abce-383944a18027 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Waiting for the task: (returnval){ [ 1044.367165] env[61923]: value = "task-1377969" [ 1044.367165] env[61923]: _type = "Task" [ 1044.367165] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1044.374353] env[61923]: DEBUG oslo_vmware.api [None req-84927854-5e83-4111-abce-383944a18027 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Task: {'id': task-1377969, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1044.879395] env[61923]: DEBUG oslo_vmware.api [None req-84927854-5e83-4111-abce-383944a18027 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Task: {'id': task-1377969, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.416687} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1044.882613] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-84927854-5e83-4111-abce-383944a18027 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk to [datastore1] 9aeac922-aac2-47fe-8c11-835bef75e55c/9aeac922-aac2-47fe-8c11-835bef75e55c.vmdk {{(pid=61923) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1044.882613] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-84927854-5e83-4111-abce-383944a18027 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: 9aeac922-aac2-47fe-8c11-835bef75e55c] Extending root virtual disk to 1048576 {{(pid=61923) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1044.882613] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-57b286d4-f227-43d8-9f7b-5fbdd850a598 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.887377] env[61923]: DEBUG oslo_vmware.api [None req-84927854-5e83-4111-abce-383944a18027 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Waiting for the task: (returnval){ [ 1044.887377] env[61923]: value = "task-1377970" [ 1044.887377] env[61923]: _type = "Task" [ 1044.887377] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1044.894696] env[61923]: DEBUG oslo_vmware.api [None req-84927854-5e83-4111-abce-383944a18027 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Task: {'id': task-1377970, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1045.168856] env[61923]: DEBUG nova.compute.manager [None req-b47c5602-b8b6-48e8-b951-0e67592f330e tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: eeb51744-e397-40f8-8e1a-813fa23ebe95] Start spawning the instance on the hypervisor. {{(pid=61923) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1045.193848] env[61923]: DEBUG nova.virt.hardware [None req-b47c5602-b8b6-48e8-b951-0e67592f330e tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-29T20:07:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-29T20:07:35Z,direct_url=,disk_format='vmdk',id=0f153f63-ae0a-45b1-b7f5-7f9b673c947f,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='4e7b5e3b3c3443c8bd5c70f8a15739f5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-29T20:07:36Z,virtual_size=,visibility=), allow threads: False {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1045.194114] env[61923]: DEBUG nova.virt.hardware [None req-b47c5602-b8b6-48e8-b951-0e67592f330e tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Flavor limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1045.194276] env[61923]: DEBUG nova.virt.hardware [None req-b47c5602-b8b6-48e8-b951-0e67592f330e tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Image limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1045.194474] env[61923]: DEBUG nova.virt.hardware [None req-b47c5602-b8b6-48e8-b951-0e67592f330e tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Flavor pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1045.194620] env[61923]: DEBUG nova.virt.hardware [None req-b47c5602-b8b6-48e8-b951-0e67592f330e tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Image pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1045.194768] env[61923]: DEBUG nova.virt.hardware [None req-b47c5602-b8b6-48e8-b951-0e67592f330e tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1045.194975] env[61923]: DEBUG nova.virt.hardware [None req-b47c5602-b8b6-48e8-b951-0e67592f330e tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1045.195162] env[61923]: DEBUG nova.virt.hardware [None req-b47c5602-b8b6-48e8-b951-0e67592f330e tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1045.195325] env[61923]: DEBUG nova.virt.hardware [None req-b47c5602-b8b6-48e8-b951-0e67592f330e tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Got 1 possible topologies {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1045.195488] env[61923]: DEBUG nova.virt.hardware [None req-b47c5602-b8b6-48e8-b951-0e67592f330e tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1045.195659] env[61923]: DEBUG nova.virt.hardware [None req-b47c5602-b8b6-48e8-b951-0e67592f330e tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1045.196535] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1bb1666-5daa-4b09-9649-9b4d19d64544 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.204441] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84377f5c-0653-4374-964e-2a8399e8f81a {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.396911] env[61923]: DEBUG oslo_vmware.api [None req-84927854-5e83-4111-abce-383944a18027 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Task: {'id': task-1377970, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.064227} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1045.397221] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-84927854-5e83-4111-abce-383944a18027 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: 9aeac922-aac2-47fe-8c11-835bef75e55c] Extended root virtual disk {{(pid=61923) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1045.397976] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b3cb2ad-61f5-444f-96e5-2a19cb33dcff {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.419665] env[61923]: DEBUG nova.virt.vmwareapi.volumeops [None req-84927854-5e83-4111-abce-383944a18027 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: 9aeac922-aac2-47fe-8c11-835bef75e55c] Reconfiguring VM instance instance-00000061 to attach disk [datastore1] 9aeac922-aac2-47fe-8c11-835bef75e55c/9aeac922-aac2-47fe-8c11-835bef75e55c.vmdk or device None with type sparse {{(pid=61923) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1045.419940] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b354056e-928d-42dd-93bf-19fd3d30697d {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.438790] env[61923]: DEBUG oslo_vmware.api [None req-84927854-5e83-4111-abce-383944a18027 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Waiting for the task: (returnval){ [ 1045.438790] env[61923]: value = "task-1377971" [ 1045.438790] env[61923]: _type = "Task" [ 1045.438790] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1045.448132] env[61923]: DEBUG oslo_vmware.api [None req-84927854-5e83-4111-abce-383944a18027 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Task: {'id': task-1377971, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1045.499531] env[61923]: DEBUG nova.compute.manager [req-1e04791b-6359-4862-9a37-d3d3a620c0a2 req-373d6c28-196b-48ab-9b97-af61cec32d7d service nova] [instance: eeb51744-e397-40f8-8e1a-813fa23ebe95] Received event network-vif-plugged-494148c6-b0ca-4198-bd04-465adbaffd47 {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1045.499810] env[61923]: DEBUG oslo_concurrency.lockutils [req-1e04791b-6359-4862-9a37-d3d3a620c0a2 req-373d6c28-196b-48ab-9b97-af61cec32d7d service nova] Acquiring lock "eeb51744-e397-40f8-8e1a-813fa23ebe95-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1045.500042] env[61923]: DEBUG oslo_concurrency.lockutils [req-1e04791b-6359-4862-9a37-d3d3a620c0a2 req-373d6c28-196b-48ab-9b97-af61cec32d7d service nova] Lock "eeb51744-e397-40f8-8e1a-813fa23ebe95-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1045.500222] env[61923]: DEBUG oslo_concurrency.lockutils [req-1e04791b-6359-4862-9a37-d3d3a620c0a2 req-373d6c28-196b-48ab-9b97-af61cec32d7d service nova] Lock "eeb51744-e397-40f8-8e1a-813fa23ebe95-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1045.500457] env[61923]: DEBUG nova.compute.manager [req-1e04791b-6359-4862-9a37-d3d3a620c0a2 req-373d6c28-196b-48ab-9b97-af61cec32d7d service nova] [instance: eeb51744-e397-40f8-8e1a-813fa23ebe95] No waiting events found dispatching network-vif-plugged-494148c6-b0ca-4198-bd04-465adbaffd47 {{(pid=61923) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1045.500636] env[61923]: WARNING nova.compute.manager [req-1e04791b-6359-4862-9a37-d3d3a620c0a2 req-373d6c28-196b-48ab-9b97-af61cec32d7d service nova] [instance: eeb51744-e397-40f8-8e1a-813fa23ebe95] Received unexpected event network-vif-plugged-494148c6-b0ca-4198-bd04-465adbaffd47 for instance with vm_state building and task_state spawning. [ 1045.579190] env[61923]: DEBUG nova.network.neutron [None req-b47c5602-b8b6-48e8-b951-0e67592f330e tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: eeb51744-e397-40f8-8e1a-813fa23ebe95] Successfully updated port: 494148c6-b0ca-4198-bd04-465adbaffd47 {{(pid=61923) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1045.949261] env[61923]: DEBUG oslo_vmware.api [None req-84927854-5e83-4111-abce-383944a18027 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Task: {'id': task-1377971, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1046.082234] env[61923]: DEBUG oslo_concurrency.lockutils [None req-b47c5602-b8b6-48e8-b951-0e67592f330e tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Acquiring lock "refresh_cache-eeb51744-e397-40f8-8e1a-813fa23ebe95" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1046.082408] env[61923]: DEBUG oslo_concurrency.lockutils [None req-b47c5602-b8b6-48e8-b951-0e67592f330e tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Acquired lock "refresh_cache-eeb51744-e397-40f8-8e1a-813fa23ebe95" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1046.082531] env[61923]: DEBUG nova.network.neutron [None req-b47c5602-b8b6-48e8-b951-0e67592f330e tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: eeb51744-e397-40f8-8e1a-813fa23ebe95] Building network info cache for instance {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1046.448980] env[61923]: DEBUG oslo_vmware.api [None req-84927854-5e83-4111-abce-383944a18027 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Task: {'id': task-1377971, 'name': ReconfigVM_Task, 'duration_secs': 0.8061} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1046.451819] env[61923]: DEBUG nova.virt.vmwareapi.volumeops [None req-84927854-5e83-4111-abce-383944a18027 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: 9aeac922-aac2-47fe-8c11-835bef75e55c] Reconfigured VM instance instance-00000061 to attach disk [datastore1] 9aeac922-aac2-47fe-8c11-835bef75e55c/9aeac922-aac2-47fe-8c11-835bef75e55c.vmdk or device None with type sparse {{(pid=61923) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1046.451819] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-5d333f51-235b-4293-8820-a00b0cf0af54 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.457050] env[61923]: DEBUG oslo_vmware.api [None req-84927854-5e83-4111-abce-383944a18027 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Waiting for the task: (returnval){ [ 1046.457050] env[61923]: value = "task-1377972" [ 1046.457050] env[61923]: _type = "Task" [ 1046.457050] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1046.463891] env[61923]: DEBUG oslo_vmware.api [None req-84927854-5e83-4111-abce-383944a18027 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Task: {'id': task-1377972, 'name': Rename_Task} progress is 5%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1046.614288] env[61923]: DEBUG nova.network.neutron [None req-b47c5602-b8b6-48e8-b951-0e67592f330e tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: eeb51744-e397-40f8-8e1a-813fa23ebe95] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1046.741976] env[61923]: DEBUG nova.network.neutron [None req-b47c5602-b8b6-48e8-b951-0e67592f330e tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: eeb51744-e397-40f8-8e1a-813fa23ebe95] Updating instance_info_cache with network_info: [{"id": "494148c6-b0ca-4198-bd04-465adbaffd47", "address": "fa:16:3e:6f:11:63", "network": {"id": "9533e79d-fde2-4c10-86a0-86a36845a8cf", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1736743137-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "780d3541d9604417b977bb62390c4299", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6d650b26-c3e7-4de7-98db-5e4b816d123a", "external-id": "nsx-vlan-transportzone-757", "segmentation_id": 757, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap494148c6-b0", "ovs_interfaceid": "494148c6-b0ca-4198-bd04-465adbaffd47", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1046.966893] env[61923]: DEBUG oslo_vmware.api [None req-84927854-5e83-4111-abce-383944a18027 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Task: {'id': task-1377972, 'name': Rename_Task, 'duration_secs': 0.128127} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1046.967288] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-84927854-5e83-4111-abce-383944a18027 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: 9aeac922-aac2-47fe-8c11-835bef75e55c] Powering on the VM {{(pid=61923) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1046.967476] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-dc758591-e18a-44a6-b92c-c825a7dc01bb {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.973082] env[61923]: DEBUG oslo_vmware.api [None req-84927854-5e83-4111-abce-383944a18027 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Waiting for the task: (returnval){ [ 1046.973082] env[61923]: value = "task-1377973" [ 1046.973082] env[61923]: _type = "Task" [ 1046.973082] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1046.980497] env[61923]: DEBUG oslo_vmware.api [None req-84927854-5e83-4111-abce-383944a18027 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Task: {'id': task-1377973, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1047.244784] env[61923]: DEBUG oslo_concurrency.lockutils [None req-b47c5602-b8b6-48e8-b951-0e67592f330e tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Releasing lock "refresh_cache-eeb51744-e397-40f8-8e1a-813fa23ebe95" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1047.245094] env[61923]: DEBUG nova.compute.manager [None req-b47c5602-b8b6-48e8-b951-0e67592f330e tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: eeb51744-e397-40f8-8e1a-813fa23ebe95] Instance network_info: |[{"id": "494148c6-b0ca-4198-bd04-465adbaffd47", "address": "fa:16:3e:6f:11:63", "network": {"id": "9533e79d-fde2-4c10-86a0-86a36845a8cf", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1736743137-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "780d3541d9604417b977bb62390c4299", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6d650b26-c3e7-4de7-98db-5e4b816d123a", "external-id": "nsx-vlan-transportzone-757", "segmentation_id": 757, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap494148c6-b0", "ovs_interfaceid": "494148c6-b0ca-4198-bd04-465adbaffd47", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61923) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1047.245559] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-b47c5602-b8b6-48e8-b951-0e67592f330e tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: eeb51744-e397-40f8-8e1a-813fa23ebe95] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:6f:11:63', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '6d650b26-c3e7-4de7-98db-5e4b816d123a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '494148c6-b0ca-4198-bd04-465adbaffd47', 'vif_model': 'vmxnet3'}] {{(pid=61923) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1047.253049] env[61923]: DEBUG oslo.service.loopingcall [None req-b47c5602-b8b6-48e8-b951-0e67592f330e tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61923) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1047.253273] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: eeb51744-e397-40f8-8e1a-813fa23ebe95] Creating VM on the ESX host {{(pid=61923) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1047.253499] env[61923]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-15bbf470-0da1-4ab7-8db8-f24e0325deb8 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.272908] env[61923]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1047.272908] env[61923]: value = "task-1377974" [ 1047.272908] env[61923]: _type = "Task" [ 1047.272908] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1047.279841] env[61923]: DEBUG oslo_vmware.api [-] Task: {'id': task-1377974, 'name': CreateVM_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1047.483429] env[61923]: DEBUG oslo_vmware.api [None req-84927854-5e83-4111-abce-383944a18027 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Task: {'id': task-1377973, 'name': PowerOnVM_Task, 'duration_secs': 0.437643} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1047.483750] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-84927854-5e83-4111-abce-383944a18027 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: 9aeac922-aac2-47fe-8c11-835bef75e55c] Powered on the VM {{(pid=61923) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1047.483970] env[61923]: INFO nova.compute.manager [None req-84927854-5e83-4111-abce-383944a18027 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: 9aeac922-aac2-47fe-8c11-835bef75e55c] Took 8.73 seconds to spawn the instance on the hypervisor. [ 1047.484175] env[61923]: DEBUG nova.compute.manager [None req-84927854-5e83-4111-abce-383944a18027 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: 9aeac922-aac2-47fe-8c11-835bef75e55c] Checking state {{(pid=61923) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1047.484988] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf4f72df-e9b1-4466-846a-1cc04370ff9c {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.526174] env[61923]: DEBUG nova.compute.manager [req-930e91ef-615c-469f-9e80-b04673aac6e1 req-fbed52ec-aaf0-4472-9d1c-d2dd2b5b7c3b service nova] [instance: eeb51744-e397-40f8-8e1a-813fa23ebe95] Received event network-changed-494148c6-b0ca-4198-bd04-465adbaffd47 {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1047.526404] env[61923]: DEBUG nova.compute.manager [req-930e91ef-615c-469f-9e80-b04673aac6e1 req-fbed52ec-aaf0-4472-9d1c-d2dd2b5b7c3b service nova] [instance: eeb51744-e397-40f8-8e1a-813fa23ebe95] Refreshing instance network info cache due to event network-changed-494148c6-b0ca-4198-bd04-465adbaffd47. {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1047.526633] env[61923]: DEBUG oslo_concurrency.lockutils [req-930e91ef-615c-469f-9e80-b04673aac6e1 req-fbed52ec-aaf0-4472-9d1c-d2dd2b5b7c3b service nova] Acquiring lock "refresh_cache-eeb51744-e397-40f8-8e1a-813fa23ebe95" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1047.526796] env[61923]: DEBUG oslo_concurrency.lockutils [req-930e91ef-615c-469f-9e80-b04673aac6e1 req-fbed52ec-aaf0-4472-9d1c-d2dd2b5b7c3b service nova] Acquired lock "refresh_cache-eeb51744-e397-40f8-8e1a-813fa23ebe95" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1047.526985] env[61923]: DEBUG nova.network.neutron [req-930e91ef-615c-469f-9e80-b04673aac6e1 req-fbed52ec-aaf0-4472-9d1c-d2dd2b5b7c3b service nova] [instance: eeb51744-e397-40f8-8e1a-813fa23ebe95] Refreshing network info cache for port 494148c6-b0ca-4198-bd04-465adbaffd47 {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1047.783432] env[61923]: DEBUG oslo_vmware.api [-] Task: {'id': task-1377974, 'name': CreateVM_Task, 'duration_secs': 0.33357} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1047.783432] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: eeb51744-e397-40f8-8e1a-813fa23ebe95] Created VM on the ESX host {{(pid=61923) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1047.784049] env[61923]: DEBUG oslo_concurrency.lockutils [None req-b47c5602-b8b6-48e8-b951-0e67592f330e tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1047.784227] env[61923]: DEBUG oslo_concurrency.lockutils [None req-b47c5602-b8b6-48e8-b951-0e67592f330e tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1047.784547] env[61923]: DEBUG oslo_concurrency.lockutils [None req-b47c5602-b8b6-48e8-b951-0e67592f330e tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1047.784820] env[61923]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d0caa4de-2441-4544-ac6d-f72ee64df0cc {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.789300] env[61923]: DEBUG oslo_vmware.api [None req-b47c5602-b8b6-48e8-b951-0e67592f330e tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Waiting for the task: (returnval){ [ 1047.789300] env[61923]: value = "session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52c572e2-b737-e650-7db4-110e57218095" [ 1047.789300] env[61923]: _type = "Task" [ 1047.789300] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1047.796312] env[61923]: DEBUG oslo_vmware.api [None req-b47c5602-b8b6-48e8-b951-0e67592f330e tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52c572e2-b737-e650-7db4-110e57218095, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1048.002425] env[61923]: INFO nova.compute.manager [None req-84927854-5e83-4111-abce-383944a18027 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: 9aeac922-aac2-47fe-8c11-835bef75e55c] Took 14.63 seconds to build instance. [ 1048.213384] env[61923]: DEBUG nova.network.neutron [req-930e91ef-615c-469f-9e80-b04673aac6e1 req-fbed52ec-aaf0-4472-9d1c-d2dd2b5b7c3b service nova] [instance: eeb51744-e397-40f8-8e1a-813fa23ebe95] Updated VIF entry in instance network info cache for port 494148c6-b0ca-4198-bd04-465adbaffd47. {{(pid=61923) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1048.213735] env[61923]: DEBUG nova.network.neutron [req-930e91ef-615c-469f-9e80-b04673aac6e1 req-fbed52ec-aaf0-4472-9d1c-d2dd2b5b7c3b service nova] [instance: eeb51744-e397-40f8-8e1a-813fa23ebe95] Updating instance_info_cache with network_info: [{"id": "494148c6-b0ca-4198-bd04-465adbaffd47", "address": "fa:16:3e:6f:11:63", "network": {"id": "9533e79d-fde2-4c10-86a0-86a36845a8cf", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1736743137-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "780d3541d9604417b977bb62390c4299", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6d650b26-c3e7-4de7-98db-5e4b816d123a", "external-id": "nsx-vlan-transportzone-757", "segmentation_id": 757, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap494148c6-b0", "ovs_interfaceid": "494148c6-b0ca-4198-bd04-465adbaffd47", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1048.299255] env[61923]: DEBUG oslo_vmware.api [None req-b47c5602-b8b6-48e8-b951-0e67592f330e tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52c572e2-b737-e650-7db4-110e57218095, 'name': SearchDatastore_Task, 'duration_secs': 0.00845} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1048.299622] env[61923]: DEBUG oslo_concurrency.lockutils [None req-b47c5602-b8b6-48e8-b951-0e67592f330e tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1048.299835] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-b47c5602-b8b6-48e8-b951-0e67592f330e tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: eeb51744-e397-40f8-8e1a-813fa23ebe95] Processing image 0f153f63-ae0a-45b1-b7f5-7f9b673c947f {{(pid=61923) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1048.300082] env[61923]: DEBUG oslo_concurrency.lockutils [None req-b47c5602-b8b6-48e8-b951-0e67592f330e tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1048.300235] env[61923]: DEBUG oslo_concurrency.lockutils [None req-b47c5602-b8b6-48e8-b951-0e67592f330e tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1048.300446] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-b47c5602-b8b6-48e8-b951-0e67592f330e tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61923) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1048.300702] env[61923]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f22476b9-f097-4e00-af5f-e0bb7a887951 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.308307] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-b47c5602-b8b6-48e8-b951-0e67592f330e tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61923) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1048.308470] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-b47c5602-b8b6-48e8-b951-0e67592f330e tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61923) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1048.309150] env[61923]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6299fed3-9809-44fb-a25c-51dfd9eba97f {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.313904] env[61923]: DEBUG oslo_vmware.api [None req-b47c5602-b8b6-48e8-b951-0e67592f330e tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Waiting for the task: (returnval){ [ 1048.313904] env[61923]: value = "session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52e3ec56-82d8-756c-9c79-c57eae38a621" [ 1048.313904] env[61923]: _type = "Task" [ 1048.313904] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1048.320722] env[61923]: DEBUG oslo_vmware.api [None req-b47c5602-b8b6-48e8-b951-0e67592f330e tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52e3ec56-82d8-756c-9c79-c57eae38a621, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1048.504549] env[61923]: DEBUG oslo_concurrency.lockutils [None req-84927854-5e83-4111-abce-383944a18027 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Lock "9aeac922-aac2-47fe-8c11-835bef75e55c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 16.148s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1048.717342] env[61923]: DEBUG oslo_concurrency.lockutils [req-930e91ef-615c-469f-9e80-b04673aac6e1 req-fbed52ec-aaf0-4472-9d1c-d2dd2b5b7c3b service nova] Releasing lock "refresh_cache-eeb51744-e397-40f8-8e1a-813fa23ebe95" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1048.826049] env[61923]: DEBUG oslo_vmware.api [None req-b47c5602-b8b6-48e8-b951-0e67592f330e tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52e3ec56-82d8-756c-9c79-c57eae38a621, 'name': SearchDatastore_Task, 'duration_secs': 0.007765} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1048.826782] env[61923]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-499eb6da-7997-458d-a233-243f45670663 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.832047] env[61923]: DEBUG oslo_vmware.api [None req-b47c5602-b8b6-48e8-b951-0e67592f330e tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Waiting for the task: (returnval){ [ 1048.832047] env[61923]: value = "session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]527ee499-4e68-33be-be17-06b9f77b2956" [ 1048.832047] env[61923]: _type = "Task" [ 1048.832047] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1048.839459] env[61923]: DEBUG oslo_vmware.api [None req-b47c5602-b8b6-48e8-b951-0e67592f330e tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]527ee499-4e68-33be-be17-06b9f77b2956, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1049.243113] env[61923]: INFO nova.compute.manager [None req-0ab6e8de-2300-4b6f-8852-0f5735e09090 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: 9aeac922-aac2-47fe-8c11-835bef75e55c] Rebuilding instance [ 1049.280580] env[61923]: DEBUG nova.compute.manager [None req-0ab6e8de-2300-4b6f-8852-0f5735e09090 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: 9aeac922-aac2-47fe-8c11-835bef75e55c] Checking state {{(pid=61923) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1049.281450] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d2075d3-f460-4497-8917-130443755178 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.341818] env[61923]: DEBUG oslo_vmware.api [None req-b47c5602-b8b6-48e8-b951-0e67592f330e tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]527ee499-4e68-33be-be17-06b9f77b2956, 'name': SearchDatastore_Task, 'duration_secs': 0.008953} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1049.342089] env[61923]: DEBUG oslo_concurrency.lockutils [None req-b47c5602-b8b6-48e8-b951-0e67592f330e tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1049.342347] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-b47c5602-b8b6-48e8-b951-0e67592f330e tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk to [datastore1] eeb51744-e397-40f8-8e1a-813fa23ebe95/eeb51744-e397-40f8-8e1a-813fa23ebe95.vmdk {{(pid=61923) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1049.342590] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-836ccd3d-8349-43df-9ab4-ec510bfe5834 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.349189] env[61923]: DEBUG oslo_vmware.api [None req-b47c5602-b8b6-48e8-b951-0e67592f330e tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Waiting for the task: (returnval){ [ 1049.349189] env[61923]: value = "task-1377975" [ 1049.349189] env[61923]: _type = "Task" [ 1049.349189] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1049.356023] env[61923]: DEBUG oslo_vmware.api [None req-b47c5602-b8b6-48e8-b951-0e67592f330e tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Task: {'id': task-1377975, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1049.791951] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-0ab6e8de-2300-4b6f-8852-0f5735e09090 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: 9aeac922-aac2-47fe-8c11-835bef75e55c] Powering off the VM {{(pid=61923) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1049.792277] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-674dc4c3-892e-4c19-ae4f-0707ce340ec4 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.798473] env[61923]: DEBUG oslo_vmware.api [None req-0ab6e8de-2300-4b6f-8852-0f5735e09090 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Waiting for the task: (returnval){ [ 1049.798473] env[61923]: value = "task-1377976" [ 1049.798473] env[61923]: _type = "Task" [ 1049.798473] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1049.806307] env[61923]: DEBUG oslo_vmware.api [None req-0ab6e8de-2300-4b6f-8852-0f5735e09090 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Task: {'id': task-1377976, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1049.858861] env[61923]: DEBUG oslo_vmware.api [None req-b47c5602-b8b6-48e8-b951-0e67592f330e tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Task: {'id': task-1377975, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.417374} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1049.859186] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-b47c5602-b8b6-48e8-b951-0e67592f330e tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk to [datastore1] eeb51744-e397-40f8-8e1a-813fa23ebe95/eeb51744-e397-40f8-8e1a-813fa23ebe95.vmdk {{(pid=61923) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1049.859403] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-b47c5602-b8b6-48e8-b951-0e67592f330e tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: eeb51744-e397-40f8-8e1a-813fa23ebe95] Extending root virtual disk to 1048576 {{(pid=61923) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1049.859657] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ee1ffe4d-bc60-439d-aed2-09b31b1c0726 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.866745] env[61923]: DEBUG oslo_vmware.api [None req-b47c5602-b8b6-48e8-b951-0e67592f330e tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Waiting for the task: (returnval){ [ 1049.866745] env[61923]: value = "task-1377977" [ 1049.866745] env[61923]: _type = "Task" [ 1049.866745] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1049.874530] env[61923]: DEBUG oslo_vmware.api [None req-b47c5602-b8b6-48e8-b951-0e67592f330e tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Task: {'id': task-1377977, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1050.307876] env[61923]: DEBUG oslo_vmware.api [None req-0ab6e8de-2300-4b6f-8852-0f5735e09090 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Task: {'id': task-1377976, 'name': PowerOffVM_Task, 'duration_secs': 0.192413} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1050.308237] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-0ab6e8de-2300-4b6f-8852-0f5735e09090 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: 9aeac922-aac2-47fe-8c11-835bef75e55c] Powered off the VM {{(pid=61923) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1050.308385] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-0ab6e8de-2300-4b6f-8852-0f5735e09090 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: 9aeac922-aac2-47fe-8c11-835bef75e55c] Destroying instance {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1050.309101] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48791ba8-8940-4706-bf55-baa568d6d66d {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.315185] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-0ab6e8de-2300-4b6f-8852-0f5735e09090 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: 9aeac922-aac2-47fe-8c11-835bef75e55c] Unregistering the VM {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1050.315395] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-0cfd4815-289a-4bc9-8755-38532db7fb2c {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.372863] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-0ab6e8de-2300-4b6f-8852-0f5735e09090 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: 9aeac922-aac2-47fe-8c11-835bef75e55c] Unregistered the VM {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1050.373083] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-0ab6e8de-2300-4b6f-8852-0f5735e09090 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: 9aeac922-aac2-47fe-8c11-835bef75e55c] Deleting contents of the VM from datastore datastore1 {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1050.373303] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-0ab6e8de-2300-4b6f-8852-0f5735e09090 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Deleting the datastore file [datastore1] 9aeac922-aac2-47fe-8c11-835bef75e55c {{(pid=61923) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1050.373949] env[61923]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-80db41fd-a583-4978-a49d-62da83d6f0f8 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.377780] env[61923]: DEBUG oslo_vmware.api [None req-b47c5602-b8b6-48e8-b951-0e67592f330e tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Task: {'id': task-1377977, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.057258} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1050.378387] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-b47c5602-b8b6-48e8-b951-0e67592f330e tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: eeb51744-e397-40f8-8e1a-813fa23ebe95] Extended root virtual disk {{(pid=61923) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1050.379110] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d35da7df-37d4-472a-b26f-f2c2fa6a6042 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.382417] env[61923]: DEBUG oslo_vmware.api [None req-0ab6e8de-2300-4b6f-8852-0f5735e09090 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Waiting for the task: (returnval){ [ 1050.382417] env[61923]: value = "task-1377979" [ 1050.382417] env[61923]: _type = "Task" [ 1050.382417] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1050.403471] env[61923]: DEBUG nova.virt.vmwareapi.volumeops [None req-b47c5602-b8b6-48e8-b951-0e67592f330e tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: eeb51744-e397-40f8-8e1a-813fa23ebe95] Reconfiguring VM instance instance-00000062 to attach disk [datastore1] eeb51744-e397-40f8-8e1a-813fa23ebe95/eeb51744-e397-40f8-8e1a-813fa23ebe95.vmdk or device None with type sparse {{(pid=61923) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1050.404058] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2c282b18-634c-47d7-a2b9-3e5dad2fe299 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.421736] env[61923]: DEBUG oslo_vmware.api [None req-0ab6e8de-2300-4b6f-8852-0f5735e09090 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Task: {'id': task-1377979, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1050.426107] env[61923]: DEBUG oslo_vmware.api [None req-b47c5602-b8b6-48e8-b951-0e67592f330e tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Waiting for the task: (returnval){ [ 1050.426107] env[61923]: value = "task-1377980" [ 1050.426107] env[61923]: _type = "Task" [ 1050.426107] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1050.433014] env[61923]: DEBUG oslo_vmware.api [None req-b47c5602-b8b6-48e8-b951-0e67592f330e tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Task: {'id': task-1377980, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1050.892358] env[61923]: DEBUG oslo_vmware.api [None req-0ab6e8de-2300-4b6f-8852-0f5735e09090 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Task: {'id': task-1377979, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.3425} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1050.892609] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-0ab6e8de-2300-4b6f-8852-0f5735e09090 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Deleted the datastore file {{(pid=61923) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1050.892806] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-0ab6e8de-2300-4b6f-8852-0f5735e09090 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: 9aeac922-aac2-47fe-8c11-835bef75e55c] Deleted contents of the VM from datastore datastore1 {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1050.892981] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-0ab6e8de-2300-4b6f-8852-0f5735e09090 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: 9aeac922-aac2-47fe-8c11-835bef75e55c] Instance destroyed {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1050.936374] env[61923]: DEBUG oslo_vmware.api [None req-b47c5602-b8b6-48e8-b951-0e67592f330e tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Task: {'id': task-1377980, 'name': ReconfigVM_Task} progress is 14%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1051.437630] env[61923]: DEBUG oslo_vmware.api [None req-b47c5602-b8b6-48e8-b951-0e67592f330e tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Task: {'id': task-1377980, 'name': ReconfigVM_Task, 'duration_secs': 0.819656} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1051.438061] env[61923]: DEBUG nova.virt.vmwareapi.volumeops [None req-b47c5602-b8b6-48e8-b951-0e67592f330e tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: eeb51744-e397-40f8-8e1a-813fa23ebe95] Reconfigured VM instance instance-00000062 to attach disk [datastore1] eeb51744-e397-40f8-8e1a-813fa23ebe95/eeb51744-e397-40f8-8e1a-813fa23ebe95.vmdk or device None with type sparse {{(pid=61923) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1051.438609] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-9d55d1ad-a5d0-4655-9808-33cdc5bdd9fe {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.444642] env[61923]: DEBUG oslo_vmware.api [None req-b47c5602-b8b6-48e8-b951-0e67592f330e tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Waiting for the task: (returnval){ [ 1051.444642] env[61923]: value = "task-1377981" [ 1051.444642] env[61923]: _type = "Task" [ 1051.444642] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1051.452944] env[61923]: DEBUG oslo_vmware.api [None req-b47c5602-b8b6-48e8-b951-0e67592f330e tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Task: {'id': task-1377981, 'name': Rename_Task} progress is 5%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1051.923368] env[61923]: DEBUG nova.virt.hardware [None req-0ab6e8de-2300-4b6f-8852-0f5735e09090 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-29T20:07:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-29T20:07:35Z,direct_url=,disk_format='vmdk',id=0f153f63-ae0a-45b1-b7f5-7f9b673c947f,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='4e7b5e3b3c3443c8bd5c70f8a15739f5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-29T20:07:36Z,virtual_size=,visibility=), allow threads: False {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1051.923620] env[61923]: DEBUG nova.virt.hardware [None req-0ab6e8de-2300-4b6f-8852-0f5735e09090 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Flavor limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1051.923785] env[61923]: DEBUG nova.virt.hardware [None req-0ab6e8de-2300-4b6f-8852-0f5735e09090 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Image limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1051.923968] env[61923]: DEBUG nova.virt.hardware [None req-0ab6e8de-2300-4b6f-8852-0f5735e09090 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Flavor pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1051.924131] env[61923]: DEBUG nova.virt.hardware [None req-0ab6e8de-2300-4b6f-8852-0f5735e09090 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Image pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1051.924284] env[61923]: DEBUG nova.virt.hardware [None req-0ab6e8de-2300-4b6f-8852-0f5735e09090 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1051.924489] env[61923]: DEBUG nova.virt.hardware [None req-0ab6e8de-2300-4b6f-8852-0f5735e09090 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1051.924683] env[61923]: DEBUG nova.virt.hardware [None req-0ab6e8de-2300-4b6f-8852-0f5735e09090 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1051.924859] env[61923]: DEBUG nova.virt.hardware [None req-0ab6e8de-2300-4b6f-8852-0f5735e09090 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Got 1 possible topologies {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1051.925066] env[61923]: DEBUG nova.virt.hardware [None req-0ab6e8de-2300-4b6f-8852-0f5735e09090 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1051.925209] env[61923]: DEBUG nova.virt.hardware [None req-0ab6e8de-2300-4b6f-8852-0f5735e09090 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1051.926059] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3dcfacb-06cd-42be-af4c-2b13b9d8b569 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.933787] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9785808f-5145-4cef-b8b8-fd7b1ca49880 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.947173] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-0ab6e8de-2300-4b6f-8852-0f5735e09090 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: 9aeac922-aac2-47fe-8c11-835bef75e55c] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:00:7d:78', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '56b944d8-803d-43f2-945d-0f334ee4ea1c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '7f812cb5-e162-4dc7-9683-39a4c4048623', 'vif_model': 'vmxnet3'}] {{(pid=61923) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1051.954612] env[61923]: DEBUG oslo.service.loopingcall [None req-0ab6e8de-2300-4b6f-8852-0f5735e09090 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61923) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1051.957251] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9aeac922-aac2-47fe-8c11-835bef75e55c] Creating VM on the ESX host {{(pid=61923) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1051.957467] env[61923]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-8483b883-3f0d-4dce-8483-35decc6f71df {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.975342] env[61923]: DEBUG oslo_vmware.api [None req-b47c5602-b8b6-48e8-b951-0e67592f330e tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Task: {'id': task-1377981, 'name': Rename_Task, 'duration_secs': 0.130787} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1051.976388] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-b47c5602-b8b6-48e8-b951-0e67592f330e tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: eeb51744-e397-40f8-8e1a-813fa23ebe95] Powering on the VM {{(pid=61923) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1051.976594] env[61923]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1051.976594] env[61923]: value = "task-1377982" [ 1051.976594] env[61923]: _type = "Task" [ 1051.976594] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1051.976775] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d83a8bf6-b766-4722-bff6-a67fd16f963e {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.984813] env[61923]: DEBUG oslo_vmware.api [-] Task: {'id': task-1377982, 'name': CreateVM_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1051.985881] env[61923]: DEBUG oslo_vmware.api [None req-b47c5602-b8b6-48e8-b951-0e67592f330e tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Waiting for the task: (returnval){ [ 1051.985881] env[61923]: value = "task-1377983" [ 1051.985881] env[61923]: _type = "Task" [ 1051.985881] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1051.992461] env[61923]: DEBUG oslo_vmware.api [None req-b47c5602-b8b6-48e8-b951-0e67592f330e tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Task: {'id': task-1377983, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1052.487333] env[61923]: DEBUG oslo_vmware.api [-] Task: {'id': task-1377982, 'name': CreateVM_Task, 'duration_secs': 0.280193} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1052.489986] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9aeac922-aac2-47fe-8c11-835bef75e55c] Created VM on the ESX host {{(pid=61923) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1052.490695] env[61923]: DEBUG oslo_concurrency.lockutils [None req-0ab6e8de-2300-4b6f-8852-0f5735e09090 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1052.490860] env[61923]: DEBUG oslo_concurrency.lockutils [None req-0ab6e8de-2300-4b6f-8852-0f5735e09090 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1052.491192] env[61923]: DEBUG oslo_concurrency.lockutils [None req-0ab6e8de-2300-4b6f-8852-0f5735e09090 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1052.491810] env[61923]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-dbfee54d-b970-45f8-8880-b58fb4dae420 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.496219] env[61923]: DEBUG oslo_vmware.api [None req-b47c5602-b8b6-48e8-b951-0e67592f330e tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Task: {'id': task-1377983, 'name': PowerOnVM_Task} progress is 66%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1052.498853] env[61923]: DEBUG oslo_vmware.api [None req-0ab6e8de-2300-4b6f-8852-0f5735e09090 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Waiting for the task: (returnval){ [ 1052.498853] env[61923]: value = "session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52ac659d-9b5f-af70-d516-dbcf13053582" [ 1052.498853] env[61923]: _type = "Task" [ 1052.498853] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1052.505316] env[61923]: DEBUG oslo_vmware.api [None req-0ab6e8de-2300-4b6f-8852-0f5735e09090 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52ac659d-9b5f-af70-d516-dbcf13053582, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1052.997958] env[61923]: DEBUG oslo_vmware.api [None req-b47c5602-b8b6-48e8-b951-0e67592f330e tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Task: {'id': task-1377983, 'name': PowerOnVM_Task, 'duration_secs': 0.63942} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1052.998247] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-b47c5602-b8b6-48e8-b951-0e67592f330e tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: eeb51744-e397-40f8-8e1a-813fa23ebe95] Powered on the VM {{(pid=61923) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1052.998447] env[61923]: INFO nova.compute.manager [None req-b47c5602-b8b6-48e8-b951-0e67592f330e tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: eeb51744-e397-40f8-8e1a-813fa23ebe95] Took 7.83 seconds to spawn the instance on the hypervisor. [ 1052.998630] env[61923]: DEBUG nova.compute.manager [None req-b47c5602-b8b6-48e8-b951-0e67592f330e tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: eeb51744-e397-40f8-8e1a-813fa23ebe95] Checking state {{(pid=61923) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1052.999414] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad040771-c2c7-43c1-821b-e4ec428c000f {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.013861] env[61923]: DEBUG oslo_vmware.api [None req-0ab6e8de-2300-4b6f-8852-0f5735e09090 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52ac659d-9b5f-af70-d516-dbcf13053582, 'name': SearchDatastore_Task, 'duration_secs': 0.010514} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1053.014145] env[61923]: DEBUG oslo_concurrency.lockutils [None req-0ab6e8de-2300-4b6f-8852-0f5735e09090 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1053.014371] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-0ab6e8de-2300-4b6f-8852-0f5735e09090 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: 9aeac922-aac2-47fe-8c11-835bef75e55c] Processing image 0f153f63-ae0a-45b1-b7f5-7f9b673c947f {{(pid=61923) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1053.014603] env[61923]: DEBUG oslo_concurrency.lockutils [None req-0ab6e8de-2300-4b6f-8852-0f5735e09090 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1053.014755] env[61923]: DEBUG oslo_concurrency.lockutils [None req-0ab6e8de-2300-4b6f-8852-0f5735e09090 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1053.014930] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-0ab6e8de-2300-4b6f-8852-0f5735e09090 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61923) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1053.015178] env[61923]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b59c4179-f6fd-4b3a-9928-c70493e992d8 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.022505] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-0ab6e8de-2300-4b6f-8852-0f5735e09090 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61923) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1053.022688] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-0ab6e8de-2300-4b6f-8852-0f5735e09090 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61923) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1053.023373] env[61923]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6b328a29-1785-49b6-a684-8890a3991a9d {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.027718] env[61923]: DEBUG oslo_vmware.api [None req-0ab6e8de-2300-4b6f-8852-0f5735e09090 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Waiting for the task: (returnval){ [ 1053.027718] env[61923]: value = "session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]522e0572-5480-1e7b-b104-2d5fa89e2f83" [ 1053.027718] env[61923]: _type = "Task" [ 1053.027718] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1053.034769] env[61923]: DEBUG oslo_vmware.api [None req-0ab6e8de-2300-4b6f-8852-0f5735e09090 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]522e0572-5480-1e7b-b104-2d5fa89e2f83, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1053.518600] env[61923]: INFO nova.compute.manager [None req-b47c5602-b8b6-48e8-b951-0e67592f330e tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: eeb51744-e397-40f8-8e1a-813fa23ebe95] Took 13.51 seconds to build instance. [ 1053.537463] env[61923]: DEBUG oslo_vmware.api [None req-0ab6e8de-2300-4b6f-8852-0f5735e09090 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]522e0572-5480-1e7b-b104-2d5fa89e2f83, 'name': SearchDatastore_Task, 'duration_secs': 0.009162} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1053.538219] env[61923]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-19ecbc16-14d8-441d-b28a-1f8a40d442d7 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.543149] env[61923]: DEBUG oslo_vmware.api [None req-0ab6e8de-2300-4b6f-8852-0f5735e09090 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Waiting for the task: (returnval){ [ 1053.543149] env[61923]: value = "session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]525da284-3b21-94db-9561-fb66c239f92d" [ 1053.543149] env[61923]: _type = "Task" [ 1053.543149] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1053.550244] env[61923]: DEBUG oslo_vmware.api [None req-0ab6e8de-2300-4b6f-8852-0f5735e09090 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]525da284-3b21-94db-9561-fb66c239f92d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1054.020721] env[61923]: DEBUG oslo_concurrency.lockutils [None req-b47c5602-b8b6-48e8-b951-0e67592f330e tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Lock "eeb51744-e397-40f8-8e1a-813fa23ebe95" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.020s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1054.053907] env[61923]: DEBUG oslo_vmware.api [None req-0ab6e8de-2300-4b6f-8852-0f5735e09090 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]525da284-3b21-94db-9561-fb66c239f92d, 'name': SearchDatastore_Task, 'duration_secs': 0.008499} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1054.054175] env[61923]: DEBUG oslo_concurrency.lockutils [None req-0ab6e8de-2300-4b6f-8852-0f5735e09090 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1054.054477] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-0ab6e8de-2300-4b6f-8852-0f5735e09090 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk to [datastore2] 9aeac922-aac2-47fe-8c11-835bef75e55c/9aeac922-aac2-47fe-8c11-835bef75e55c.vmdk {{(pid=61923) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1054.054772] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-49c88a26-7175-43e9-adcb-410185758249 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.061784] env[61923]: DEBUG oslo_vmware.api [None req-0ab6e8de-2300-4b6f-8852-0f5735e09090 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Waiting for the task: (returnval){ [ 1054.061784] env[61923]: value = "task-1377984" [ 1054.061784] env[61923]: _type = "Task" [ 1054.061784] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1054.069096] env[61923]: DEBUG oslo_vmware.api [None req-0ab6e8de-2300-4b6f-8852-0f5735e09090 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Task: {'id': task-1377984, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1054.339862] env[61923]: DEBUG oslo_concurrency.lockutils [None req-1998a0c6-f793-41cc-9eb2-a3641f26cd31 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Acquiring lock "eeb51744-e397-40f8-8e1a-813fa23ebe95" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1054.340164] env[61923]: DEBUG oslo_concurrency.lockutils [None req-1998a0c6-f793-41cc-9eb2-a3641f26cd31 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Lock "eeb51744-e397-40f8-8e1a-813fa23ebe95" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1054.340351] env[61923]: DEBUG nova.compute.manager [None req-1998a0c6-f793-41cc-9eb2-a3641f26cd31 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: eeb51744-e397-40f8-8e1a-813fa23ebe95] Checking state {{(pid=61923) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1054.341331] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab693d00-9f0b-4f1a-a64a-2149706121f2 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.351686] env[61923]: DEBUG nova.compute.manager [None req-1998a0c6-f793-41cc-9eb2-a3641f26cd31 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: eeb51744-e397-40f8-8e1a-813fa23ebe95] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=61923) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 1054.352362] env[61923]: DEBUG nova.objects.instance [None req-1998a0c6-f793-41cc-9eb2-a3641f26cd31 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Lazy-loading 'flavor' on Instance uuid eeb51744-e397-40f8-8e1a-813fa23ebe95 {{(pid=61923) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1054.571747] env[61923]: DEBUG oslo_vmware.api [None req-0ab6e8de-2300-4b6f-8852-0f5735e09090 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Task: {'id': task-1377984, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.435204} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1054.572199] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-0ab6e8de-2300-4b6f-8852-0f5735e09090 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk to [datastore2] 9aeac922-aac2-47fe-8c11-835bef75e55c/9aeac922-aac2-47fe-8c11-835bef75e55c.vmdk {{(pid=61923) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1054.572278] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-0ab6e8de-2300-4b6f-8852-0f5735e09090 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: 9aeac922-aac2-47fe-8c11-835bef75e55c] Extending root virtual disk to 1048576 {{(pid=61923) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1054.572477] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-696c9da8-a562-4060-9c87-29f3b5da1259 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.579644] env[61923]: DEBUG oslo_vmware.api [None req-0ab6e8de-2300-4b6f-8852-0f5735e09090 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Waiting for the task: (returnval){ [ 1054.579644] env[61923]: value = "task-1377985" [ 1054.579644] env[61923]: _type = "Task" [ 1054.579644] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1054.587079] env[61923]: DEBUG oslo_vmware.api [None req-0ab6e8de-2300-4b6f-8852-0f5735e09090 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Task: {'id': task-1377985, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1054.857800] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-1998a0c6-f793-41cc-9eb2-a3641f26cd31 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: eeb51744-e397-40f8-8e1a-813fa23ebe95] Powering off the VM {{(pid=61923) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1054.858052] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1e417de7-89ac-4159-860e-5a05d972ee86 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1054.865577] env[61923]: DEBUG oslo_vmware.api [None req-1998a0c6-f793-41cc-9eb2-a3641f26cd31 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Waiting for the task: (returnval){ [ 1054.865577] env[61923]: value = "task-1377986" [ 1054.865577] env[61923]: _type = "Task" [ 1054.865577] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1054.875901] env[61923]: DEBUG oslo_vmware.api [None req-1998a0c6-f793-41cc-9eb2-a3641f26cd31 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Task: {'id': task-1377986, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1055.089498] env[61923]: DEBUG oslo_vmware.api [None req-0ab6e8de-2300-4b6f-8852-0f5735e09090 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Task: {'id': task-1377985, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.062859} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1055.089765] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-0ab6e8de-2300-4b6f-8852-0f5735e09090 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: 9aeac922-aac2-47fe-8c11-835bef75e55c] Extended root virtual disk {{(pid=61923) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1055.090542] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf2131fe-335f-4c19-8f13-afa06da5d2ee {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.111895] env[61923]: DEBUG nova.virt.vmwareapi.volumeops [None req-0ab6e8de-2300-4b6f-8852-0f5735e09090 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: 9aeac922-aac2-47fe-8c11-835bef75e55c] Reconfiguring VM instance instance-00000061 to attach disk [datastore2] 9aeac922-aac2-47fe-8c11-835bef75e55c/9aeac922-aac2-47fe-8c11-835bef75e55c.vmdk or device None with type sparse {{(pid=61923) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1055.112135] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-070aa960-36d6-4e92-a029-ba836d01deea {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.129662] env[61923]: DEBUG oslo_vmware.api [None req-0ab6e8de-2300-4b6f-8852-0f5735e09090 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Waiting for the task: (returnval){ [ 1055.129662] env[61923]: value = "task-1377987" [ 1055.129662] env[61923]: _type = "Task" [ 1055.129662] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1055.136544] env[61923]: DEBUG oslo_vmware.api [None req-0ab6e8de-2300-4b6f-8852-0f5735e09090 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Task: {'id': task-1377987, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1055.376478] env[61923]: DEBUG oslo_vmware.api [None req-1998a0c6-f793-41cc-9eb2-a3641f26cd31 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Task: {'id': task-1377986, 'name': PowerOffVM_Task, 'duration_secs': 0.164161} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1055.376608] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-1998a0c6-f793-41cc-9eb2-a3641f26cd31 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: eeb51744-e397-40f8-8e1a-813fa23ebe95] Powered off the VM {{(pid=61923) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1055.376757] env[61923]: DEBUG nova.compute.manager [None req-1998a0c6-f793-41cc-9eb2-a3641f26cd31 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: eeb51744-e397-40f8-8e1a-813fa23ebe95] Checking state {{(pid=61923) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1055.377533] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2941fc62-a023-4ce8-80c3-38138f7cc069 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.639449] env[61923]: DEBUG oslo_vmware.api [None req-0ab6e8de-2300-4b6f-8852-0f5735e09090 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Task: {'id': task-1377987, 'name': ReconfigVM_Task, 'duration_secs': 0.259457} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1055.639835] env[61923]: DEBUG nova.virt.vmwareapi.volumeops [None req-0ab6e8de-2300-4b6f-8852-0f5735e09090 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: 9aeac922-aac2-47fe-8c11-835bef75e55c] Reconfigured VM instance instance-00000061 to attach disk [datastore2] 9aeac922-aac2-47fe-8c11-835bef75e55c/9aeac922-aac2-47fe-8c11-835bef75e55c.vmdk or device None with type sparse {{(pid=61923) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1055.640396] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-2704a91f-0fe6-4793-af43-8f91b7e467e2 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.647388] env[61923]: DEBUG oslo_vmware.api [None req-0ab6e8de-2300-4b6f-8852-0f5735e09090 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Waiting for the task: (returnval){ [ 1055.647388] env[61923]: value = "task-1377988" [ 1055.647388] env[61923]: _type = "Task" [ 1055.647388] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1055.654724] env[61923]: DEBUG oslo_vmware.api [None req-0ab6e8de-2300-4b6f-8852-0f5735e09090 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Task: {'id': task-1377988, 'name': Rename_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1055.891216] env[61923]: DEBUG oslo_concurrency.lockutils [None req-1998a0c6-f793-41cc-9eb2-a3641f26cd31 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Lock "eeb51744-e397-40f8-8e1a-813fa23ebe95" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.551s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1056.157186] env[61923]: DEBUG oslo_vmware.api [None req-0ab6e8de-2300-4b6f-8852-0f5735e09090 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Task: {'id': task-1377988, 'name': Rename_Task, 'duration_secs': 0.139122} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1056.157459] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-0ab6e8de-2300-4b6f-8852-0f5735e09090 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: 9aeac922-aac2-47fe-8c11-835bef75e55c] Powering on the VM {{(pid=61923) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1056.157702] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-60e6dfd5-c6b5-4eef-9b85-06b8eee41d00 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.163347] env[61923]: DEBUG oslo_vmware.api [None req-0ab6e8de-2300-4b6f-8852-0f5735e09090 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Waiting for the task: (returnval){ [ 1056.163347] env[61923]: value = "task-1377989" [ 1056.163347] env[61923]: _type = "Task" [ 1056.163347] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1056.170310] env[61923]: DEBUG oslo_vmware.api [None req-0ab6e8de-2300-4b6f-8852-0f5735e09090 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Task: {'id': task-1377989, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1056.673620] env[61923]: DEBUG oslo_vmware.api [None req-0ab6e8de-2300-4b6f-8852-0f5735e09090 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Task: {'id': task-1377989, 'name': PowerOnVM_Task, 'duration_secs': 0.393692} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1056.673997] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-0ab6e8de-2300-4b6f-8852-0f5735e09090 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: 9aeac922-aac2-47fe-8c11-835bef75e55c] Powered on the VM {{(pid=61923) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1056.674140] env[61923]: DEBUG nova.compute.manager [None req-0ab6e8de-2300-4b6f-8852-0f5735e09090 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: 9aeac922-aac2-47fe-8c11-835bef75e55c] Checking state {{(pid=61923) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1056.674931] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf80d91f-a09c-4b03-80f6-eace60dcc231 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.696043] env[61923]: DEBUG oslo_concurrency.lockutils [None req-dde308d1-eed0-4731-b261-71025e36a144 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Acquiring lock "eeb51744-e397-40f8-8e1a-813fa23ebe95" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1056.696192] env[61923]: DEBUG oslo_concurrency.lockutils [None req-dde308d1-eed0-4731-b261-71025e36a144 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Lock "eeb51744-e397-40f8-8e1a-813fa23ebe95" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1056.696343] env[61923]: DEBUG oslo_concurrency.lockutils [None req-dde308d1-eed0-4731-b261-71025e36a144 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Acquiring lock "eeb51744-e397-40f8-8e1a-813fa23ebe95-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1056.696527] env[61923]: DEBUG oslo_concurrency.lockutils [None req-dde308d1-eed0-4731-b261-71025e36a144 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Lock "eeb51744-e397-40f8-8e1a-813fa23ebe95-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1056.696701] env[61923]: DEBUG oslo_concurrency.lockutils [None req-dde308d1-eed0-4731-b261-71025e36a144 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Lock "eeb51744-e397-40f8-8e1a-813fa23ebe95-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1056.698973] env[61923]: INFO nova.compute.manager [None req-dde308d1-eed0-4731-b261-71025e36a144 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: eeb51744-e397-40f8-8e1a-813fa23ebe95] Terminating instance [ 1056.700874] env[61923]: DEBUG nova.compute.manager [None req-dde308d1-eed0-4731-b261-71025e36a144 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: eeb51744-e397-40f8-8e1a-813fa23ebe95] Start destroying the instance on the hypervisor. {{(pid=61923) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1056.701076] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-dde308d1-eed0-4731-b261-71025e36a144 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: eeb51744-e397-40f8-8e1a-813fa23ebe95] Destroying instance {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1056.701852] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92a8c45d-cc9f-477c-aceb-ae06e73499ce {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.708321] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-dde308d1-eed0-4731-b261-71025e36a144 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: eeb51744-e397-40f8-8e1a-813fa23ebe95] Unregistering the VM {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1056.708538] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c8186b1a-daaf-43e6-b815-6ffc808240ac {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.775647] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-dde308d1-eed0-4731-b261-71025e36a144 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: eeb51744-e397-40f8-8e1a-813fa23ebe95] Unregistered the VM {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1056.775870] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-dde308d1-eed0-4731-b261-71025e36a144 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: eeb51744-e397-40f8-8e1a-813fa23ebe95] Deleting contents of the VM from datastore datastore1 {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1056.776067] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-dde308d1-eed0-4731-b261-71025e36a144 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Deleting the datastore file [datastore1] eeb51744-e397-40f8-8e1a-813fa23ebe95 {{(pid=61923) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1056.776332] env[61923]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e87ff6fb-60f9-4312-8ebd-fa75ce004969 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.782683] env[61923]: DEBUG oslo_vmware.api [None req-dde308d1-eed0-4731-b261-71025e36a144 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Waiting for the task: (returnval){ [ 1056.782683] env[61923]: value = "task-1377991" [ 1056.782683] env[61923]: _type = "Task" [ 1056.782683] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1056.789878] env[61923]: DEBUG oslo_vmware.api [None req-dde308d1-eed0-4731-b261-71025e36a144 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Task: {'id': task-1377991, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1057.192282] env[61923]: DEBUG oslo_concurrency.lockutils [None req-0ab6e8de-2300-4b6f-8852-0f5735e09090 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1057.192282] env[61923]: DEBUG oslo_concurrency.lockutils [None req-0ab6e8de-2300-4b6f-8852-0f5735e09090 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1057.192282] env[61923]: DEBUG nova.objects.instance [None req-0ab6e8de-2300-4b6f-8852-0f5735e09090 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: 9aeac922-aac2-47fe-8c11-835bef75e55c] Trying to apply a migration context that does not seem to be set for this instance {{(pid=61923) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 1057.292520] env[61923]: DEBUG oslo_vmware.api [None req-dde308d1-eed0-4731-b261-71025e36a144 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Task: {'id': task-1377991, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.191028} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1057.292761] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-dde308d1-eed0-4731-b261-71025e36a144 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Deleted the datastore file {{(pid=61923) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1057.292946] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-dde308d1-eed0-4731-b261-71025e36a144 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: eeb51744-e397-40f8-8e1a-813fa23ebe95] Deleted contents of the VM from datastore datastore1 {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1057.293164] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-dde308d1-eed0-4731-b261-71025e36a144 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: eeb51744-e397-40f8-8e1a-813fa23ebe95] Instance destroyed {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1057.293351] env[61923]: INFO nova.compute.manager [None req-dde308d1-eed0-4731-b261-71025e36a144 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: eeb51744-e397-40f8-8e1a-813fa23ebe95] Took 0.59 seconds to destroy the instance on the hypervisor. [ 1057.293586] env[61923]: DEBUG oslo.service.loopingcall [None req-dde308d1-eed0-4731-b261-71025e36a144 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61923) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1057.293778] env[61923]: DEBUG nova.compute.manager [-] [instance: eeb51744-e397-40f8-8e1a-813fa23ebe95] Deallocating network for instance {{(pid=61923) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1057.293874] env[61923]: DEBUG nova.network.neutron [-] [instance: eeb51744-e397-40f8-8e1a-813fa23ebe95] deallocate_for_instance() {{(pid=61923) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1057.582579] env[61923]: DEBUG nova.compute.manager [req-da0ae1aa-4f3f-479a-8c38-278a5aff1670 req-baed0443-908d-4da3-82a3-acc2b2d33e4c service nova] [instance: eeb51744-e397-40f8-8e1a-813fa23ebe95] Received event network-vif-deleted-494148c6-b0ca-4198-bd04-465adbaffd47 {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1057.582891] env[61923]: INFO nova.compute.manager [req-da0ae1aa-4f3f-479a-8c38-278a5aff1670 req-baed0443-908d-4da3-82a3-acc2b2d33e4c service nova] [instance: eeb51744-e397-40f8-8e1a-813fa23ebe95] Neutron deleted interface 494148c6-b0ca-4198-bd04-465adbaffd47; detaching it from the instance and deleting it from the info cache [ 1057.582984] env[61923]: DEBUG nova.network.neutron [req-da0ae1aa-4f3f-479a-8c38-278a5aff1670 req-baed0443-908d-4da3-82a3-acc2b2d33e4c service nova] [instance: eeb51744-e397-40f8-8e1a-813fa23ebe95] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1058.058695] env[61923]: DEBUG nova.network.neutron [-] [instance: eeb51744-e397-40f8-8e1a-813fa23ebe95] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1058.085766] env[61923]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-9fd06348-60fa-4f63-a3aa-3bdb421764e5 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.095653] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a460e85e-f68b-4725-92d1-320fad1244de {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.121193] env[61923]: DEBUG nova.compute.manager [req-da0ae1aa-4f3f-479a-8c38-278a5aff1670 req-baed0443-908d-4da3-82a3-acc2b2d33e4c service nova] [instance: eeb51744-e397-40f8-8e1a-813fa23ebe95] Detach interface failed, port_id=494148c6-b0ca-4198-bd04-465adbaffd47, reason: Instance eeb51744-e397-40f8-8e1a-813fa23ebe95 could not be found. {{(pid=61923) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1058.199276] env[61923]: DEBUG oslo_concurrency.lockutils [None req-0ab6e8de-2300-4b6f-8852-0f5735e09090 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.008s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1058.301060] env[61923]: DEBUG oslo_concurrency.lockutils [None req-cad45a90-3f63-4a20-9553-2e39f212668e tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Acquiring lock "9aeac922-aac2-47fe-8c11-835bef75e55c" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1058.301060] env[61923]: DEBUG oslo_concurrency.lockutils [None req-cad45a90-3f63-4a20-9553-2e39f212668e tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Lock "9aeac922-aac2-47fe-8c11-835bef75e55c" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1058.301299] env[61923]: DEBUG oslo_concurrency.lockutils [None req-cad45a90-3f63-4a20-9553-2e39f212668e tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Acquiring lock "9aeac922-aac2-47fe-8c11-835bef75e55c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1058.301471] env[61923]: DEBUG oslo_concurrency.lockutils [None req-cad45a90-3f63-4a20-9553-2e39f212668e tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Lock "9aeac922-aac2-47fe-8c11-835bef75e55c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1058.301646] env[61923]: DEBUG oslo_concurrency.lockutils [None req-cad45a90-3f63-4a20-9553-2e39f212668e tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Lock "9aeac922-aac2-47fe-8c11-835bef75e55c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1058.303777] env[61923]: INFO nova.compute.manager [None req-cad45a90-3f63-4a20-9553-2e39f212668e tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: 9aeac922-aac2-47fe-8c11-835bef75e55c] Terminating instance [ 1058.305501] env[61923]: DEBUG nova.compute.manager [None req-cad45a90-3f63-4a20-9553-2e39f212668e tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: 9aeac922-aac2-47fe-8c11-835bef75e55c] Start destroying the instance on the hypervisor. {{(pid=61923) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1058.305697] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-cad45a90-3f63-4a20-9553-2e39f212668e tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: 9aeac922-aac2-47fe-8c11-835bef75e55c] Destroying instance {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1058.306535] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e44a2e8-2b9b-4117-923e-28537d77d6f0 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.314090] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-cad45a90-3f63-4a20-9553-2e39f212668e tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: 9aeac922-aac2-47fe-8c11-835bef75e55c] Powering off the VM {{(pid=61923) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1058.314291] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f6d6b14e-9bf9-4f94-a3d4-1ab057dbbea2 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.319666] env[61923]: DEBUG oslo_vmware.api [None req-cad45a90-3f63-4a20-9553-2e39f212668e tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Waiting for the task: (returnval){ [ 1058.319666] env[61923]: value = "task-1377992" [ 1058.319666] env[61923]: _type = "Task" [ 1058.319666] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1058.326973] env[61923]: DEBUG oslo_vmware.api [None req-cad45a90-3f63-4a20-9553-2e39f212668e tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Task: {'id': task-1377992, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1058.565074] env[61923]: INFO nova.compute.manager [-] [instance: eeb51744-e397-40f8-8e1a-813fa23ebe95] Took 1.27 seconds to deallocate network for instance. [ 1058.681889] env[61923]: DEBUG oslo_concurrency.lockutils [None req-2e0136c0-8e92-4295-b28a-93095da1d6f6 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Acquiring lock "ab3edc53-37a5-4119-84c5-fd408ed66b69" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1058.682196] env[61923]: DEBUG oslo_concurrency.lockutils [None req-2e0136c0-8e92-4295-b28a-93095da1d6f6 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Lock "ab3edc53-37a5-4119-84c5-fd408ed66b69" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1058.830459] env[61923]: DEBUG oslo_vmware.api [None req-cad45a90-3f63-4a20-9553-2e39f212668e tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Task: {'id': task-1377992, 'name': PowerOffVM_Task, 'duration_secs': 0.195266} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1058.830722] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-cad45a90-3f63-4a20-9553-2e39f212668e tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: 9aeac922-aac2-47fe-8c11-835bef75e55c] Powered off the VM {{(pid=61923) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1058.830881] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-cad45a90-3f63-4a20-9553-2e39f212668e tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: 9aeac922-aac2-47fe-8c11-835bef75e55c] Unregistering the VM {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1058.831152] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-8533855d-7752-4cbd-be3a-6a8d2245e30e {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.889941] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-cad45a90-3f63-4a20-9553-2e39f212668e tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: 9aeac922-aac2-47fe-8c11-835bef75e55c] Unregistered the VM {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1058.890181] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-cad45a90-3f63-4a20-9553-2e39f212668e tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: 9aeac922-aac2-47fe-8c11-835bef75e55c] Deleting contents of the VM from datastore datastore2 {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1058.890366] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-cad45a90-3f63-4a20-9553-2e39f212668e tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Deleting the datastore file [datastore2] 9aeac922-aac2-47fe-8c11-835bef75e55c {{(pid=61923) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1058.890678] env[61923]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b2f4d68a-3ac0-4899-8c46-0ae4ba6df04d {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.896362] env[61923]: DEBUG oslo_vmware.api [None req-cad45a90-3f63-4a20-9553-2e39f212668e tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Waiting for the task: (returnval){ [ 1058.896362] env[61923]: value = "task-1377994" [ 1058.896362] env[61923]: _type = "Task" [ 1058.896362] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1058.903496] env[61923]: DEBUG oslo_vmware.api [None req-cad45a90-3f63-4a20-9553-2e39f212668e tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Task: {'id': task-1377994, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1059.071866] env[61923]: DEBUG oslo_concurrency.lockutils [None req-dde308d1-eed0-4731-b261-71025e36a144 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1059.072266] env[61923]: DEBUG oslo_concurrency.lockutils [None req-dde308d1-eed0-4731-b261-71025e36a144 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1059.072397] env[61923]: DEBUG nova.objects.instance [None req-dde308d1-eed0-4731-b261-71025e36a144 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Lazy-loading 'resources' on Instance uuid eeb51744-e397-40f8-8e1a-813fa23ebe95 {{(pid=61923) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1059.184729] env[61923]: DEBUG nova.compute.manager [None req-2e0136c0-8e92-4295-b28a-93095da1d6f6 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: ab3edc53-37a5-4119-84c5-fd408ed66b69] Starting instance... {{(pid=61923) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1059.406148] env[61923]: DEBUG oslo_vmware.api [None req-cad45a90-3f63-4a20-9553-2e39f212668e tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Task: {'id': task-1377994, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.133696} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1059.406367] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-cad45a90-3f63-4a20-9553-2e39f212668e tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Deleted the datastore file {{(pid=61923) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1059.406552] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-cad45a90-3f63-4a20-9553-2e39f212668e tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: 9aeac922-aac2-47fe-8c11-835bef75e55c] Deleted contents of the VM from datastore datastore2 {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1059.406733] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-cad45a90-3f63-4a20-9553-2e39f212668e tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: 9aeac922-aac2-47fe-8c11-835bef75e55c] Instance destroyed {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1059.406904] env[61923]: INFO nova.compute.manager [None req-cad45a90-3f63-4a20-9553-2e39f212668e tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: 9aeac922-aac2-47fe-8c11-835bef75e55c] Took 1.10 seconds to destroy the instance on the hypervisor. [ 1059.407168] env[61923]: DEBUG oslo.service.loopingcall [None req-cad45a90-3f63-4a20-9553-2e39f212668e tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61923) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1059.407366] env[61923]: DEBUG nova.compute.manager [-] [instance: 9aeac922-aac2-47fe-8c11-835bef75e55c] Deallocating network for instance {{(pid=61923) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1059.407460] env[61923]: DEBUG nova.network.neutron [-] [instance: 9aeac922-aac2-47fe-8c11-835bef75e55c] deallocate_for_instance() {{(pid=61923) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1059.637385] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3acf37e6-44e6-4978-abe4-c0f900f89388 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.645682] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4f9cc74-0b44-4183-99d1-adc75db14fd3 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.682743] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-375ad825-4994-40f8-950f-732bc04153cb {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.686448] env[61923]: DEBUG nova.compute.manager [req-fd9879c7-c548-4fba-a727-a014c1fc2bdd req-8b30a4bf-cd16-41fb-942e-7a1315c127b2 service nova] [instance: 9aeac922-aac2-47fe-8c11-835bef75e55c] Received event network-vif-deleted-7f812cb5-e162-4dc7-9683-39a4c4048623 {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1059.686643] env[61923]: INFO nova.compute.manager [req-fd9879c7-c548-4fba-a727-a014c1fc2bdd req-8b30a4bf-cd16-41fb-942e-7a1315c127b2 service nova] [instance: 9aeac922-aac2-47fe-8c11-835bef75e55c] Neutron deleted interface 7f812cb5-e162-4dc7-9683-39a4c4048623; detaching it from the instance and deleting it from the info cache [ 1059.686814] env[61923]: DEBUG nova.network.neutron [req-fd9879c7-c548-4fba-a727-a014c1fc2bdd req-8b30a4bf-cd16-41fb-942e-7a1315c127b2 service nova] [instance: 9aeac922-aac2-47fe-8c11-835bef75e55c] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1059.695940] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf01b1dc-1a53-4420-bb42-1d3eb6fc7b6f {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.702624] env[61923]: DEBUG oslo_concurrency.lockutils [None req-2e0136c0-8e92-4295-b28a-93095da1d6f6 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1059.710539] env[61923]: DEBUG nova.compute.provider_tree [None req-dde308d1-eed0-4731-b261-71025e36a144 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1060.151031] env[61923]: DEBUG nova.network.neutron [-] [instance: 9aeac922-aac2-47fe-8c11-835bef75e55c] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1060.189835] env[61923]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-fd8e79a7-64d2-4e87-b820-374fc06db272 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.199617] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-deebced8-9ddf-4661-88a9-efc74f321220 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.213557] env[61923]: DEBUG nova.scheduler.client.report [None req-dde308d1-eed0-4731-b261-71025e36a144 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1060.223096] env[61923]: DEBUG nova.compute.manager [req-fd9879c7-c548-4fba-a727-a014c1fc2bdd req-8b30a4bf-cd16-41fb-942e-7a1315c127b2 service nova] [instance: 9aeac922-aac2-47fe-8c11-835bef75e55c] Detach interface failed, port_id=7f812cb5-e162-4dc7-9683-39a4c4048623, reason: Instance 9aeac922-aac2-47fe-8c11-835bef75e55c could not be found. {{(pid=61923) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1060.654165] env[61923]: INFO nova.compute.manager [-] [instance: 9aeac922-aac2-47fe-8c11-835bef75e55c] Took 1.25 seconds to deallocate network for instance. [ 1060.717861] env[61923]: DEBUG oslo_concurrency.lockutils [None req-dde308d1-eed0-4731-b261-71025e36a144 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.646s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1060.720066] env[61923]: DEBUG oslo_concurrency.lockutils [None req-2e0136c0-8e92-4295-b28a-93095da1d6f6 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.017s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1060.721556] env[61923]: INFO nova.compute.claims [None req-2e0136c0-8e92-4295-b28a-93095da1d6f6 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: ab3edc53-37a5-4119-84c5-fd408ed66b69] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1060.735617] env[61923]: INFO nova.scheduler.client.report [None req-dde308d1-eed0-4731-b261-71025e36a144 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Deleted allocations for instance eeb51744-e397-40f8-8e1a-813fa23ebe95 [ 1061.160714] env[61923]: DEBUG oslo_concurrency.lockutils [None req-cad45a90-3f63-4a20-9553-2e39f212668e tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1061.241619] env[61923]: DEBUG oslo_concurrency.lockutils [None req-dde308d1-eed0-4731-b261-71025e36a144 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Lock "eeb51744-e397-40f8-8e1a-813fa23ebe95" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 4.545s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1061.776257] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76da8055-f31e-4066-91c0-5c0769bc70f3 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.784231] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-232059f0-6324-4d2e-b893-707e912fc7f9 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.812672] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa0d5626-a749-4ccb-bfbd-fc9e81ea8173 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.819470] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4455859-2112-4b52-99ea-46d920b5f070 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.832150] env[61923]: DEBUG nova.compute.provider_tree [None req-2e0136c0-8e92-4295-b28a-93095da1d6f6 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1062.335562] env[61923]: DEBUG nova.scheduler.client.report [None req-2e0136c0-8e92-4295-b28a-93095da1d6f6 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1062.508146] env[61923]: DEBUG oslo_concurrency.lockutils [None req-26bb2be5-f543-490f-89c5-3e7df42ca732 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Acquiring lock "38b08866-9b44-4fc7-957b-d329715af18e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1062.508390] env[61923]: DEBUG oslo_concurrency.lockutils [None req-26bb2be5-f543-490f-89c5-3e7df42ca732 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Lock "38b08866-9b44-4fc7-957b-d329715af18e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1062.840500] env[61923]: DEBUG oslo_concurrency.lockutils [None req-2e0136c0-8e92-4295-b28a-93095da1d6f6 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.120s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1062.841079] env[61923]: DEBUG nova.compute.manager [None req-2e0136c0-8e92-4295-b28a-93095da1d6f6 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: ab3edc53-37a5-4119-84c5-fd408ed66b69] Start building networks asynchronously for instance. {{(pid=61923) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1062.843666] env[61923]: DEBUG oslo_concurrency.lockutils [None req-cad45a90-3f63-4a20-9553-2e39f212668e tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.683s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1062.843900] env[61923]: DEBUG nova.objects.instance [None req-cad45a90-3f63-4a20-9553-2e39f212668e tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Lazy-loading 'resources' on Instance uuid 9aeac922-aac2-47fe-8c11-835bef75e55c {{(pid=61923) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1063.010793] env[61923]: DEBUG nova.compute.manager [None req-26bb2be5-f543-490f-89c5-3e7df42ca732 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: 38b08866-9b44-4fc7-957b-d329715af18e] Starting instance... {{(pid=61923) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1063.347108] env[61923]: DEBUG nova.compute.utils [None req-2e0136c0-8e92-4295-b28a-93095da1d6f6 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Using /dev/sd instead of None {{(pid=61923) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1063.351538] env[61923]: DEBUG nova.compute.manager [None req-2e0136c0-8e92-4295-b28a-93095da1d6f6 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: ab3edc53-37a5-4119-84c5-fd408ed66b69] Allocating IP information in the background. {{(pid=61923) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1063.351716] env[61923]: DEBUG nova.network.neutron [None req-2e0136c0-8e92-4295-b28a-93095da1d6f6 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: ab3edc53-37a5-4119-84c5-fd408ed66b69] allocate_for_instance() {{(pid=61923) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1063.400370] env[61923]: DEBUG nova.policy [None req-2e0136c0-8e92-4295-b28a-93095da1d6f6 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6087a385c42e4383aa26a7f586cca87f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5a4db4f8aa194a219e80c538cc715d43', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61923) authorize /opt/stack/nova/nova/policy.py:201}} [ 1063.408373] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d590504-bffe-405e-9a09-6b0a430026c5 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.416391] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe7733b5-0ed2-49b1-a7d4-daa99f351ff9 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.446030] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3f93fa2-d591-4f8e-886d-4d2a962026ef {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.452947] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-564afe3b-c389-4b3a-87c8-7652f734cf57 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.466699] env[61923]: DEBUG nova.compute.provider_tree [None req-cad45a90-3f63-4a20-9553-2e39f212668e tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1063.528042] env[61923]: DEBUG oslo_concurrency.lockutils [None req-26bb2be5-f543-490f-89c5-3e7df42ca732 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1063.636618] env[61923]: DEBUG nova.network.neutron [None req-2e0136c0-8e92-4295-b28a-93095da1d6f6 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: ab3edc53-37a5-4119-84c5-fd408ed66b69] Successfully created port: f2c777e2-f8f4-437b-b990-20e42dd2491b {{(pid=61923) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1063.852540] env[61923]: DEBUG nova.compute.manager [None req-2e0136c0-8e92-4295-b28a-93095da1d6f6 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: ab3edc53-37a5-4119-84c5-fd408ed66b69] Start building block device mappings for instance. {{(pid=61923) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1063.969858] env[61923]: DEBUG nova.scheduler.client.report [None req-cad45a90-3f63-4a20-9553-2e39f212668e tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1064.474357] env[61923]: DEBUG oslo_concurrency.lockutils [None req-cad45a90-3f63-4a20-9553-2e39f212668e tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.631s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1064.476667] env[61923]: DEBUG oslo_concurrency.lockutils [None req-26bb2be5-f543-490f-89c5-3e7df42ca732 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.949s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1064.478099] env[61923]: INFO nova.compute.claims [None req-26bb2be5-f543-490f-89c5-3e7df42ca732 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: 38b08866-9b44-4fc7-957b-d329715af18e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1064.495832] env[61923]: INFO nova.scheduler.client.report [None req-cad45a90-3f63-4a20-9553-2e39f212668e tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Deleted allocations for instance 9aeac922-aac2-47fe-8c11-835bef75e55c [ 1064.862653] env[61923]: DEBUG nova.compute.manager [None req-2e0136c0-8e92-4295-b28a-93095da1d6f6 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: ab3edc53-37a5-4119-84c5-fd408ed66b69] Start spawning the instance on the hypervisor. {{(pid=61923) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1064.885489] env[61923]: DEBUG nova.virt.hardware [None req-2e0136c0-8e92-4295-b28a-93095da1d6f6 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-29T20:07:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-29T20:07:35Z,direct_url=,disk_format='vmdk',id=0f153f63-ae0a-45b1-b7f5-7f9b673c947f,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='4e7b5e3b3c3443c8bd5c70f8a15739f5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-29T20:07:36Z,virtual_size=,visibility=), allow threads: False {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1064.885727] env[61923]: DEBUG nova.virt.hardware [None req-2e0136c0-8e92-4295-b28a-93095da1d6f6 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Flavor limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1064.885883] env[61923]: DEBUG nova.virt.hardware [None req-2e0136c0-8e92-4295-b28a-93095da1d6f6 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Image limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1064.886076] env[61923]: DEBUG nova.virt.hardware [None req-2e0136c0-8e92-4295-b28a-93095da1d6f6 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Flavor pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1064.886228] env[61923]: DEBUG nova.virt.hardware [None req-2e0136c0-8e92-4295-b28a-93095da1d6f6 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Image pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1064.886374] env[61923]: DEBUG nova.virt.hardware [None req-2e0136c0-8e92-4295-b28a-93095da1d6f6 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1064.886580] env[61923]: DEBUG nova.virt.hardware [None req-2e0136c0-8e92-4295-b28a-93095da1d6f6 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1064.886740] env[61923]: DEBUG nova.virt.hardware [None req-2e0136c0-8e92-4295-b28a-93095da1d6f6 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1064.886909] env[61923]: DEBUG nova.virt.hardware [None req-2e0136c0-8e92-4295-b28a-93095da1d6f6 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Got 1 possible topologies {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1064.887158] env[61923]: DEBUG nova.virt.hardware [None req-2e0136c0-8e92-4295-b28a-93095da1d6f6 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1064.887346] env[61923]: DEBUG nova.virt.hardware [None req-2e0136c0-8e92-4295-b28a-93095da1d6f6 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1064.888338] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-abd5b191-c136-4164-8646-1eecaac5e91b {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.895767] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b3fa6ad-f32b-4714-af1c-e738b06607ee {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.002063] env[61923]: DEBUG oslo_concurrency.lockutils [None req-cad45a90-3f63-4a20-9553-2e39f212668e tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Lock "9aeac922-aac2-47fe-8c11-835bef75e55c" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.701s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1065.044166] env[61923]: DEBUG nova.compute.manager [req-fdb2d9f8-2b7b-4bde-9f32-7e14437c84ba req-c5712089-c4db-4d56-9d63-0aedb3468ae2 service nova] [instance: ab3edc53-37a5-4119-84c5-fd408ed66b69] Received event network-vif-plugged-f2c777e2-f8f4-437b-b990-20e42dd2491b {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1065.044166] env[61923]: DEBUG oslo_concurrency.lockutils [req-fdb2d9f8-2b7b-4bde-9f32-7e14437c84ba req-c5712089-c4db-4d56-9d63-0aedb3468ae2 service nova] Acquiring lock "ab3edc53-37a5-4119-84c5-fd408ed66b69-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1065.044166] env[61923]: DEBUG oslo_concurrency.lockutils [req-fdb2d9f8-2b7b-4bde-9f32-7e14437c84ba req-c5712089-c4db-4d56-9d63-0aedb3468ae2 service nova] Lock "ab3edc53-37a5-4119-84c5-fd408ed66b69-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1065.044166] env[61923]: DEBUG oslo_concurrency.lockutils [req-fdb2d9f8-2b7b-4bde-9f32-7e14437c84ba req-c5712089-c4db-4d56-9d63-0aedb3468ae2 service nova] Lock "ab3edc53-37a5-4119-84c5-fd408ed66b69-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1065.044758] env[61923]: DEBUG nova.compute.manager [req-fdb2d9f8-2b7b-4bde-9f32-7e14437c84ba req-c5712089-c4db-4d56-9d63-0aedb3468ae2 service nova] [instance: ab3edc53-37a5-4119-84c5-fd408ed66b69] No waiting events found dispatching network-vif-plugged-f2c777e2-f8f4-437b-b990-20e42dd2491b {{(pid=61923) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1065.044987] env[61923]: WARNING nova.compute.manager [req-fdb2d9f8-2b7b-4bde-9f32-7e14437c84ba req-c5712089-c4db-4d56-9d63-0aedb3468ae2 service nova] [instance: ab3edc53-37a5-4119-84c5-fd408ed66b69] Received unexpected event network-vif-plugged-f2c777e2-f8f4-437b-b990-20e42dd2491b for instance with vm_state building and task_state spawning. [ 1065.129634] env[61923]: DEBUG nova.network.neutron [None req-2e0136c0-8e92-4295-b28a-93095da1d6f6 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: ab3edc53-37a5-4119-84c5-fd408ed66b69] Successfully updated port: f2c777e2-f8f4-437b-b990-20e42dd2491b {{(pid=61923) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1065.537256] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4510a6d3-c03f-49cb-99db-29593ff83e1d {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.545172] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17c33fed-cf3a-4e4c-8054-c2b3480d5c0d {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.575638] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e734cca-bd41-413d-b9ff-ff860e2d69ba {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.582840] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06884044-fcb1-447a-b8a9-f9abe7eccbb8 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.597125] env[61923]: DEBUG nova.compute.provider_tree [None req-26bb2be5-f543-490f-89c5-3e7df42ca732 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1065.631976] env[61923]: DEBUG oslo_concurrency.lockutils [None req-2e0136c0-8e92-4295-b28a-93095da1d6f6 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Acquiring lock "refresh_cache-ab3edc53-37a5-4119-84c5-fd408ed66b69" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1065.632151] env[61923]: DEBUG oslo_concurrency.lockutils [None req-2e0136c0-8e92-4295-b28a-93095da1d6f6 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Acquired lock "refresh_cache-ab3edc53-37a5-4119-84c5-fd408ed66b69" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1065.632280] env[61923]: DEBUG nova.network.neutron [None req-2e0136c0-8e92-4295-b28a-93095da1d6f6 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: ab3edc53-37a5-4119-84c5-fd408ed66b69] Building network info cache for instance {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1066.100441] env[61923]: DEBUG nova.scheduler.client.report [None req-26bb2be5-f543-490f-89c5-3e7df42ca732 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1066.164036] env[61923]: DEBUG nova.network.neutron [None req-2e0136c0-8e92-4295-b28a-93095da1d6f6 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: ab3edc53-37a5-4119-84c5-fd408ed66b69] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1066.286894] env[61923]: DEBUG nova.network.neutron [None req-2e0136c0-8e92-4295-b28a-93095da1d6f6 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: ab3edc53-37a5-4119-84c5-fd408ed66b69] Updating instance_info_cache with network_info: [{"id": "f2c777e2-f8f4-437b-b990-20e42dd2491b", "address": "fa:16:3e:b3:a4:a2", "network": {"id": "77aec6ca-d1be-4b39-9d51-08853a1183a1", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1813085414-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5a4db4f8aa194a219e80c538cc715d43", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "56b944d8-803d-43f2-945d-0f334ee4ea1c", "external-id": "nsx-vlan-transportzone-799", "segmentation_id": 799, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf2c777e2-f8", "ovs_interfaceid": "f2c777e2-f8f4-437b-b990-20e42dd2491b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1066.355291] env[61923]: DEBUG oslo_service.periodic_task [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Running periodic task ComputeManager._cleanup_incomplete_migrations {{(pid=61923) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1066.355457] env[61923]: DEBUG nova.compute.manager [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Cleaning up deleted instances with incomplete migration {{(pid=61923) _cleanup_incomplete_migrations /opt/stack/nova/nova/compute/manager.py:11266}} [ 1066.605214] env[61923]: DEBUG oslo_concurrency.lockutils [None req-26bb2be5-f543-490f-89c5-3e7df42ca732 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.128s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1066.605605] env[61923]: DEBUG nova.compute.manager [None req-26bb2be5-f543-490f-89c5-3e7df42ca732 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: 38b08866-9b44-4fc7-957b-d329715af18e] Start building networks asynchronously for instance. {{(pid=61923) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1066.789847] env[61923]: DEBUG oslo_concurrency.lockutils [None req-2e0136c0-8e92-4295-b28a-93095da1d6f6 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Releasing lock "refresh_cache-ab3edc53-37a5-4119-84c5-fd408ed66b69" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1066.790208] env[61923]: DEBUG nova.compute.manager [None req-2e0136c0-8e92-4295-b28a-93095da1d6f6 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: ab3edc53-37a5-4119-84c5-fd408ed66b69] Instance network_info: |[{"id": "f2c777e2-f8f4-437b-b990-20e42dd2491b", "address": "fa:16:3e:b3:a4:a2", "network": {"id": "77aec6ca-d1be-4b39-9d51-08853a1183a1", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1813085414-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5a4db4f8aa194a219e80c538cc715d43", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "56b944d8-803d-43f2-945d-0f334ee4ea1c", "external-id": "nsx-vlan-transportzone-799", "segmentation_id": 799, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf2c777e2-f8", "ovs_interfaceid": "f2c777e2-f8f4-437b-b990-20e42dd2491b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61923) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1066.790692] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-2e0136c0-8e92-4295-b28a-93095da1d6f6 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: ab3edc53-37a5-4119-84c5-fd408ed66b69] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:b3:a4:a2', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '56b944d8-803d-43f2-945d-0f334ee4ea1c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'f2c777e2-f8f4-437b-b990-20e42dd2491b', 'vif_model': 'vmxnet3'}] {{(pid=61923) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1066.798296] env[61923]: DEBUG oslo.service.loopingcall [None req-2e0136c0-8e92-4295-b28a-93095da1d6f6 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61923) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1066.798516] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ab3edc53-37a5-4119-84c5-fd408ed66b69] Creating VM on the ESX host {{(pid=61923) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1066.798769] env[61923]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-216151f2-14d5-4d43-9138-395fa2648d93 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.818646] env[61923]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1066.818646] env[61923]: value = "task-1377995" [ 1066.818646] env[61923]: _type = "Task" [ 1066.818646] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1066.826734] env[61923]: DEBUG oslo_vmware.api [-] Task: {'id': task-1377995, 'name': CreateVM_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1067.068416] env[61923]: DEBUG nova.compute.manager [req-8958c214-363a-4b1d-aa04-0b2a5f8a4bab req-21d76aa5-fd09-48a5-b5f0-332e3447b509 service nova] [instance: ab3edc53-37a5-4119-84c5-fd408ed66b69] Received event network-changed-f2c777e2-f8f4-437b-b990-20e42dd2491b {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1067.068651] env[61923]: DEBUG nova.compute.manager [req-8958c214-363a-4b1d-aa04-0b2a5f8a4bab req-21d76aa5-fd09-48a5-b5f0-332e3447b509 service nova] [instance: ab3edc53-37a5-4119-84c5-fd408ed66b69] Refreshing instance network info cache due to event network-changed-f2c777e2-f8f4-437b-b990-20e42dd2491b. {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1067.068873] env[61923]: DEBUG oslo_concurrency.lockutils [req-8958c214-363a-4b1d-aa04-0b2a5f8a4bab req-21d76aa5-fd09-48a5-b5f0-332e3447b509 service nova] Acquiring lock "refresh_cache-ab3edc53-37a5-4119-84c5-fd408ed66b69" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1067.069160] env[61923]: DEBUG oslo_concurrency.lockutils [req-8958c214-363a-4b1d-aa04-0b2a5f8a4bab req-21d76aa5-fd09-48a5-b5f0-332e3447b509 service nova] Acquired lock "refresh_cache-ab3edc53-37a5-4119-84c5-fd408ed66b69" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1067.069415] env[61923]: DEBUG nova.network.neutron [req-8958c214-363a-4b1d-aa04-0b2a5f8a4bab req-21d76aa5-fd09-48a5-b5f0-332e3447b509 service nova] [instance: ab3edc53-37a5-4119-84c5-fd408ed66b69] Refreshing network info cache for port f2c777e2-f8f4-437b-b990-20e42dd2491b {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1067.109821] env[61923]: DEBUG nova.compute.utils [None req-26bb2be5-f543-490f-89c5-3e7df42ca732 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Using /dev/sd instead of None {{(pid=61923) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1067.111149] env[61923]: DEBUG nova.compute.manager [None req-26bb2be5-f543-490f-89c5-3e7df42ca732 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: 38b08866-9b44-4fc7-957b-d329715af18e] Allocating IP information in the background. {{(pid=61923) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1067.111312] env[61923]: DEBUG nova.network.neutron [None req-26bb2be5-f543-490f-89c5-3e7df42ca732 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: 38b08866-9b44-4fc7-957b-d329715af18e] allocate_for_instance() {{(pid=61923) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1067.150160] env[61923]: DEBUG nova.policy [None req-26bb2be5-f543-490f-89c5-3e7df42ca732 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '69be89006ff141b686261d643b3b9a73', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '780d3541d9604417b977bb62390c4299', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61923) authorize /opt/stack/nova/nova/policy.py:201}} [ 1067.328279] env[61923]: DEBUG oslo_vmware.api [-] Task: {'id': task-1377995, 'name': CreateVM_Task, 'duration_secs': 0.293634} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1067.328493] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ab3edc53-37a5-4119-84c5-fd408ed66b69] Created VM on the ESX host {{(pid=61923) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1067.329211] env[61923]: DEBUG oslo_concurrency.lockutils [None req-2e0136c0-8e92-4295-b28a-93095da1d6f6 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1067.329439] env[61923]: DEBUG oslo_concurrency.lockutils [None req-2e0136c0-8e92-4295-b28a-93095da1d6f6 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1067.329740] env[61923]: DEBUG oslo_concurrency.lockutils [None req-2e0136c0-8e92-4295-b28a-93095da1d6f6 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1067.330286] env[61923]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cd342aa0-3a14-4291-bd24-cda101844da5 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.334449] env[61923]: DEBUG oslo_vmware.api [None req-2e0136c0-8e92-4295-b28a-93095da1d6f6 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Waiting for the task: (returnval){ [ 1067.334449] env[61923]: value = "session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52b03f99-f041-9e02-d2b6-c9da4f88bf96" [ 1067.334449] env[61923]: _type = "Task" [ 1067.334449] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1067.341448] env[61923]: DEBUG oslo_vmware.api [None req-2e0136c0-8e92-4295-b28a-93095da1d6f6 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52b03f99-f041-9e02-d2b6-c9da4f88bf96, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1067.390406] env[61923]: DEBUG nova.network.neutron [None req-26bb2be5-f543-490f-89c5-3e7df42ca732 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: 38b08866-9b44-4fc7-957b-d329715af18e] Successfully created port: d5c39dc6-35e2-4535-941d-f5673169edca {{(pid=61923) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1067.614723] env[61923]: DEBUG nova.compute.manager [None req-26bb2be5-f543-490f-89c5-3e7df42ca732 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: 38b08866-9b44-4fc7-957b-d329715af18e] Start building block device mappings for instance. {{(pid=61923) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1067.773453] env[61923]: DEBUG nova.network.neutron [req-8958c214-363a-4b1d-aa04-0b2a5f8a4bab req-21d76aa5-fd09-48a5-b5f0-332e3447b509 service nova] [instance: ab3edc53-37a5-4119-84c5-fd408ed66b69] Updated VIF entry in instance network info cache for port f2c777e2-f8f4-437b-b990-20e42dd2491b. {{(pid=61923) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1067.773794] env[61923]: DEBUG nova.network.neutron [req-8958c214-363a-4b1d-aa04-0b2a5f8a4bab req-21d76aa5-fd09-48a5-b5f0-332e3447b509 service nova] [instance: ab3edc53-37a5-4119-84c5-fd408ed66b69] Updating instance_info_cache with network_info: [{"id": "f2c777e2-f8f4-437b-b990-20e42dd2491b", "address": "fa:16:3e:b3:a4:a2", "network": {"id": "77aec6ca-d1be-4b39-9d51-08853a1183a1", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1813085414-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5a4db4f8aa194a219e80c538cc715d43", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "56b944d8-803d-43f2-945d-0f334ee4ea1c", "external-id": "nsx-vlan-transportzone-799", "segmentation_id": 799, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf2c777e2-f8", "ovs_interfaceid": "f2c777e2-f8f4-437b-b990-20e42dd2491b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1067.844844] env[61923]: DEBUG oslo_vmware.api [None req-2e0136c0-8e92-4295-b28a-93095da1d6f6 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52b03f99-f041-9e02-d2b6-c9da4f88bf96, 'name': SearchDatastore_Task, 'duration_secs': 0.00913} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1067.845181] env[61923]: DEBUG oslo_concurrency.lockutils [None req-2e0136c0-8e92-4295-b28a-93095da1d6f6 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1067.845455] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-2e0136c0-8e92-4295-b28a-93095da1d6f6 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: ab3edc53-37a5-4119-84c5-fd408ed66b69] Processing image 0f153f63-ae0a-45b1-b7f5-7f9b673c947f {{(pid=61923) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1067.845721] env[61923]: DEBUG oslo_concurrency.lockutils [None req-2e0136c0-8e92-4295-b28a-93095da1d6f6 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1067.845885] env[61923]: DEBUG oslo_concurrency.lockutils [None req-2e0136c0-8e92-4295-b28a-93095da1d6f6 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1067.846108] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-2e0136c0-8e92-4295-b28a-93095da1d6f6 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61923) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1067.846405] env[61923]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-695598da-417f-49ad-ae66-15fd64db012d {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.856053] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-2e0136c0-8e92-4295-b28a-93095da1d6f6 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61923) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1067.856053] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-2e0136c0-8e92-4295-b28a-93095da1d6f6 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61923) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1067.856053] env[61923]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b11cc45a-42dc-4b32-aefb-bbd1ea449671 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.861108] env[61923]: DEBUG oslo_vmware.api [None req-2e0136c0-8e92-4295-b28a-93095da1d6f6 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Waiting for the task: (returnval){ [ 1067.861108] env[61923]: value = "session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]529d750b-0731-de93-778f-14db4b29491f" [ 1067.861108] env[61923]: _type = "Task" [ 1067.861108] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1067.868079] env[61923]: DEBUG oslo_vmware.api [None req-2e0136c0-8e92-4295-b28a-93095da1d6f6 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]529d750b-0731-de93-778f-14db4b29491f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1068.276927] env[61923]: DEBUG oslo_concurrency.lockutils [req-8958c214-363a-4b1d-aa04-0b2a5f8a4bab req-21d76aa5-fd09-48a5-b5f0-332e3447b509 service nova] Releasing lock "refresh_cache-ab3edc53-37a5-4119-84c5-fd408ed66b69" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1068.371796] env[61923]: DEBUG oslo_vmware.api [None req-2e0136c0-8e92-4295-b28a-93095da1d6f6 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]529d750b-0731-de93-778f-14db4b29491f, 'name': SearchDatastore_Task, 'duration_secs': 0.007468} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1068.372599] env[61923]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c208bfba-d284-4fff-8038-881520488a4b {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.378120] env[61923]: DEBUG oslo_vmware.api [None req-2e0136c0-8e92-4295-b28a-93095da1d6f6 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Waiting for the task: (returnval){ [ 1068.378120] env[61923]: value = "session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]527d12d8-a44d-fa1d-23db-c68d259985ed" [ 1068.378120] env[61923]: _type = "Task" [ 1068.378120] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1068.385172] env[61923]: DEBUG oslo_vmware.api [None req-2e0136c0-8e92-4295-b28a-93095da1d6f6 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]527d12d8-a44d-fa1d-23db-c68d259985ed, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1068.623239] env[61923]: DEBUG nova.compute.manager [None req-26bb2be5-f543-490f-89c5-3e7df42ca732 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: 38b08866-9b44-4fc7-957b-d329715af18e] Start spawning the instance on the hypervisor. {{(pid=61923) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1068.649663] env[61923]: DEBUG nova.virt.hardware [None req-26bb2be5-f543-490f-89c5-3e7df42ca732 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-29T20:07:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-29T20:07:35Z,direct_url=,disk_format='vmdk',id=0f153f63-ae0a-45b1-b7f5-7f9b673c947f,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='4e7b5e3b3c3443c8bd5c70f8a15739f5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-29T20:07:36Z,virtual_size=,visibility=), allow threads: False {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1068.649921] env[61923]: DEBUG nova.virt.hardware [None req-26bb2be5-f543-490f-89c5-3e7df42ca732 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Flavor limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1068.650092] env[61923]: DEBUG nova.virt.hardware [None req-26bb2be5-f543-490f-89c5-3e7df42ca732 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Image limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1068.650279] env[61923]: DEBUG nova.virt.hardware [None req-26bb2be5-f543-490f-89c5-3e7df42ca732 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Flavor pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1068.650425] env[61923]: DEBUG nova.virt.hardware [None req-26bb2be5-f543-490f-89c5-3e7df42ca732 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Image pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1068.650571] env[61923]: DEBUG nova.virt.hardware [None req-26bb2be5-f543-490f-89c5-3e7df42ca732 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1068.650810] env[61923]: DEBUG nova.virt.hardware [None req-26bb2be5-f543-490f-89c5-3e7df42ca732 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1068.650985] env[61923]: DEBUG nova.virt.hardware [None req-26bb2be5-f543-490f-89c5-3e7df42ca732 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1068.651174] env[61923]: DEBUG nova.virt.hardware [None req-26bb2be5-f543-490f-89c5-3e7df42ca732 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Got 1 possible topologies {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1068.651338] env[61923]: DEBUG nova.virt.hardware [None req-26bb2be5-f543-490f-89c5-3e7df42ca732 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1068.651512] env[61923]: DEBUG nova.virt.hardware [None req-26bb2be5-f543-490f-89c5-3e7df42ca732 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1068.652385] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5fb089e7-38ce-40c9-8abe-8155e110daf1 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.660153] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee0f805a-22eb-4beb-bd3d-f126ffb08cbc {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.836789] env[61923]: DEBUG nova.network.neutron [None req-26bb2be5-f543-490f-89c5-3e7df42ca732 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: 38b08866-9b44-4fc7-957b-d329715af18e] Successfully updated port: d5c39dc6-35e2-4535-941d-f5673169edca {{(pid=61923) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1068.890203] env[61923]: DEBUG oslo_vmware.api [None req-2e0136c0-8e92-4295-b28a-93095da1d6f6 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]527d12d8-a44d-fa1d-23db-c68d259985ed, 'name': SearchDatastore_Task, 'duration_secs': 0.009392} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1068.890499] env[61923]: DEBUG oslo_concurrency.lockutils [None req-2e0136c0-8e92-4295-b28a-93095da1d6f6 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1068.890801] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-2e0136c0-8e92-4295-b28a-93095da1d6f6 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk to [datastore1] ab3edc53-37a5-4119-84c5-fd408ed66b69/ab3edc53-37a5-4119-84c5-fd408ed66b69.vmdk {{(pid=61923) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1068.891078] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d49657cc-2abf-454f-b5b3-111e7fcd65bd {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.897380] env[61923]: DEBUG oslo_vmware.api [None req-2e0136c0-8e92-4295-b28a-93095da1d6f6 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Waiting for the task: (returnval){ [ 1068.897380] env[61923]: value = "task-1377996" [ 1068.897380] env[61923]: _type = "Task" [ 1068.897380] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1068.905281] env[61923]: DEBUG oslo_vmware.api [None req-2e0136c0-8e92-4295-b28a-93095da1d6f6 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Task: {'id': task-1377996, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1069.094470] env[61923]: DEBUG nova.compute.manager [req-b88ef97a-6b3a-41fa-b783-f68a08ed0ea3 req-efcca927-78a1-411e-ab1f-e5ff5f359cb0 service nova] [instance: 38b08866-9b44-4fc7-957b-d329715af18e] Received event network-vif-plugged-d5c39dc6-35e2-4535-941d-f5673169edca {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1069.094657] env[61923]: DEBUG oslo_concurrency.lockutils [req-b88ef97a-6b3a-41fa-b783-f68a08ed0ea3 req-efcca927-78a1-411e-ab1f-e5ff5f359cb0 service nova] Acquiring lock "38b08866-9b44-4fc7-957b-d329715af18e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1069.094856] env[61923]: DEBUG oslo_concurrency.lockutils [req-b88ef97a-6b3a-41fa-b783-f68a08ed0ea3 req-efcca927-78a1-411e-ab1f-e5ff5f359cb0 service nova] Lock "38b08866-9b44-4fc7-957b-d329715af18e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1069.095184] env[61923]: DEBUG oslo_concurrency.lockutils [req-b88ef97a-6b3a-41fa-b783-f68a08ed0ea3 req-efcca927-78a1-411e-ab1f-e5ff5f359cb0 service nova] Lock "38b08866-9b44-4fc7-957b-d329715af18e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1069.095397] env[61923]: DEBUG nova.compute.manager [req-b88ef97a-6b3a-41fa-b783-f68a08ed0ea3 req-efcca927-78a1-411e-ab1f-e5ff5f359cb0 service nova] [instance: 38b08866-9b44-4fc7-957b-d329715af18e] No waiting events found dispatching network-vif-plugged-d5c39dc6-35e2-4535-941d-f5673169edca {{(pid=61923) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1069.095573] env[61923]: WARNING nova.compute.manager [req-b88ef97a-6b3a-41fa-b783-f68a08ed0ea3 req-efcca927-78a1-411e-ab1f-e5ff5f359cb0 service nova] [instance: 38b08866-9b44-4fc7-957b-d329715af18e] Received unexpected event network-vif-plugged-d5c39dc6-35e2-4535-941d-f5673169edca for instance with vm_state building and task_state spawning. [ 1069.095740] env[61923]: DEBUG nova.compute.manager [req-b88ef97a-6b3a-41fa-b783-f68a08ed0ea3 req-efcca927-78a1-411e-ab1f-e5ff5f359cb0 service nova] [instance: 38b08866-9b44-4fc7-957b-d329715af18e] Received event network-changed-d5c39dc6-35e2-4535-941d-f5673169edca {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1069.095899] env[61923]: DEBUG nova.compute.manager [req-b88ef97a-6b3a-41fa-b783-f68a08ed0ea3 req-efcca927-78a1-411e-ab1f-e5ff5f359cb0 service nova] [instance: 38b08866-9b44-4fc7-957b-d329715af18e] Refreshing instance network info cache due to event network-changed-d5c39dc6-35e2-4535-941d-f5673169edca. {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1069.096101] env[61923]: DEBUG oslo_concurrency.lockutils [req-b88ef97a-6b3a-41fa-b783-f68a08ed0ea3 req-efcca927-78a1-411e-ab1f-e5ff5f359cb0 service nova] Acquiring lock "refresh_cache-38b08866-9b44-4fc7-957b-d329715af18e" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1069.096240] env[61923]: DEBUG oslo_concurrency.lockutils [req-b88ef97a-6b3a-41fa-b783-f68a08ed0ea3 req-efcca927-78a1-411e-ab1f-e5ff5f359cb0 service nova] Acquired lock "refresh_cache-38b08866-9b44-4fc7-957b-d329715af18e" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1069.096400] env[61923]: DEBUG nova.network.neutron [req-b88ef97a-6b3a-41fa-b783-f68a08ed0ea3 req-efcca927-78a1-411e-ab1f-e5ff5f359cb0 service nova] [instance: 38b08866-9b44-4fc7-957b-d329715af18e] Refreshing network info cache for port d5c39dc6-35e2-4535-941d-f5673169edca {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1069.340375] env[61923]: DEBUG oslo_concurrency.lockutils [None req-26bb2be5-f543-490f-89c5-3e7df42ca732 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Acquiring lock "refresh_cache-38b08866-9b44-4fc7-957b-d329715af18e" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1069.407018] env[61923]: DEBUG oslo_vmware.api [None req-2e0136c0-8e92-4295-b28a-93095da1d6f6 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Task: {'id': task-1377996, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.465905} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1069.407393] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-2e0136c0-8e92-4295-b28a-93095da1d6f6 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk to [datastore1] ab3edc53-37a5-4119-84c5-fd408ed66b69/ab3edc53-37a5-4119-84c5-fd408ed66b69.vmdk {{(pid=61923) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1069.407622] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-2e0136c0-8e92-4295-b28a-93095da1d6f6 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: ab3edc53-37a5-4119-84c5-fd408ed66b69] Extending root virtual disk to 1048576 {{(pid=61923) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1069.407870] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-7ebe661f-65cf-4c8d-9da4-d886ce2fea6c {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.413848] env[61923]: DEBUG oslo_vmware.api [None req-2e0136c0-8e92-4295-b28a-93095da1d6f6 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Waiting for the task: (returnval){ [ 1069.413848] env[61923]: value = "task-1377997" [ 1069.413848] env[61923]: _type = "Task" [ 1069.413848] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1069.420844] env[61923]: DEBUG oslo_vmware.api [None req-2e0136c0-8e92-4295-b28a-93095da1d6f6 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Task: {'id': task-1377997, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1069.627214] env[61923]: DEBUG nova.network.neutron [req-b88ef97a-6b3a-41fa-b783-f68a08ed0ea3 req-efcca927-78a1-411e-ab1f-e5ff5f359cb0 service nova] [instance: 38b08866-9b44-4fc7-957b-d329715af18e] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1069.693408] env[61923]: DEBUG nova.network.neutron [req-b88ef97a-6b3a-41fa-b783-f68a08ed0ea3 req-efcca927-78a1-411e-ab1f-e5ff5f359cb0 service nova] [instance: 38b08866-9b44-4fc7-957b-d329715af18e] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1069.923692] env[61923]: DEBUG oslo_vmware.api [None req-2e0136c0-8e92-4295-b28a-93095da1d6f6 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Task: {'id': task-1377997, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.100646} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1069.925184] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-2e0136c0-8e92-4295-b28a-93095da1d6f6 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: ab3edc53-37a5-4119-84c5-fd408ed66b69] Extended root virtual disk {{(pid=61923) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1069.925184] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41478b19-289c-4405-b72f-4b10ce9bd0da {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.945584] env[61923]: DEBUG nova.virt.vmwareapi.volumeops [None req-2e0136c0-8e92-4295-b28a-93095da1d6f6 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: ab3edc53-37a5-4119-84c5-fd408ed66b69] Reconfiguring VM instance instance-00000063 to attach disk [datastore1] ab3edc53-37a5-4119-84c5-fd408ed66b69/ab3edc53-37a5-4119-84c5-fd408ed66b69.vmdk or device None with type sparse {{(pid=61923) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1069.945810] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-fe5849d1-f353-4c57-b2a3-9c0e1b7be090 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.964384] env[61923]: DEBUG oslo_vmware.api [None req-2e0136c0-8e92-4295-b28a-93095da1d6f6 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Waiting for the task: (returnval){ [ 1069.964384] env[61923]: value = "task-1377998" [ 1069.964384] env[61923]: _type = "Task" [ 1069.964384] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1069.971460] env[61923]: DEBUG oslo_vmware.api [None req-2e0136c0-8e92-4295-b28a-93095da1d6f6 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Task: {'id': task-1377998, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1070.196350] env[61923]: DEBUG oslo_concurrency.lockutils [req-b88ef97a-6b3a-41fa-b783-f68a08ed0ea3 req-efcca927-78a1-411e-ab1f-e5ff5f359cb0 service nova] Releasing lock "refresh_cache-38b08866-9b44-4fc7-957b-d329715af18e" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1070.196782] env[61923]: DEBUG oslo_concurrency.lockutils [None req-26bb2be5-f543-490f-89c5-3e7df42ca732 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Acquired lock "refresh_cache-38b08866-9b44-4fc7-957b-d329715af18e" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1070.196961] env[61923]: DEBUG nova.network.neutron [None req-26bb2be5-f543-490f-89c5-3e7df42ca732 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: 38b08866-9b44-4fc7-957b-d329715af18e] Building network info cache for instance {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1070.355546] env[61923]: DEBUG oslo_service.periodic_task [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Running periodic task ComputeManager._cleanup_expired_console_auth_tokens {{(pid=61923) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1070.474224] env[61923]: DEBUG oslo_vmware.api [None req-2e0136c0-8e92-4295-b28a-93095da1d6f6 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Task: {'id': task-1377998, 'name': ReconfigVM_Task, 'duration_secs': 0.257978} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1070.474549] env[61923]: DEBUG nova.virt.vmwareapi.volumeops [None req-2e0136c0-8e92-4295-b28a-93095da1d6f6 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: ab3edc53-37a5-4119-84c5-fd408ed66b69] Reconfigured VM instance instance-00000063 to attach disk [datastore1] ab3edc53-37a5-4119-84c5-fd408ed66b69/ab3edc53-37a5-4119-84c5-fd408ed66b69.vmdk or device None with type sparse {{(pid=61923) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1070.475153] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-39bb236b-19c9-4c46-b35d-766564d9e385 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.485934] env[61923]: DEBUG oslo_vmware.api [None req-2e0136c0-8e92-4295-b28a-93095da1d6f6 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Waiting for the task: (returnval){ [ 1070.485934] env[61923]: value = "task-1377999" [ 1070.485934] env[61923]: _type = "Task" [ 1070.485934] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1070.494583] env[61923]: DEBUG oslo_vmware.api [None req-2e0136c0-8e92-4295-b28a-93095da1d6f6 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Task: {'id': task-1377999, 'name': Rename_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1070.728164] env[61923]: DEBUG nova.network.neutron [None req-26bb2be5-f543-490f-89c5-3e7df42ca732 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: 38b08866-9b44-4fc7-957b-d329715af18e] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1070.848027] env[61923]: DEBUG nova.network.neutron [None req-26bb2be5-f543-490f-89c5-3e7df42ca732 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: 38b08866-9b44-4fc7-957b-d329715af18e] Updating instance_info_cache with network_info: [{"id": "d5c39dc6-35e2-4535-941d-f5673169edca", "address": "fa:16:3e:92:f6:bb", "network": {"id": "9533e79d-fde2-4c10-86a0-86a36845a8cf", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1736743137-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "780d3541d9604417b977bb62390c4299", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6d650b26-c3e7-4de7-98db-5e4b816d123a", "external-id": "nsx-vlan-transportzone-757", "segmentation_id": 757, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd5c39dc6-35", "ovs_interfaceid": "d5c39dc6-35e2-4535-941d-f5673169edca", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1070.995988] env[61923]: DEBUG oslo_vmware.api [None req-2e0136c0-8e92-4295-b28a-93095da1d6f6 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Task: {'id': task-1377999, 'name': Rename_Task, 'duration_secs': 0.158045} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1070.996221] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-2e0136c0-8e92-4295-b28a-93095da1d6f6 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: ab3edc53-37a5-4119-84c5-fd408ed66b69] Powering on the VM {{(pid=61923) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1070.996450] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-336f27f2-1c7b-49e9-8740-fbede81f7ad2 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.002576] env[61923]: DEBUG oslo_vmware.api [None req-2e0136c0-8e92-4295-b28a-93095da1d6f6 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Waiting for the task: (returnval){ [ 1071.002576] env[61923]: value = "task-1378000" [ 1071.002576] env[61923]: _type = "Task" [ 1071.002576] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1071.009333] env[61923]: DEBUG oslo_vmware.api [None req-2e0136c0-8e92-4295-b28a-93095da1d6f6 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Task: {'id': task-1378000, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1071.350476] env[61923]: DEBUG oslo_concurrency.lockutils [None req-26bb2be5-f543-490f-89c5-3e7df42ca732 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Releasing lock "refresh_cache-38b08866-9b44-4fc7-957b-d329715af18e" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1071.350770] env[61923]: DEBUG nova.compute.manager [None req-26bb2be5-f543-490f-89c5-3e7df42ca732 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: 38b08866-9b44-4fc7-957b-d329715af18e] Instance network_info: |[{"id": "d5c39dc6-35e2-4535-941d-f5673169edca", "address": "fa:16:3e:92:f6:bb", "network": {"id": "9533e79d-fde2-4c10-86a0-86a36845a8cf", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1736743137-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "780d3541d9604417b977bb62390c4299", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6d650b26-c3e7-4de7-98db-5e4b816d123a", "external-id": "nsx-vlan-transportzone-757", "segmentation_id": 757, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd5c39dc6-35", "ovs_interfaceid": "d5c39dc6-35e2-4535-941d-f5673169edca", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61923) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1071.351243] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-26bb2be5-f543-490f-89c5-3e7df42ca732 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: 38b08866-9b44-4fc7-957b-d329715af18e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:92:f6:bb', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '6d650b26-c3e7-4de7-98db-5e4b816d123a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'd5c39dc6-35e2-4535-941d-f5673169edca', 'vif_model': 'vmxnet3'}] {{(pid=61923) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1071.358491] env[61923]: DEBUG oslo.service.loopingcall [None req-26bb2be5-f543-490f-89c5-3e7df42ca732 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61923) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1071.358702] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 38b08866-9b44-4fc7-957b-d329715af18e] Creating VM on the ESX host {{(pid=61923) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1071.358916] env[61923]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-67a908e1-8ff0-4ce8-9b91-2d33d7f99abc {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.378963] env[61923]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1071.378963] env[61923]: value = "task-1378001" [ 1071.378963] env[61923]: _type = "Task" [ 1071.378963] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1071.386975] env[61923]: DEBUG oslo_vmware.api [-] Task: {'id': task-1378001, 'name': CreateVM_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1071.512604] env[61923]: DEBUG oslo_vmware.api [None req-2e0136c0-8e92-4295-b28a-93095da1d6f6 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Task: {'id': task-1378000, 'name': PowerOnVM_Task, 'duration_secs': 0.498711} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1071.512858] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-2e0136c0-8e92-4295-b28a-93095da1d6f6 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: ab3edc53-37a5-4119-84c5-fd408ed66b69] Powered on the VM {{(pid=61923) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1071.513095] env[61923]: INFO nova.compute.manager [None req-2e0136c0-8e92-4295-b28a-93095da1d6f6 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: ab3edc53-37a5-4119-84c5-fd408ed66b69] Took 6.65 seconds to spawn the instance on the hypervisor. [ 1071.513270] env[61923]: DEBUG nova.compute.manager [None req-2e0136c0-8e92-4295-b28a-93095da1d6f6 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: ab3edc53-37a5-4119-84c5-fd408ed66b69] Checking state {{(pid=61923) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1071.514038] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d910b117-a296-4182-824a-05fcb7b34b8d {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.857740] env[61923]: DEBUG oslo_service.periodic_task [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61923) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1071.858189] env[61923]: DEBUG nova.compute.manager [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61923) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 1071.888183] env[61923]: DEBUG oslo_vmware.api [-] Task: {'id': task-1378001, 'name': CreateVM_Task, 'duration_secs': 0.29641} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1071.888338] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 38b08866-9b44-4fc7-957b-d329715af18e] Created VM on the ESX host {{(pid=61923) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1071.889011] env[61923]: DEBUG oslo_concurrency.lockutils [None req-26bb2be5-f543-490f-89c5-3e7df42ca732 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1071.889187] env[61923]: DEBUG oslo_concurrency.lockutils [None req-26bb2be5-f543-490f-89c5-3e7df42ca732 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1071.889514] env[61923]: DEBUG oslo_concurrency.lockutils [None req-26bb2be5-f543-490f-89c5-3e7df42ca732 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1071.889759] env[61923]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-aa5b849c-a294-44f6-8f6f-a1860e0de8a2 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.893785] env[61923]: DEBUG oslo_vmware.api [None req-26bb2be5-f543-490f-89c5-3e7df42ca732 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Waiting for the task: (returnval){ [ 1071.893785] env[61923]: value = "session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52db5678-4fbc-aacf-f119-ce4a51196377" [ 1071.893785] env[61923]: _type = "Task" [ 1071.893785] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1071.900525] env[61923]: DEBUG oslo_vmware.api [None req-26bb2be5-f543-490f-89c5-3e7df42ca732 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52db5678-4fbc-aacf-f119-ce4a51196377, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1072.029647] env[61923]: INFO nova.compute.manager [None req-2e0136c0-8e92-4295-b28a-93095da1d6f6 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: ab3edc53-37a5-4119-84c5-fd408ed66b69] Took 12.34 seconds to build instance. [ 1072.356293] env[61923]: DEBUG oslo_service.periodic_task [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61923) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1072.356508] env[61923]: DEBUG nova.compute.manager [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Starting heal instance info cache {{(pid=61923) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 1072.356594] env[61923]: DEBUG nova.compute.manager [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Rebuilding the list of instances to heal {{(pid=61923) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 1072.404214] env[61923]: DEBUG oslo_vmware.api [None req-26bb2be5-f543-490f-89c5-3e7df42ca732 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52db5678-4fbc-aacf-f119-ce4a51196377, 'name': SearchDatastore_Task, 'duration_secs': 0.009185} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1072.404531] env[61923]: DEBUG oslo_concurrency.lockutils [None req-26bb2be5-f543-490f-89c5-3e7df42ca732 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1072.404770] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-26bb2be5-f543-490f-89c5-3e7df42ca732 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: 38b08866-9b44-4fc7-957b-d329715af18e] Processing image 0f153f63-ae0a-45b1-b7f5-7f9b673c947f {{(pid=61923) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1072.405017] env[61923]: DEBUG oslo_concurrency.lockutils [None req-26bb2be5-f543-490f-89c5-3e7df42ca732 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1072.405179] env[61923]: DEBUG oslo_concurrency.lockutils [None req-26bb2be5-f543-490f-89c5-3e7df42ca732 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1072.405359] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-26bb2be5-f543-490f-89c5-3e7df42ca732 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61923) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1072.405620] env[61923]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-9a8a4550-ddc0-4cde-b013-ade593378abd {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.414072] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-26bb2be5-f543-490f-89c5-3e7df42ca732 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61923) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1072.414266] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-26bb2be5-f543-490f-89c5-3e7df42ca732 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61923) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1072.414969] env[61923]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-20a3fb06-6c13-468e-932e-0c9db5953345 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.419939] env[61923]: DEBUG oslo_vmware.api [None req-26bb2be5-f543-490f-89c5-3e7df42ca732 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Waiting for the task: (returnval){ [ 1072.419939] env[61923]: value = "session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]522d4438-8b82-c670-4881-ca129de4fa2a" [ 1072.419939] env[61923]: _type = "Task" [ 1072.419939] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1072.427635] env[61923]: DEBUG oslo_vmware.api [None req-26bb2be5-f543-490f-89c5-3e7df42ca732 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]522d4438-8b82-c670-4881-ca129de4fa2a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1072.532362] env[61923]: DEBUG oslo_concurrency.lockutils [None req-2e0136c0-8e92-4295-b28a-93095da1d6f6 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Lock "ab3edc53-37a5-4119-84c5-fd408ed66b69" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 13.850s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1072.860785] env[61923]: DEBUG nova.compute.manager [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] [instance: 38b08866-9b44-4fc7-957b-d329715af18e] Skipping network cache update for instance because it is Building. {{(pid=61923) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 1072.897732] env[61923]: DEBUG oslo_concurrency.lockutils [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Acquiring lock "refresh_cache-35bdbd45-5838-4b09-a232-7f4a02322ad7" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1072.897883] env[61923]: DEBUG oslo_concurrency.lockutils [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Acquired lock "refresh_cache-35bdbd45-5838-4b09-a232-7f4a02322ad7" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1072.898050] env[61923]: DEBUG nova.network.neutron [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] [instance: 35bdbd45-5838-4b09-a232-7f4a02322ad7] Forcefully refreshing network info cache for instance {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 1072.898209] env[61923]: DEBUG nova.objects.instance [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Lazy-loading 'info_cache' on Instance uuid 35bdbd45-5838-4b09-a232-7f4a02322ad7 {{(pid=61923) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1072.930663] env[61923]: DEBUG oslo_vmware.api [None req-26bb2be5-f543-490f-89c5-3e7df42ca732 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]522d4438-8b82-c670-4881-ca129de4fa2a, 'name': SearchDatastore_Task, 'duration_secs': 0.012946} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1072.931489] env[61923]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d5f612d7-e723-484d-978a-3b0f9a99ba04 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.936844] env[61923]: DEBUG oslo_vmware.api [None req-26bb2be5-f543-490f-89c5-3e7df42ca732 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Waiting for the task: (returnval){ [ 1072.936844] env[61923]: value = "session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]5211f33c-9d49-da6d-0e97-c8bf9f10fa40" [ 1072.936844] env[61923]: _type = "Task" [ 1072.936844] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1072.945047] env[61923]: DEBUG oslo_vmware.api [None req-26bb2be5-f543-490f-89c5-3e7df42ca732 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]5211f33c-9d49-da6d-0e97-c8bf9f10fa40, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1073.447487] env[61923]: DEBUG oslo_vmware.api [None req-26bb2be5-f543-490f-89c5-3e7df42ca732 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]5211f33c-9d49-da6d-0e97-c8bf9f10fa40, 'name': SearchDatastore_Task, 'duration_secs': 0.00926} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1073.447709] env[61923]: DEBUG oslo_concurrency.lockutils [None req-26bb2be5-f543-490f-89c5-3e7df42ca732 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1073.447969] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-26bb2be5-f543-490f-89c5-3e7df42ca732 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk to [datastore1] 38b08866-9b44-4fc7-957b-d329715af18e/38b08866-9b44-4fc7-957b-d329715af18e.vmdk {{(pid=61923) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1073.448246] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-7d983bd7-4e97-401d-b1a1-739351556e2a {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.454439] env[61923]: DEBUG oslo_vmware.api [None req-26bb2be5-f543-490f-89c5-3e7df42ca732 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Waiting for the task: (returnval){ [ 1073.454439] env[61923]: value = "task-1378002" [ 1073.454439] env[61923]: _type = "Task" [ 1073.454439] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1073.462176] env[61923]: DEBUG oslo_vmware.api [None req-26bb2be5-f543-490f-89c5-3e7df42ca732 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Task: {'id': task-1378002, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1073.963594] env[61923]: DEBUG oslo_vmware.api [None req-26bb2be5-f543-490f-89c5-3e7df42ca732 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Task: {'id': task-1378002, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.479408} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1073.963948] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-26bb2be5-f543-490f-89c5-3e7df42ca732 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk to [datastore1] 38b08866-9b44-4fc7-957b-d329715af18e/38b08866-9b44-4fc7-957b-d329715af18e.vmdk {{(pid=61923) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1073.964113] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-26bb2be5-f543-490f-89c5-3e7df42ca732 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: 38b08866-9b44-4fc7-957b-d329715af18e] Extending root virtual disk to 1048576 {{(pid=61923) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1073.964302] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-3d9dd149-dc53-4350-a6c0-277c20a14405 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.970350] env[61923]: DEBUG oslo_vmware.api [None req-26bb2be5-f543-490f-89c5-3e7df42ca732 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Waiting for the task: (returnval){ [ 1073.970350] env[61923]: value = "task-1378003" [ 1073.970350] env[61923]: _type = "Task" [ 1073.970350] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1073.977425] env[61923]: DEBUG oslo_vmware.api [None req-26bb2be5-f543-490f-89c5-3e7df42ca732 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Task: {'id': task-1378003, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1074.034923] env[61923]: DEBUG nova.compute.manager [None req-e5fe1a98-bb9f-4924-80b3-7de0d4805067 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: ab3edc53-37a5-4119-84c5-fd408ed66b69] Stashing vm_state: active {{(pid=61923) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 1074.479827] env[61923]: DEBUG oslo_vmware.api [None req-26bb2be5-f543-490f-89c5-3e7df42ca732 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Task: {'id': task-1378003, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.068236} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1074.480113] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-26bb2be5-f543-490f-89c5-3e7df42ca732 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: 38b08866-9b44-4fc7-957b-d329715af18e] Extended root virtual disk {{(pid=61923) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1074.480929] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a1ce4b2-b869-4a57-9ad3-b4e534cd4c2e {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.503645] env[61923]: DEBUG nova.virt.vmwareapi.volumeops [None req-26bb2be5-f543-490f-89c5-3e7df42ca732 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: 38b08866-9b44-4fc7-957b-d329715af18e] Reconfiguring VM instance instance-00000064 to attach disk [datastore1] 38b08866-9b44-4fc7-957b-d329715af18e/38b08866-9b44-4fc7-957b-d329715af18e.vmdk or device None with type sparse {{(pid=61923) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1074.503923] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-bcd7e9a4-ef48-4b6f-ad4a-b0cc1408386a {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.522834] env[61923]: DEBUG oslo_vmware.api [None req-26bb2be5-f543-490f-89c5-3e7df42ca732 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Waiting for the task: (returnval){ [ 1074.522834] env[61923]: value = "task-1378004" [ 1074.522834] env[61923]: _type = "Task" [ 1074.522834] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1074.535397] env[61923]: DEBUG oslo_vmware.api [None req-26bb2be5-f543-490f-89c5-3e7df42ca732 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Task: {'id': task-1378004, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1074.554027] env[61923]: DEBUG oslo_concurrency.lockutils [None req-e5fe1a98-bb9f-4924-80b3-7de0d4805067 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1074.554027] env[61923]: DEBUG oslo_concurrency.lockutils [None req-e5fe1a98-bb9f-4924-80b3-7de0d4805067 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1074.646042] env[61923]: DEBUG nova.network.neutron [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] [instance: 35bdbd45-5838-4b09-a232-7f4a02322ad7] Updating instance_info_cache with network_info: [{"id": "42e3ee29-941c-43d0-a7f3-81fc7755f64b", "address": "fa:16:3e:88:34:3b", "network": {"id": "d2905082-7ca4-4f92-a142-bb85211bb7c4", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-885218622-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.159", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a13d827933fa4597984afdb91dbbdd39", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "03ac2c9c-6ad2-4a85-bfab-c7e336df859a", "external-id": "nsx-vlan-transportzone-379", "segmentation_id": 379, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap42e3ee29-94", "ovs_interfaceid": "42e3ee29-941c-43d0-a7f3-81fc7755f64b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1075.033310] env[61923]: DEBUG oslo_vmware.api [None req-26bb2be5-f543-490f-89c5-3e7df42ca732 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Task: {'id': task-1378004, 'name': ReconfigVM_Task, 'duration_secs': 0.364176} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1075.033633] env[61923]: DEBUG nova.virt.vmwareapi.volumeops [None req-26bb2be5-f543-490f-89c5-3e7df42ca732 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: 38b08866-9b44-4fc7-957b-d329715af18e] Reconfigured VM instance instance-00000064 to attach disk [datastore1] 38b08866-9b44-4fc7-957b-d329715af18e/38b08866-9b44-4fc7-957b-d329715af18e.vmdk or device None with type sparse {{(pid=61923) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1075.034078] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-d1400fa6-4fd5-41c6-aae1-dfaebdda945e {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.039915] env[61923]: DEBUG oslo_vmware.api [None req-26bb2be5-f543-490f-89c5-3e7df42ca732 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Waiting for the task: (returnval){ [ 1075.039915] env[61923]: value = "task-1378005" [ 1075.039915] env[61923]: _type = "Task" [ 1075.039915] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1075.047388] env[61923]: DEBUG oslo_vmware.api [None req-26bb2be5-f543-490f-89c5-3e7df42ca732 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Task: {'id': task-1378005, 'name': Rename_Task} progress is 5%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1075.058421] env[61923]: INFO nova.compute.claims [None req-e5fe1a98-bb9f-4924-80b3-7de0d4805067 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: ab3edc53-37a5-4119-84c5-fd408ed66b69] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1075.148762] env[61923]: DEBUG oslo_concurrency.lockutils [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Releasing lock "refresh_cache-35bdbd45-5838-4b09-a232-7f4a02322ad7" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1075.149052] env[61923]: DEBUG nova.compute.manager [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] [instance: 35bdbd45-5838-4b09-a232-7f4a02322ad7] Updated the network info_cache for instance {{(pid=61923) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9999}} [ 1075.149512] env[61923]: DEBUG oslo_service.periodic_task [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61923) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1075.149682] env[61923]: DEBUG oslo_service.periodic_task [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61923) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1075.149827] env[61923]: DEBUG oslo_service.periodic_task [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61923) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1075.149976] env[61923]: DEBUG oslo_service.periodic_task [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61923) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1075.150134] env[61923]: DEBUG oslo_service.periodic_task [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61923) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1075.150279] env[61923]: DEBUG oslo_service.periodic_task [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Running periodic task ComputeManager.update_available_resource {{(pid=61923) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1075.507013] env[61923]: DEBUG oslo_concurrency.lockutils [None req-0ae5b13f-f9fc-4836-a56d-6f93c3843994 tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Acquiring lock "35bdbd45-5838-4b09-a232-7f4a02322ad7" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1075.507292] env[61923]: DEBUG oslo_concurrency.lockutils [None req-0ae5b13f-f9fc-4836-a56d-6f93c3843994 tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Lock "35bdbd45-5838-4b09-a232-7f4a02322ad7" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.001s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1075.549884] env[61923]: DEBUG oslo_vmware.api [None req-26bb2be5-f543-490f-89c5-3e7df42ca732 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Task: {'id': task-1378005, 'name': Rename_Task, 'duration_secs': 0.163884} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1075.550165] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-26bb2be5-f543-490f-89c5-3e7df42ca732 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: 38b08866-9b44-4fc7-957b-d329715af18e] Powering on the VM {{(pid=61923) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1075.550398] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6b7f4c26-9e8f-4e2f-be8f-5ab4b51f4d6b {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.555863] env[61923]: DEBUG oslo_vmware.api [None req-26bb2be5-f543-490f-89c5-3e7df42ca732 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Waiting for the task: (returnval){ [ 1075.555863] env[61923]: value = "task-1378006" [ 1075.555863] env[61923]: _type = "Task" [ 1075.555863] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1075.563849] env[61923]: INFO nova.compute.resource_tracker [None req-e5fe1a98-bb9f-4924-80b3-7de0d4805067 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: ab3edc53-37a5-4119-84c5-fd408ed66b69] Updating resource usage from migration 2c6aa9f7-6c7b-47e6-8e98-c1574b78f4a1 [ 1075.566380] env[61923]: DEBUG oslo_vmware.api [None req-26bb2be5-f543-490f-89c5-3e7df42ca732 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Task: {'id': task-1378006, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1075.632255] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a700c69-c324-4944-bb98-27c19931cd57 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.639469] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9101f07-053d-4d12-9732-198a49b5ff55 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.330798] env[61923]: DEBUG oslo_concurrency.lockutils [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1076.331953] env[61923]: DEBUG nova.compute.utils [None req-0ae5b13f-f9fc-4836-a56d-6f93c3843994 tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Using /dev/sd instead of None {{(pid=61923) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1076.336978] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2fdd397d-ed0b-4443-907f-141e3829e730 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.346748] env[61923]: DEBUG oslo_vmware.api [None req-26bb2be5-f543-490f-89c5-3e7df42ca732 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Task: {'id': task-1378006, 'name': PowerOnVM_Task, 'duration_secs': 0.442947} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1076.347852] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3577221-156a-4d09-a373-73f2e19328c3 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.351304] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-26bb2be5-f543-490f-89c5-3e7df42ca732 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: 38b08866-9b44-4fc7-957b-d329715af18e] Powered on the VM {{(pid=61923) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1076.351503] env[61923]: INFO nova.compute.manager [None req-26bb2be5-f543-490f-89c5-3e7df42ca732 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: 38b08866-9b44-4fc7-957b-d329715af18e] Took 7.73 seconds to spawn the instance on the hypervisor. [ 1076.351678] env[61923]: DEBUG nova.compute.manager [None req-26bb2be5-f543-490f-89c5-3e7df42ca732 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: 38b08866-9b44-4fc7-957b-d329715af18e] Checking state {{(pid=61923) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1076.352565] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53709fe9-5d17-45d3-b164-fae168c1b6a0 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.363992] env[61923]: DEBUG nova.compute.provider_tree [None req-e5fe1a98-bb9f-4924-80b3-7de0d4805067 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1076.834563] env[61923]: DEBUG oslo_concurrency.lockutils [None req-0ae5b13f-f9fc-4836-a56d-6f93c3843994 tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Lock "35bdbd45-5838-4b09-a232-7f4a02322ad7" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.327s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1076.870424] env[61923]: DEBUG nova.scheduler.client.report [None req-e5fe1a98-bb9f-4924-80b3-7de0d4805067 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1076.878885] env[61923]: INFO nova.compute.manager [None req-26bb2be5-f543-490f-89c5-3e7df42ca732 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: 38b08866-9b44-4fc7-957b-d329715af18e] Took 13.36 seconds to build instance. [ 1077.381239] env[61923]: DEBUG oslo_concurrency.lockutils [None req-e5fe1a98-bb9f-4924-80b3-7de0d4805067 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.827s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1077.381612] env[61923]: INFO nova.compute.manager [None req-e5fe1a98-bb9f-4924-80b3-7de0d4805067 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: ab3edc53-37a5-4119-84c5-fd408ed66b69] Migrating [ 1077.381686] env[61923]: DEBUG oslo_concurrency.lockutils [None req-e5fe1a98-bb9f-4924-80b3-7de0d4805067 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Acquiring lock "compute-rpcapi-router" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1077.381825] env[61923]: DEBUG oslo_concurrency.lockutils [None req-e5fe1a98-bb9f-4924-80b3-7de0d4805067 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Acquired lock "compute-rpcapi-router" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1077.383122] env[61923]: DEBUG oslo_concurrency.lockutils [None req-26bb2be5-f543-490f-89c5-3e7df42ca732 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Lock "38b08866-9b44-4fc7-957b-d329715af18e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.875s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1077.383337] env[61923]: DEBUG oslo_concurrency.lockutils [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 1.053s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1077.383496] env[61923]: DEBUG oslo_concurrency.lockutils [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1077.383636] env[61923]: DEBUG nova.compute.resource_tracker [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61923) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1077.384657] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac29d991-00ea-4a2b-a62a-136cb8fbdcf6 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.393271] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-baa4019b-b786-4cd3-985f-490e71bc94ac {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.408022] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-abafc976-e3af-4dc5-8021-707a38a8da89 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.413904] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eaa14b0f-f6c5-4c6c-af06-99a59a0a67fd {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.441930] env[61923]: DEBUG nova.compute.resource_tracker [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181301MB free_disk=178GB free_vcpus=48 pci_devices=None {{(pid=61923) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1077.442111] env[61923]: DEBUG oslo_concurrency.lockutils [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1077.442260] env[61923]: DEBUG oslo_concurrency.lockutils [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1077.573222] env[61923]: DEBUG oslo_concurrency.lockutils [None req-0ae5b13f-f9fc-4836-a56d-6f93c3843994 tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Acquiring lock "35bdbd45-5838-4b09-a232-7f4a02322ad7" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1077.573494] env[61923]: DEBUG oslo_concurrency.lockutils [None req-0ae5b13f-f9fc-4836-a56d-6f93c3843994 tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Lock "35bdbd45-5838-4b09-a232-7f4a02322ad7" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1077.573725] env[61923]: INFO nova.compute.manager [None req-0ae5b13f-f9fc-4836-a56d-6f93c3843994 tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] [instance: 35bdbd45-5838-4b09-a232-7f4a02322ad7] Attaching volume f908034d-f063-4b8d-9877-08531afe6e62 to /dev/sdb [ 1077.604360] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e5be505-b893-4d23-a42b-f6ee762d1af2 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.611312] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1921528-5d13-4632-bd1b-df041a7df73f {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.623618] env[61923]: DEBUG nova.virt.block_device [None req-0ae5b13f-f9fc-4836-a56d-6f93c3843994 tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] [instance: 35bdbd45-5838-4b09-a232-7f4a02322ad7] Updating existing volume attachment record: 32bbfe3d-62d2-4344-8df6-b4d3027ae930 {{(pid=61923) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1077.775640] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d838b19-0429-47b7-a35d-deabc21e0717 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.781808] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-d345334c-5da0-4395-85ef-9e40fc2b7ede tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: 38b08866-9b44-4fc7-957b-d329715af18e] Suspending the VM {{(pid=61923) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1162}} [ 1077.782080] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.SuspendVM_Task with opID=oslo.vmware-e3fb0286-3b8f-4160-8cce-cbd1ccb351a6 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.786922] env[61923]: DEBUG oslo_vmware.api [None req-d345334c-5da0-4395-85ef-9e40fc2b7ede tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Waiting for the task: (returnval){ [ 1077.786922] env[61923]: value = "task-1378008" [ 1077.786922] env[61923]: _type = "Task" [ 1077.786922] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1077.795386] env[61923]: DEBUG oslo_vmware.api [None req-d345334c-5da0-4395-85ef-9e40fc2b7ede tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Task: {'id': task-1378008, 'name': SuspendVM_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1077.885550] env[61923]: INFO nova.compute.rpcapi [None req-e5fe1a98-bb9f-4924-80b3-7de0d4805067 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Automatically selected compute RPC version 6.3 from minimum service version 67 [ 1077.886252] env[61923]: DEBUG oslo_concurrency.lockutils [None req-e5fe1a98-bb9f-4924-80b3-7de0d4805067 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Releasing lock "compute-rpcapi-router" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1078.299074] env[61923]: DEBUG oslo_vmware.api [None req-d345334c-5da0-4395-85ef-9e40fc2b7ede tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Task: {'id': task-1378008, 'name': SuspendVM_Task} progress is 66%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1078.404738] env[61923]: DEBUG oslo_concurrency.lockutils [None req-e5fe1a98-bb9f-4924-80b3-7de0d4805067 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Acquiring lock "refresh_cache-ab3edc53-37a5-4119-84c5-fd408ed66b69" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1078.405134] env[61923]: DEBUG oslo_concurrency.lockutils [None req-e5fe1a98-bb9f-4924-80b3-7de0d4805067 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Acquired lock "refresh_cache-ab3edc53-37a5-4119-84c5-fd408ed66b69" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1078.405380] env[61923]: DEBUG nova.network.neutron [None req-e5fe1a98-bb9f-4924-80b3-7de0d4805067 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: ab3edc53-37a5-4119-84c5-fd408ed66b69] Building network info cache for instance {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1078.448996] env[61923]: DEBUG nova.compute.resource_tracker [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Applying migration context for instance ab3edc53-37a5-4119-84c5-fd408ed66b69 as it has an incoming, in-progress migration 2c6aa9f7-6c7b-47e6-8e98-c1574b78f4a1. Migration status is pre-migrating {{(pid=61923) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1016}} [ 1078.449923] env[61923]: INFO nova.compute.resource_tracker [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] [instance: ab3edc53-37a5-4119-84c5-fd408ed66b69] Updating resource usage from migration 2c6aa9f7-6c7b-47e6-8e98-c1574b78f4a1 [ 1078.547519] env[61923]: DEBUG nova.compute.resource_tracker [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Instance 35bdbd45-5838-4b09-a232-7f4a02322ad7 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61923) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1078.547673] env[61923]: DEBUG nova.compute.resource_tracker [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Instance 38b08866-9b44-4fc7-957b-d329715af18e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61923) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1078.547806] env[61923]: DEBUG nova.compute.resource_tracker [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Migration 2c6aa9f7-6c7b-47e6-8e98-c1574b78f4a1 is active on this compute host and has allocations in placement: {'resources': {'VCPU': 1, 'MEMORY_MB': 192, 'DISK_GB': 1}}. {{(pid=61923) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1712}} [ 1078.548038] env[61923]: DEBUG nova.compute.resource_tracker [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Instance ab3edc53-37a5-4119-84c5-fd408ed66b69 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=61923) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1078.548351] env[61923]: DEBUG nova.compute.resource_tracker [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Total usable vcpus: 48, total allocated vcpus: 4 {{(pid=61923) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1078.548500] env[61923]: DEBUG nova.compute.resource_tracker [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1344MB phys_disk=200GB used_disk=4GB total_vcpus=48 used_vcpus=4 pci_stats=[] {{(pid=61923) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1078.611634] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9173f47-270c-41c1-a03c-76cfba5a5bab {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.619694] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70e74b27-c295-4931-bd11-9eb150ec0096 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.652168] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-260294ec-c4b2-4c79-9599-0c9fdc643495 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.659943] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0cb7da5-2707-4a92-91b3-29f86c6df29a {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.674589] env[61923]: DEBUG nova.compute.provider_tree [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1078.797801] env[61923]: DEBUG oslo_vmware.api [None req-d345334c-5da0-4395-85ef-9e40fc2b7ede tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Task: {'id': task-1378008, 'name': SuspendVM_Task, 'duration_secs': 0.798879} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1078.798240] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-d345334c-5da0-4395-85ef-9e40fc2b7ede tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: 38b08866-9b44-4fc7-957b-d329715af18e] Suspended the VM {{(pid=61923) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1166}} [ 1078.798538] env[61923]: DEBUG nova.compute.manager [None req-d345334c-5da0-4395-85ef-9e40fc2b7ede tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: 38b08866-9b44-4fc7-957b-d329715af18e] Checking state {{(pid=61923) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1078.799422] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5ae3bfd-e591-4791-a294-acbac239425e {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.103989] env[61923]: DEBUG nova.network.neutron [None req-e5fe1a98-bb9f-4924-80b3-7de0d4805067 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: ab3edc53-37a5-4119-84c5-fd408ed66b69] Updating instance_info_cache with network_info: [{"id": "f2c777e2-f8f4-437b-b990-20e42dd2491b", "address": "fa:16:3e:b3:a4:a2", "network": {"id": "77aec6ca-d1be-4b39-9d51-08853a1183a1", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1813085414-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5a4db4f8aa194a219e80c538cc715d43", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "56b944d8-803d-43f2-945d-0f334ee4ea1c", "external-id": "nsx-vlan-transportzone-799", "segmentation_id": 799, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf2c777e2-f8", "ovs_interfaceid": "f2c777e2-f8f4-437b-b990-20e42dd2491b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1079.177745] env[61923]: DEBUG nova.scheduler.client.report [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1079.607016] env[61923]: DEBUG oslo_concurrency.lockutils [None req-e5fe1a98-bb9f-4924-80b3-7de0d4805067 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Releasing lock "refresh_cache-ab3edc53-37a5-4119-84c5-fd408ed66b69" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1079.682425] env[61923]: DEBUG nova.compute.resource_tracker [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61923) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1079.682631] env[61923]: DEBUG oslo_concurrency.lockutils [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.240s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1079.682927] env[61923]: DEBUG oslo_service.periodic_task [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Running periodic task ComputeManager._run_pending_deletes {{(pid=61923) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1079.683083] env[61923]: DEBUG nova.compute.manager [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Cleaning up deleted instances {{(pid=61923) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11228}} [ 1080.194785] env[61923]: DEBUG nova.compute.manager [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] There are 36 instances to clean {{(pid=61923) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11237}} [ 1080.195088] env[61923]: DEBUG nova.compute.manager [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] [instance: eeb51744-e397-40f8-8e1a-813fa23ebe95] Instance has had 0 of 5 cleanup attempts {{(pid=61923) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1080.217522] env[61923]: DEBUG oslo_concurrency.lockutils [None req-b76e4b68-94ff-4a62-be9c-54322062fc01 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Acquiring lock "38b08866-9b44-4fc7-957b-d329715af18e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1080.217814] env[61923]: DEBUG oslo_concurrency.lockutils [None req-b76e4b68-94ff-4a62-be9c-54322062fc01 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Lock "38b08866-9b44-4fc7-957b-d329715af18e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1080.218045] env[61923]: DEBUG oslo_concurrency.lockutils [None req-b76e4b68-94ff-4a62-be9c-54322062fc01 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Acquiring lock "38b08866-9b44-4fc7-957b-d329715af18e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1080.218240] env[61923]: DEBUG oslo_concurrency.lockutils [None req-b76e4b68-94ff-4a62-be9c-54322062fc01 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Lock "38b08866-9b44-4fc7-957b-d329715af18e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1080.218417] env[61923]: DEBUG oslo_concurrency.lockutils [None req-b76e4b68-94ff-4a62-be9c-54322062fc01 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Lock "38b08866-9b44-4fc7-957b-d329715af18e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1080.220771] env[61923]: INFO nova.compute.manager [None req-b76e4b68-94ff-4a62-be9c-54322062fc01 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: 38b08866-9b44-4fc7-957b-d329715af18e] Terminating instance [ 1080.222579] env[61923]: DEBUG nova.compute.manager [None req-b76e4b68-94ff-4a62-be9c-54322062fc01 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: 38b08866-9b44-4fc7-957b-d329715af18e] Start destroying the instance on the hypervisor. {{(pid=61923) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1080.222774] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-b76e4b68-94ff-4a62-be9c-54322062fc01 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: 38b08866-9b44-4fc7-957b-d329715af18e] Destroying instance {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1080.223631] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9793bec-05b3-4ffc-9c63-51a0daa8de76 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.231973] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-b76e4b68-94ff-4a62-be9c-54322062fc01 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: 38b08866-9b44-4fc7-957b-d329715af18e] Unregistering the VM {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1080.232253] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9c713cff-8ef7-4d4f-8e4e-0a6ed125b9b2 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.312456] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-b76e4b68-94ff-4a62-be9c-54322062fc01 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: 38b08866-9b44-4fc7-957b-d329715af18e] Unregistered the VM {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1080.312720] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-b76e4b68-94ff-4a62-be9c-54322062fc01 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: 38b08866-9b44-4fc7-957b-d329715af18e] Deleting contents of the VM from datastore datastore1 {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1080.312860] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-b76e4b68-94ff-4a62-be9c-54322062fc01 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Deleting the datastore file [datastore1] 38b08866-9b44-4fc7-957b-d329715af18e {{(pid=61923) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1080.313137] env[61923]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e83ec79a-de7a-4dca-94c5-d47a4f2702cf {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.319254] env[61923]: DEBUG oslo_vmware.api [None req-b76e4b68-94ff-4a62-be9c-54322062fc01 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Waiting for the task: (returnval){ [ 1080.319254] env[61923]: value = "task-1378011" [ 1080.319254] env[61923]: _type = "Task" [ 1080.319254] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1080.326957] env[61923]: DEBUG oslo_vmware.api [None req-b76e4b68-94ff-4a62-be9c-54322062fc01 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Task: {'id': task-1378011, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1080.698797] env[61923]: DEBUG nova.compute.manager [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] [instance: 9aeac922-aac2-47fe-8c11-835bef75e55c] Instance has had 0 of 5 cleanup attempts {{(pid=61923) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1080.832627] env[61923]: DEBUG oslo_vmware.api [None req-b76e4b68-94ff-4a62-be9c-54322062fc01 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Task: {'id': task-1378011, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.13258} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1080.833039] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-b76e4b68-94ff-4a62-be9c-54322062fc01 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Deleted the datastore file {{(pid=61923) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1080.833354] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-b76e4b68-94ff-4a62-be9c-54322062fc01 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: 38b08866-9b44-4fc7-957b-d329715af18e] Deleted contents of the VM from datastore datastore1 {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1080.833654] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-b76e4b68-94ff-4a62-be9c-54322062fc01 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: 38b08866-9b44-4fc7-957b-d329715af18e] Instance destroyed {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1080.833954] env[61923]: INFO nova.compute.manager [None req-b76e4b68-94ff-4a62-be9c-54322062fc01 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: 38b08866-9b44-4fc7-957b-d329715af18e] Took 0.61 seconds to destroy the instance on the hypervisor. [ 1080.834362] env[61923]: DEBUG oslo.service.loopingcall [None req-b76e4b68-94ff-4a62-be9c-54322062fc01 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61923) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1080.834649] env[61923]: DEBUG nova.compute.manager [-] [instance: 38b08866-9b44-4fc7-957b-d329715af18e] Deallocating network for instance {{(pid=61923) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1080.834788] env[61923]: DEBUG nova.network.neutron [-] [instance: 38b08866-9b44-4fc7-957b-d329715af18e] deallocate_for_instance() {{(pid=61923) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1081.121570] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e064592d-276d-44e2-93de-828d7e939f4e {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.142349] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-e5fe1a98-bb9f-4924-80b3-7de0d4805067 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: ab3edc53-37a5-4119-84c5-fd408ed66b69] Updating instance 'ab3edc53-37a5-4119-84c5-fd408ed66b69' progress to 0 {{(pid=61923) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1081.187453] env[61923]: DEBUG nova.compute.manager [req-b606256b-a863-458a-bba0-b229118bdec8 req-bbfdb9f8-4565-4e01-afd7-a7fe466990b6 service nova] [instance: 38b08866-9b44-4fc7-957b-d329715af18e] Received event network-vif-deleted-d5c39dc6-35e2-4535-941d-f5673169edca {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1081.187666] env[61923]: INFO nova.compute.manager [req-b606256b-a863-458a-bba0-b229118bdec8 req-bbfdb9f8-4565-4e01-afd7-a7fe466990b6 service nova] [instance: 38b08866-9b44-4fc7-957b-d329715af18e] Neutron deleted interface d5c39dc6-35e2-4535-941d-f5673169edca; detaching it from the instance and deleting it from the info cache [ 1081.187845] env[61923]: DEBUG nova.network.neutron [req-b606256b-a863-458a-bba0-b229118bdec8 req-bbfdb9f8-4565-4e01-afd7-a7fe466990b6 service nova] [instance: 38b08866-9b44-4fc7-957b-d329715af18e] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1081.202518] env[61923]: DEBUG nova.compute.manager [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] [instance: 0c3788fc-6cce-4806-b4f8-84f84489fb93] Instance has had 0 of 5 cleanup attempts {{(pid=61923) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1081.651210] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-e5fe1a98-bb9f-4924-80b3-7de0d4805067 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: ab3edc53-37a5-4119-84c5-fd408ed66b69] Powering off the VM {{(pid=61923) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1081.651540] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7c12ab3e-1470-45b5-ae6b-71264f1a6e03 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.658870] env[61923]: DEBUG oslo_vmware.api [None req-e5fe1a98-bb9f-4924-80b3-7de0d4805067 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Waiting for the task: (returnval){ [ 1081.658870] env[61923]: value = "task-1378012" [ 1081.658870] env[61923]: _type = "Task" [ 1081.658870] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1081.663579] env[61923]: DEBUG nova.network.neutron [-] [instance: 38b08866-9b44-4fc7-957b-d329715af18e] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1081.667843] env[61923]: DEBUG oslo_vmware.api [None req-e5fe1a98-bb9f-4924-80b3-7de0d4805067 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Task: {'id': task-1378012, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1081.690698] env[61923]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-68affc6f-f705-4a93-b947-a14d810fdc0c {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.700949] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a966148-0f19-4de7-966f-7a78d2ce3e49 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.711172] env[61923]: DEBUG nova.compute.manager [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] [instance: 3fcedef5-9040-4b45-9e27-f0700112d528] Instance has had 0 of 5 cleanup attempts {{(pid=61923) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1081.726872] env[61923]: DEBUG nova.compute.manager [req-b606256b-a863-458a-bba0-b229118bdec8 req-bbfdb9f8-4565-4e01-afd7-a7fe466990b6 service nova] [instance: 38b08866-9b44-4fc7-957b-d329715af18e] Detach interface failed, port_id=d5c39dc6-35e2-4535-941d-f5673169edca, reason: Instance 38b08866-9b44-4fc7-957b-d329715af18e could not be found. {{(pid=61923) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1082.165103] env[61923]: DEBUG nova.virt.vmwareapi.volumeops [None req-0ae5b13f-f9fc-4836-a56d-6f93c3843994 tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] [instance: 35bdbd45-5838-4b09-a232-7f4a02322ad7] Volume attach. Driver type: vmdk {{(pid=61923) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1082.165376] env[61923]: DEBUG nova.virt.vmwareapi.volumeops [None req-0ae5b13f-f9fc-4836-a56d-6f93c3843994 tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] [instance: 35bdbd45-5838-4b09-a232-7f4a02322ad7] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-292753', 'volume_id': 'f908034d-f063-4b8d-9877-08531afe6e62', 'name': 'volume-f908034d-f063-4b8d-9877-08531afe6e62', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '35bdbd45-5838-4b09-a232-7f4a02322ad7', 'attached_at': '', 'detached_at': '', 'volume_id': 'f908034d-f063-4b8d-9877-08531afe6e62', 'serial': 'f908034d-f063-4b8d-9877-08531afe6e62'} {{(pid=61923) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1082.166114] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d824e3f6-4b05-4b28-a443-1124f6d9978f {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.171496] env[61923]: INFO nova.compute.manager [-] [instance: 38b08866-9b44-4fc7-957b-d329715af18e] Took 1.34 seconds to deallocate network for instance. [ 1082.171907] env[61923]: DEBUG oslo_vmware.api [None req-e5fe1a98-bb9f-4924-80b3-7de0d4805067 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Task: {'id': task-1378012, 'name': PowerOffVM_Task, 'duration_secs': 0.201873} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1082.173614] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-e5fe1a98-bb9f-4924-80b3-7de0d4805067 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: ab3edc53-37a5-4119-84c5-fd408ed66b69] Powered off the VM {{(pid=61923) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1082.173770] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-e5fe1a98-bb9f-4924-80b3-7de0d4805067 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: ab3edc53-37a5-4119-84c5-fd408ed66b69] Updating instance 'ab3edc53-37a5-4119-84c5-fd408ed66b69' progress to 17 {{(pid=61923) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1082.188011] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4b70a9e-d214-4848-aebd-16e253a97ce0 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.217800] env[61923]: DEBUG nova.virt.vmwareapi.volumeops [None req-0ae5b13f-f9fc-4836-a56d-6f93c3843994 tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] [instance: 35bdbd45-5838-4b09-a232-7f4a02322ad7] Reconfiguring VM instance instance-00000060 to attach disk [datastore2] volume-f908034d-f063-4b8d-9877-08531afe6e62/volume-f908034d-f063-4b8d-9877-08531afe6e62.vmdk or device None with type thin {{(pid=61923) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1082.219215] env[61923]: DEBUG nova.compute.manager [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] [instance: 638bbbbb-827e-4a73-ac24-a3e1b2cd72eb] Instance has had 0 of 5 cleanup attempts {{(pid=61923) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1082.220770] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a0a1831f-8db3-4248-937b-943440215b26 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.238731] env[61923]: DEBUG oslo_vmware.api [None req-0ae5b13f-f9fc-4836-a56d-6f93c3843994 tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Waiting for the task: (returnval){ [ 1082.238731] env[61923]: value = "task-1378013" [ 1082.238731] env[61923]: _type = "Task" [ 1082.238731] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1082.247267] env[61923]: DEBUG oslo_vmware.api [None req-0ae5b13f-f9fc-4836-a56d-6f93c3843994 tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Task: {'id': task-1378013, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1082.689651] env[61923]: DEBUG nova.virt.hardware [None req-e5fe1a98-bb9f-4924-80b3-7de0d4805067 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-29T20:07:54Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=0f153f63-ae0a-45b1-b7f5-7f9b673c947f,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1082.689942] env[61923]: DEBUG nova.virt.hardware [None req-e5fe1a98-bb9f-4924-80b3-7de0d4805067 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Flavor limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1082.690087] env[61923]: DEBUG nova.virt.hardware [None req-e5fe1a98-bb9f-4924-80b3-7de0d4805067 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Image limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1082.690282] env[61923]: DEBUG nova.virt.hardware [None req-e5fe1a98-bb9f-4924-80b3-7de0d4805067 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Flavor pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1082.690431] env[61923]: DEBUG nova.virt.hardware [None req-e5fe1a98-bb9f-4924-80b3-7de0d4805067 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Image pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1082.690579] env[61923]: DEBUG nova.virt.hardware [None req-e5fe1a98-bb9f-4924-80b3-7de0d4805067 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1082.690784] env[61923]: DEBUG nova.virt.hardware [None req-e5fe1a98-bb9f-4924-80b3-7de0d4805067 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1082.690982] env[61923]: DEBUG nova.virt.hardware [None req-e5fe1a98-bb9f-4924-80b3-7de0d4805067 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1082.691176] env[61923]: DEBUG nova.virt.hardware [None req-e5fe1a98-bb9f-4924-80b3-7de0d4805067 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Got 1 possible topologies {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1082.691349] env[61923]: DEBUG nova.virt.hardware [None req-e5fe1a98-bb9f-4924-80b3-7de0d4805067 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1082.691527] env[61923]: DEBUG nova.virt.hardware [None req-e5fe1a98-bb9f-4924-80b3-7de0d4805067 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1082.697456] env[61923]: DEBUG oslo_concurrency.lockutils [None req-b76e4b68-94ff-4a62-be9c-54322062fc01 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1082.697696] env[61923]: DEBUG oslo_concurrency.lockutils [None req-b76e4b68-94ff-4a62-be9c-54322062fc01 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1082.697901] env[61923]: DEBUG nova.objects.instance [None req-b76e4b68-94ff-4a62-be9c-54322062fc01 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Lazy-loading 'resources' on Instance uuid 38b08866-9b44-4fc7-957b-d329715af18e {{(pid=61923) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1082.698882] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7599d429-2dd2-4aa0-b8a7-3cc10f44595c {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.717772] env[61923]: DEBUG oslo_vmware.api [None req-e5fe1a98-bb9f-4924-80b3-7de0d4805067 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Waiting for the task: (returnval){ [ 1082.717772] env[61923]: value = "task-1378014" [ 1082.717772] env[61923]: _type = "Task" [ 1082.717772] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1082.726019] env[61923]: DEBUG oslo_vmware.api [None req-e5fe1a98-bb9f-4924-80b3-7de0d4805067 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Task: {'id': task-1378014, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1082.734100] env[61923]: DEBUG nova.compute.manager [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] [instance: aba844c1-d488-4e95-a550-bd1cb7331fb2] Instance has had 0 of 5 cleanup attempts {{(pid=61923) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1082.751543] env[61923]: DEBUG oslo_vmware.api [None req-0ae5b13f-f9fc-4836-a56d-6f93c3843994 tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Task: {'id': task-1378013, 'name': ReconfigVM_Task, 'duration_secs': 0.314319} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1082.751850] env[61923]: DEBUG nova.virt.vmwareapi.volumeops [None req-0ae5b13f-f9fc-4836-a56d-6f93c3843994 tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] [instance: 35bdbd45-5838-4b09-a232-7f4a02322ad7] Reconfigured VM instance instance-00000060 to attach disk [datastore2] volume-f908034d-f063-4b8d-9877-08531afe6e62/volume-f908034d-f063-4b8d-9877-08531afe6e62.vmdk or device None with type thin {{(pid=61923) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1082.759657] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a0f26369-40b4-43e7-b3ab-be3ef277b76e {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.774628] env[61923]: DEBUG oslo_vmware.api [None req-0ae5b13f-f9fc-4836-a56d-6f93c3843994 tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Waiting for the task: (returnval){ [ 1082.774628] env[61923]: value = "task-1378015" [ 1082.774628] env[61923]: _type = "Task" [ 1082.774628] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1082.783749] env[61923]: DEBUG oslo_vmware.api [None req-0ae5b13f-f9fc-4836-a56d-6f93c3843994 tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Task: {'id': task-1378015, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1082.790120] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-771aa48e-34eb-49b1-bd8b-d503dd9f3c60 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.796760] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bdf297e7-0fc6-4bb6-9d6a-b0ccfef94821 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.826793] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f023c050-a74b-4ab8-994e-388b234ad133 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.834046] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b706e9a-8c46-4e61-b62d-59019ff12b8d {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.850127] env[61923]: DEBUG nova.compute.provider_tree [None req-b76e4b68-94ff-4a62-be9c-54322062fc01 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1083.227991] env[61923]: DEBUG oslo_vmware.api [None req-e5fe1a98-bb9f-4924-80b3-7de0d4805067 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Task: {'id': task-1378014, 'name': ReconfigVM_Task, 'duration_secs': 0.151839} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1083.228367] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-e5fe1a98-bb9f-4924-80b3-7de0d4805067 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: ab3edc53-37a5-4119-84c5-fd408ed66b69] Updating instance 'ab3edc53-37a5-4119-84c5-fd408ed66b69' progress to 33 {{(pid=61923) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1083.237373] env[61923]: DEBUG nova.compute.manager [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] [instance: 44196b4c-1401-40fa-bd14-04a49947ab15] Instance has had 0 of 5 cleanup attempts {{(pid=61923) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1083.284489] env[61923]: DEBUG oslo_vmware.api [None req-0ae5b13f-f9fc-4836-a56d-6f93c3843994 tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Task: {'id': task-1378015, 'name': ReconfigVM_Task, 'duration_secs': 0.145618} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1083.284731] env[61923]: DEBUG nova.virt.vmwareapi.volumeops [None req-0ae5b13f-f9fc-4836-a56d-6f93c3843994 tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] [instance: 35bdbd45-5838-4b09-a232-7f4a02322ad7] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-292753', 'volume_id': 'f908034d-f063-4b8d-9877-08531afe6e62', 'name': 'volume-f908034d-f063-4b8d-9877-08531afe6e62', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '35bdbd45-5838-4b09-a232-7f4a02322ad7', 'attached_at': '', 'detached_at': '', 'volume_id': 'f908034d-f063-4b8d-9877-08531afe6e62', 'serial': 'f908034d-f063-4b8d-9877-08531afe6e62'} {{(pid=61923) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1083.354099] env[61923]: DEBUG nova.scheduler.client.report [None req-b76e4b68-94ff-4a62-be9c-54322062fc01 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1083.734997] env[61923]: DEBUG nova.virt.hardware [None req-e5fe1a98-bb9f-4924-80b3-7de0d4805067 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-29T20:07:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=0f153f63-ae0a-45b1-b7f5-7f9b673c947f,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1083.735445] env[61923]: DEBUG nova.virt.hardware [None req-e5fe1a98-bb9f-4924-80b3-7de0d4805067 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Flavor limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1083.735445] env[61923]: DEBUG nova.virt.hardware [None req-e5fe1a98-bb9f-4924-80b3-7de0d4805067 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Image limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1083.735725] env[61923]: DEBUG nova.virt.hardware [None req-e5fe1a98-bb9f-4924-80b3-7de0d4805067 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Flavor pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1083.735882] env[61923]: DEBUG nova.virt.hardware [None req-e5fe1a98-bb9f-4924-80b3-7de0d4805067 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Image pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1083.736049] env[61923]: DEBUG nova.virt.hardware [None req-e5fe1a98-bb9f-4924-80b3-7de0d4805067 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1083.736261] env[61923]: DEBUG nova.virt.hardware [None req-e5fe1a98-bb9f-4924-80b3-7de0d4805067 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1083.736432] env[61923]: DEBUG nova.virt.hardware [None req-e5fe1a98-bb9f-4924-80b3-7de0d4805067 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1083.736604] env[61923]: DEBUG nova.virt.hardware [None req-e5fe1a98-bb9f-4924-80b3-7de0d4805067 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Got 1 possible topologies {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1083.736767] env[61923]: DEBUG nova.virt.hardware [None req-e5fe1a98-bb9f-4924-80b3-7de0d4805067 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1083.736941] env[61923]: DEBUG nova.virt.hardware [None req-e5fe1a98-bb9f-4924-80b3-7de0d4805067 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1083.742166] env[61923]: DEBUG nova.virt.vmwareapi.volumeops [None req-e5fe1a98-bb9f-4924-80b3-7de0d4805067 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: ab3edc53-37a5-4119-84c5-fd408ed66b69] Reconfiguring VM instance instance-00000063 to detach disk 2000 {{(pid=61923) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1083.742581] env[61923]: DEBUG nova.compute.manager [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] [instance: 7867c6cf-2ad6-414f-ab52-581827860836] Instance has had 0 of 5 cleanup attempts {{(pid=61923) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1083.744317] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-bea66ead-49a8-46ee-84ff-1b9fdfb9c991 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.762611] env[61923]: DEBUG oslo_vmware.api [None req-e5fe1a98-bb9f-4924-80b3-7de0d4805067 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Waiting for the task: (returnval){ [ 1083.762611] env[61923]: value = "task-1378016" [ 1083.762611] env[61923]: _type = "Task" [ 1083.762611] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1083.771017] env[61923]: DEBUG oslo_vmware.api [None req-e5fe1a98-bb9f-4924-80b3-7de0d4805067 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Task: {'id': task-1378016, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1083.858014] env[61923]: DEBUG oslo_concurrency.lockutils [None req-b76e4b68-94ff-4a62-be9c-54322062fc01 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.160s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1083.877026] env[61923]: INFO nova.scheduler.client.report [None req-b76e4b68-94ff-4a62-be9c-54322062fc01 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Deleted allocations for instance 38b08866-9b44-4fc7-957b-d329715af18e [ 1084.258169] env[61923]: DEBUG nova.compute.manager [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] [instance: 02a8f197-80bb-4cee-bdd6-b07705759986] Instance has had 0 of 5 cleanup attempts {{(pid=61923) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1084.272691] env[61923]: DEBUG oslo_vmware.api [None req-e5fe1a98-bb9f-4924-80b3-7de0d4805067 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Task: {'id': task-1378016, 'name': ReconfigVM_Task, 'duration_secs': 0.156463} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1084.272913] env[61923]: DEBUG nova.virt.vmwareapi.volumeops [None req-e5fe1a98-bb9f-4924-80b3-7de0d4805067 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: ab3edc53-37a5-4119-84c5-fd408ed66b69] Reconfigured VM instance instance-00000063 to detach disk 2000 {{(pid=61923) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1084.273690] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f081f61-99d0-4486-aa16-2a7b338d3e94 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.296402] env[61923]: DEBUG nova.virt.vmwareapi.volumeops [None req-e5fe1a98-bb9f-4924-80b3-7de0d4805067 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: ab3edc53-37a5-4119-84c5-fd408ed66b69] Reconfiguring VM instance instance-00000063 to attach disk [datastore1] ab3edc53-37a5-4119-84c5-fd408ed66b69/ab3edc53-37a5-4119-84c5-fd408ed66b69.vmdk or device None with type thin {{(pid=61923) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1084.297356] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1740983f-d75f-4638-842e-90fec5d8daac {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.315034] env[61923]: DEBUG oslo_vmware.api [None req-e5fe1a98-bb9f-4924-80b3-7de0d4805067 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Waiting for the task: (returnval){ [ 1084.315034] env[61923]: value = "task-1378017" [ 1084.315034] env[61923]: _type = "Task" [ 1084.315034] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1084.323576] env[61923]: DEBUG nova.objects.instance [None req-0ae5b13f-f9fc-4836-a56d-6f93c3843994 tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Lazy-loading 'flavor' on Instance uuid 35bdbd45-5838-4b09-a232-7f4a02322ad7 {{(pid=61923) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1084.324839] env[61923]: DEBUG oslo_vmware.api [None req-e5fe1a98-bb9f-4924-80b3-7de0d4805067 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Task: {'id': task-1378017, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1084.385817] env[61923]: DEBUG oslo_concurrency.lockutils [None req-b76e4b68-94ff-4a62-be9c-54322062fc01 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Lock "38b08866-9b44-4fc7-957b-d329715af18e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 4.168s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1084.761334] env[61923]: DEBUG nova.compute.manager [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] [instance: 62c75dfd-6046-4de1-b1e3-13a307af1394] Instance has had 0 of 5 cleanup attempts {{(pid=61923) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1084.825785] env[61923]: DEBUG oslo_vmware.api [None req-e5fe1a98-bb9f-4924-80b3-7de0d4805067 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Task: {'id': task-1378017, 'name': ReconfigVM_Task, 'duration_secs': 0.258045} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1084.827984] env[61923]: DEBUG nova.virt.vmwareapi.volumeops [None req-e5fe1a98-bb9f-4924-80b3-7de0d4805067 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: ab3edc53-37a5-4119-84c5-fd408ed66b69] Reconfigured VM instance instance-00000063 to attach disk [datastore1] ab3edc53-37a5-4119-84c5-fd408ed66b69/ab3edc53-37a5-4119-84c5-fd408ed66b69.vmdk or device None with type thin {{(pid=61923) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1084.828396] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-e5fe1a98-bb9f-4924-80b3-7de0d4805067 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: ab3edc53-37a5-4119-84c5-fd408ed66b69] Updating instance 'ab3edc53-37a5-4119-84c5-fd408ed66b69' progress to 50 {{(pid=61923) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1084.832111] env[61923]: DEBUG oslo_concurrency.lockutils [None req-0ae5b13f-f9fc-4836-a56d-6f93c3843994 tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Lock "35bdbd45-5838-4b09-a232-7f4a02322ad7" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.259s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1084.966959] env[61923]: DEBUG oslo_concurrency.lockutils [None req-affadfdd-ba30-4ce8-8d78-e18297fead5a tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Acquiring lock "c0a8a116-c7b6-4adc-849b-e0a4faa150d3" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1084.967241] env[61923]: DEBUG oslo_concurrency.lockutils [None req-affadfdd-ba30-4ce8-8d78-e18297fead5a tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Lock "c0a8a116-c7b6-4adc-849b-e0a4faa150d3" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1085.264744] env[61923]: DEBUG nova.compute.manager [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] [instance: 03c1e8ea-a322-46fa-9b07-e283080a7871] Instance has had 0 of 5 cleanup attempts {{(pid=61923) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1085.336763] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3dbd450f-291e-495d-8919-d2abcc1a3025 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.357893] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1cf6bb8-7937-4954-8cc0-de6102aacb4d {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.376217] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-e5fe1a98-bb9f-4924-80b3-7de0d4805067 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: ab3edc53-37a5-4119-84c5-fd408ed66b69] Updating instance 'ab3edc53-37a5-4119-84c5-fd408ed66b69' progress to 67 {{(pid=61923) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1085.470084] env[61923]: DEBUG nova.compute.manager [None req-affadfdd-ba30-4ce8-8d78-e18297fead5a tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: c0a8a116-c7b6-4adc-849b-e0a4faa150d3] Starting instance... {{(pid=61923) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1085.663798] env[61923]: DEBUG oslo_concurrency.lockutils [None req-717de34a-2011-46e6-8553-55b971b239b8 tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Acquiring lock "35bdbd45-5838-4b09-a232-7f4a02322ad7" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1085.664053] env[61923]: DEBUG oslo_concurrency.lockutils [None req-717de34a-2011-46e6-8553-55b971b239b8 tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Lock "35bdbd45-5838-4b09-a232-7f4a02322ad7" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1085.767951] env[61923]: DEBUG nova.compute.manager [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] [instance: 624e9207-7f24-4f45-8f52-0fc7a60c1e84] Instance has had 0 of 5 cleanup attempts {{(pid=61923) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1085.918360] env[61923]: DEBUG nova.network.neutron [None req-e5fe1a98-bb9f-4924-80b3-7de0d4805067 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: ab3edc53-37a5-4119-84c5-fd408ed66b69] Port f2c777e2-f8f4-437b-b990-20e42dd2491b binding to destination host cpu-1 is already ACTIVE {{(pid=61923) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 1085.991579] env[61923]: DEBUG oslo_concurrency.lockutils [None req-affadfdd-ba30-4ce8-8d78-e18297fead5a tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1085.991832] env[61923]: DEBUG oslo_concurrency.lockutils [None req-affadfdd-ba30-4ce8-8d78-e18297fead5a tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1085.993292] env[61923]: INFO nova.compute.claims [None req-affadfdd-ba30-4ce8-8d78-e18297fead5a tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: c0a8a116-c7b6-4adc-849b-e0a4faa150d3] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1086.167871] env[61923]: DEBUG nova.compute.utils [None req-717de34a-2011-46e6-8553-55b971b239b8 tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Using /dev/sd instead of None {{(pid=61923) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1086.271278] env[61923]: DEBUG nova.compute.manager [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] [instance: c98c96a5-d18e-4519-996a-949ded16225f] Instance has had 0 of 5 cleanup attempts {{(pid=61923) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1086.672623] env[61923]: DEBUG oslo_concurrency.lockutils [None req-717de34a-2011-46e6-8553-55b971b239b8 tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Lock "35bdbd45-5838-4b09-a232-7f4a02322ad7" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.008s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1086.774596] env[61923]: DEBUG nova.compute.manager [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] [instance: bac71328-3796-4a65-aa41-ccda93bc7926] Instance has had 0 of 5 cleanup attempts {{(pid=61923) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1086.938221] env[61923]: DEBUG oslo_concurrency.lockutils [None req-e5fe1a98-bb9f-4924-80b3-7de0d4805067 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Acquiring lock "ab3edc53-37a5-4119-84c5-fd408ed66b69-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1086.938497] env[61923]: DEBUG oslo_concurrency.lockutils [None req-e5fe1a98-bb9f-4924-80b3-7de0d4805067 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Lock "ab3edc53-37a5-4119-84c5-fd408ed66b69-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1086.938632] env[61923]: DEBUG oslo_concurrency.lockutils [None req-e5fe1a98-bb9f-4924-80b3-7de0d4805067 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Lock "ab3edc53-37a5-4119-84c5-fd408ed66b69-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1087.053402] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a577989-3372-447e-8e0f-23bbf6391b4c {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.060786] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c1659fe-7896-4dc0-8b3b-f77a4793cd8f {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.090870] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68819827-e84b-42fe-8d3f-1ee0f361c124 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.097574] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe47935f-5bba-4c31-b000-39b1f379af19 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.110356] env[61923]: DEBUG nova.compute.provider_tree [None req-affadfdd-ba30-4ce8-8d78-e18297fead5a tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1087.277429] env[61923]: DEBUG nova.compute.manager [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] [instance: 263f0866-49af-4ab6-8132-d993642ebedb] Instance has had 0 of 5 cleanup attempts {{(pid=61923) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1087.614019] env[61923]: DEBUG nova.scheduler.client.report [None req-affadfdd-ba30-4ce8-8d78-e18297fead5a tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1087.728958] env[61923]: DEBUG oslo_concurrency.lockutils [None req-717de34a-2011-46e6-8553-55b971b239b8 tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Acquiring lock "35bdbd45-5838-4b09-a232-7f4a02322ad7" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1087.729270] env[61923]: DEBUG oslo_concurrency.lockutils [None req-717de34a-2011-46e6-8553-55b971b239b8 tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Lock "35bdbd45-5838-4b09-a232-7f4a02322ad7" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.001s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1087.729511] env[61923]: INFO nova.compute.manager [None req-717de34a-2011-46e6-8553-55b971b239b8 tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] [instance: 35bdbd45-5838-4b09-a232-7f4a02322ad7] Attaching volume 8744225b-4483-40f9-b314-ebab57d963bc to /dev/sdc [ 1087.758659] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f6494fc-7a00-47d0-968f-05fb38e1f6c0 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.765575] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-769a2573-4dba-49df-a30b-ca7c7065c15f {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.779246] env[61923]: DEBUG nova.virt.block_device [None req-717de34a-2011-46e6-8553-55b971b239b8 tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] [instance: 35bdbd45-5838-4b09-a232-7f4a02322ad7] Updating existing volume attachment record: 074fccd4-ab21-4547-84f5-ffe0f46708f0 {{(pid=61923) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1087.781294] env[61923]: DEBUG nova.compute.manager [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] [instance: 4c2c973a-5d23-4b8a-8b99-6b634e337054] Instance has had 0 of 5 cleanup attempts {{(pid=61923) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1088.006076] env[61923]: DEBUG oslo_concurrency.lockutils [None req-e5fe1a98-bb9f-4924-80b3-7de0d4805067 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Acquiring lock "refresh_cache-ab3edc53-37a5-4119-84c5-fd408ed66b69" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1088.006324] env[61923]: DEBUG oslo_concurrency.lockutils [None req-e5fe1a98-bb9f-4924-80b3-7de0d4805067 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Acquired lock "refresh_cache-ab3edc53-37a5-4119-84c5-fd408ed66b69" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1088.006507] env[61923]: DEBUG nova.network.neutron [None req-e5fe1a98-bb9f-4924-80b3-7de0d4805067 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: ab3edc53-37a5-4119-84c5-fd408ed66b69] Building network info cache for instance {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1088.119367] env[61923]: DEBUG oslo_concurrency.lockutils [None req-affadfdd-ba30-4ce8-8d78-e18297fead5a tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.127s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1088.119959] env[61923]: DEBUG nova.compute.manager [None req-affadfdd-ba30-4ce8-8d78-e18297fead5a tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: c0a8a116-c7b6-4adc-849b-e0a4faa150d3] Start building networks asynchronously for instance. {{(pid=61923) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1088.284449] env[61923]: DEBUG nova.compute.manager [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] [instance: 4cb3c397-ef1a-4267-a300-26da48206f3f] Instance has had 0 of 5 cleanup attempts {{(pid=61923) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1088.625392] env[61923]: DEBUG nova.compute.utils [None req-affadfdd-ba30-4ce8-8d78-e18297fead5a tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Using /dev/sd instead of None {{(pid=61923) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1088.626791] env[61923]: DEBUG nova.compute.manager [None req-affadfdd-ba30-4ce8-8d78-e18297fead5a tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: c0a8a116-c7b6-4adc-849b-e0a4faa150d3] Allocating IP information in the background. {{(pid=61923) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1088.627051] env[61923]: DEBUG nova.network.neutron [None req-affadfdd-ba30-4ce8-8d78-e18297fead5a tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: c0a8a116-c7b6-4adc-849b-e0a4faa150d3] allocate_for_instance() {{(pid=61923) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1088.664709] env[61923]: DEBUG nova.policy [None req-affadfdd-ba30-4ce8-8d78-e18297fead5a tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '69be89006ff141b686261d643b3b9a73', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '780d3541d9604417b977bb62390c4299', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61923) authorize /opt/stack/nova/nova/policy.py:201}} [ 1088.695576] env[61923]: DEBUG nova.network.neutron [None req-e5fe1a98-bb9f-4924-80b3-7de0d4805067 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: ab3edc53-37a5-4119-84c5-fd408ed66b69] Updating instance_info_cache with network_info: [{"id": "f2c777e2-f8f4-437b-b990-20e42dd2491b", "address": "fa:16:3e:b3:a4:a2", "network": {"id": "77aec6ca-d1be-4b39-9d51-08853a1183a1", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1813085414-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5a4db4f8aa194a219e80c538cc715d43", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "56b944d8-803d-43f2-945d-0f334ee4ea1c", "external-id": "nsx-vlan-transportzone-799", "segmentation_id": 799, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf2c777e2-f8", "ovs_interfaceid": "f2c777e2-f8f4-437b-b990-20e42dd2491b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1088.787437] env[61923]: DEBUG nova.compute.manager [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] [instance: a511392e-9ab6-42fb-b07c-f90e9786dcc1] Instance has had 0 of 5 cleanup attempts {{(pid=61923) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1089.090100] env[61923]: DEBUG nova.network.neutron [None req-affadfdd-ba30-4ce8-8d78-e18297fead5a tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: c0a8a116-c7b6-4adc-849b-e0a4faa150d3] Successfully created port: c210c30f-4f3c-4698-ab24-a6a81c78e658 {{(pid=61923) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1089.131392] env[61923]: DEBUG nova.compute.manager [None req-affadfdd-ba30-4ce8-8d78-e18297fead5a tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: c0a8a116-c7b6-4adc-849b-e0a4faa150d3] Start building block device mappings for instance. {{(pid=61923) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1089.199086] env[61923]: DEBUG oslo_concurrency.lockutils [None req-e5fe1a98-bb9f-4924-80b3-7de0d4805067 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Releasing lock "refresh_cache-ab3edc53-37a5-4119-84c5-fd408ed66b69" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1089.292345] env[61923]: DEBUG nova.compute.manager [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] [instance: 6012d3b5-739c-4762-9bb4-09c51171dcd7] Instance has had 0 of 5 cleanup attempts {{(pid=61923) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1089.723665] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3aa4bed2-bf3a-433d-a1f1-272c860019d7 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.742703] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71b97938-f37e-49f0-a950-bd47d0df4181 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.751595] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-e5fe1a98-bb9f-4924-80b3-7de0d4805067 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: ab3edc53-37a5-4119-84c5-fd408ed66b69] Updating instance 'ab3edc53-37a5-4119-84c5-fd408ed66b69' progress to 83 {{(pid=61923) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1089.793392] env[61923]: DEBUG nova.compute.manager [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] [instance: b779d183-89ae-4e4d-ae99-e514e145ed43] Instance has had 0 of 5 cleanup attempts {{(pid=61923) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1090.141730] env[61923]: DEBUG nova.compute.manager [None req-affadfdd-ba30-4ce8-8d78-e18297fead5a tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: c0a8a116-c7b6-4adc-849b-e0a4faa150d3] Start spawning the instance on the hypervisor. {{(pid=61923) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1090.167796] env[61923]: DEBUG nova.virt.hardware [None req-affadfdd-ba30-4ce8-8d78-e18297fead5a tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-29T20:07:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-29T20:07:35Z,direct_url=,disk_format='vmdk',id=0f153f63-ae0a-45b1-b7f5-7f9b673c947f,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='4e7b5e3b3c3443c8bd5c70f8a15739f5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-29T20:07:36Z,virtual_size=,visibility=), allow threads: False {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1090.168076] env[61923]: DEBUG nova.virt.hardware [None req-affadfdd-ba30-4ce8-8d78-e18297fead5a tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Flavor limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1090.168248] env[61923]: DEBUG nova.virt.hardware [None req-affadfdd-ba30-4ce8-8d78-e18297fead5a tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Image limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1090.168435] env[61923]: DEBUG nova.virt.hardware [None req-affadfdd-ba30-4ce8-8d78-e18297fead5a tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Flavor pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1090.168584] env[61923]: DEBUG nova.virt.hardware [None req-affadfdd-ba30-4ce8-8d78-e18297fead5a tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Image pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1090.168734] env[61923]: DEBUG nova.virt.hardware [None req-affadfdd-ba30-4ce8-8d78-e18297fead5a tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1090.168949] env[61923]: DEBUG nova.virt.hardware [None req-affadfdd-ba30-4ce8-8d78-e18297fead5a tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1090.169160] env[61923]: DEBUG nova.virt.hardware [None req-affadfdd-ba30-4ce8-8d78-e18297fead5a tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1090.169334] env[61923]: DEBUG nova.virt.hardware [None req-affadfdd-ba30-4ce8-8d78-e18297fead5a tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Got 1 possible topologies {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1090.169495] env[61923]: DEBUG nova.virt.hardware [None req-affadfdd-ba30-4ce8-8d78-e18297fead5a tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1090.169666] env[61923]: DEBUG nova.virt.hardware [None req-affadfdd-ba30-4ce8-8d78-e18297fead5a tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1090.170533] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f6f2072-97ab-4cf0-ba15-87a5f85a9c29 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.177840] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b4ae476-aedd-4dc6-a369-2f1e83c0045b {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.260366] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-e5fe1a98-bb9f-4924-80b3-7de0d4805067 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: ab3edc53-37a5-4119-84c5-fd408ed66b69] Powering on the VM {{(pid=61923) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1090.260571] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b6e97811-85f1-4d43-b63a-e70b6172357b {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.267201] env[61923]: DEBUG oslo_vmware.api [None req-e5fe1a98-bb9f-4924-80b3-7de0d4805067 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Waiting for the task: (returnval){ [ 1090.267201] env[61923]: value = "task-1378020" [ 1090.267201] env[61923]: _type = "Task" [ 1090.267201] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1090.274410] env[61923]: DEBUG oslo_vmware.api [None req-e5fe1a98-bb9f-4924-80b3-7de0d4805067 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Task: {'id': task-1378020, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1090.296993] env[61923]: DEBUG nova.compute.manager [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] [instance: 5b73e07a-d6ef-4dcf-bdae-eea91d2aeb6f] Instance has had 0 of 5 cleanup attempts {{(pid=61923) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1090.461251] env[61923]: DEBUG nova.compute.manager [req-71062b06-92b8-43b3-bd84-6aa0d55f72c4 req-3c7ac1a8-91d1-40e3-8db7-94004d0b4447 service nova] [instance: c0a8a116-c7b6-4adc-849b-e0a4faa150d3] Received event network-vif-plugged-c210c30f-4f3c-4698-ab24-a6a81c78e658 {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1090.461736] env[61923]: DEBUG oslo_concurrency.lockutils [req-71062b06-92b8-43b3-bd84-6aa0d55f72c4 req-3c7ac1a8-91d1-40e3-8db7-94004d0b4447 service nova] Acquiring lock "c0a8a116-c7b6-4adc-849b-e0a4faa150d3-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1090.461970] env[61923]: DEBUG oslo_concurrency.lockutils [req-71062b06-92b8-43b3-bd84-6aa0d55f72c4 req-3c7ac1a8-91d1-40e3-8db7-94004d0b4447 service nova] Lock "c0a8a116-c7b6-4adc-849b-e0a4faa150d3-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1090.462191] env[61923]: DEBUG oslo_concurrency.lockutils [req-71062b06-92b8-43b3-bd84-6aa0d55f72c4 req-3c7ac1a8-91d1-40e3-8db7-94004d0b4447 service nova] Lock "c0a8a116-c7b6-4adc-849b-e0a4faa150d3-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1090.462418] env[61923]: DEBUG nova.compute.manager [req-71062b06-92b8-43b3-bd84-6aa0d55f72c4 req-3c7ac1a8-91d1-40e3-8db7-94004d0b4447 service nova] [instance: c0a8a116-c7b6-4adc-849b-e0a4faa150d3] No waiting events found dispatching network-vif-plugged-c210c30f-4f3c-4698-ab24-a6a81c78e658 {{(pid=61923) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1090.462598] env[61923]: WARNING nova.compute.manager [req-71062b06-92b8-43b3-bd84-6aa0d55f72c4 req-3c7ac1a8-91d1-40e3-8db7-94004d0b4447 service nova] [instance: c0a8a116-c7b6-4adc-849b-e0a4faa150d3] Received unexpected event network-vif-plugged-c210c30f-4f3c-4698-ab24-a6a81c78e658 for instance with vm_state building and task_state spawning. [ 1090.554686] env[61923]: DEBUG nova.network.neutron [None req-affadfdd-ba30-4ce8-8d78-e18297fead5a tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: c0a8a116-c7b6-4adc-849b-e0a4faa150d3] Successfully updated port: c210c30f-4f3c-4698-ab24-a6a81c78e658 {{(pid=61923) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1090.777517] env[61923]: DEBUG oslo_vmware.api [None req-e5fe1a98-bb9f-4924-80b3-7de0d4805067 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Task: {'id': task-1378020, 'name': PowerOnVM_Task, 'duration_secs': 0.348038} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1090.777790] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-e5fe1a98-bb9f-4924-80b3-7de0d4805067 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: ab3edc53-37a5-4119-84c5-fd408ed66b69] Powered on the VM {{(pid=61923) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1090.777983] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-e5fe1a98-bb9f-4924-80b3-7de0d4805067 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: ab3edc53-37a5-4119-84c5-fd408ed66b69] Updating instance 'ab3edc53-37a5-4119-84c5-fd408ed66b69' progress to 100 {{(pid=61923) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1090.800484] env[61923]: DEBUG nova.compute.manager [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] [instance: e413c45d-cd89-44d4-9102-3d188907e7bb] Instance has had 0 of 5 cleanup attempts {{(pid=61923) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1091.057327] env[61923]: DEBUG oslo_concurrency.lockutils [None req-affadfdd-ba30-4ce8-8d78-e18297fead5a tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Acquiring lock "refresh_cache-c0a8a116-c7b6-4adc-849b-e0a4faa150d3" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1091.057514] env[61923]: DEBUG oslo_concurrency.lockutils [None req-affadfdd-ba30-4ce8-8d78-e18297fead5a tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Acquired lock "refresh_cache-c0a8a116-c7b6-4adc-849b-e0a4faa150d3" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1091.057634] env[61923]: DEBUG nova.network.neutron [None req-affadfdd-ba30-4ce8-8d78-e18297fead5a tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: c0a8a116-c7b6-4adc-849b-e0a4faa150d3] Building network info cache for instance {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1091.304185] env[61923]: DEBUG nova.compute.manager [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] [instance: 444affa5-a7ed-4a17-9015-9fd5724aab64] Instance has had 0 of 5 cleanup attempts {{(pid=61923) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1091.592208] env[61923]: DEBUG nova.network.neutron [None req-affadfdd-ba30-4ce8-8d78-e18297fead5a tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: c0a8a116-c7b6-4adc-849b-e0a4faa150d3] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1091.727606] env[61923]: DEBUG nova.network.neutron [None req-affadfdd-ba30-4ce8-8d78-e18297fead5a tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: c0a8a116-c7b6-4adc-849b-e0a4faa150d3] Updating instance_info_cache with network_info: [{"id": "c210c30f-4f3c-4698-ab24-a6a81c78e658", "address": "fa:16:3e:ad:1c:c4", "network": {"id": "9533e79d-fde2-4c10-86a0-86a36845a8cf", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1736743137-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "780d3541d9604417b977bb62390c4299", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6d650b26-c3e7-4de7-98db-5e4b816d123a", "external-id": "nsx-vlan-transportzone-757", "segmentation_id": 757, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc210c30f-4f", "ovs_interfaceid": "c210c30f-4f3c-4698-ab24-a6a81c78e658", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1091.807375] env[61923]: DEBUG nova.compute.manager [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] [instance: 2719569b-8572-4199-8158-7bb367d17dc5] Instance has had 0 of 5 cleanup attempts {{(pid=61923) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1092.230768] env[61923]: DEBUG oslo_concurrency.lockutils [None req-affadfdd-ba30-4ce8-8d78-e18297fead5a tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Releasing lock "refresh_cache-c0a8a116-c7b6-4adc-849b-e0a4faa150d3" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1092.230976] env[61923]: DEBUG nova.compute.manager [None req-affadfdd-ba30-4ce8-8d78-e18297fead5a tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: c0a8a116-c7b6-4adc-849b-e0a4faa150d3] Instance network_info: |[{"id": "c210c30f-4f3c-4698-ab24-a6a81c78e658", "address": "fa:16:3e:ad:1c:c4", "network": {"id": "9533e79d-fde2-4c10-86a0-86a36845a8cf", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1736743137-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "780d3541d9604417b977bb62390c4299", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6d650b26-c3e7-4de7-98db-5e4b816d123a", "external-id": "nsx-vlan-transportzone-757", "segmentation_id": 757, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc210c30f-4f", "ovs_interfaceid": "c210c30f-4f3c-4698-ab24-a6a81c78e658", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61923) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1092.231459] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-affadfdd-ba30-4ce8-8d78-e18297fead5a tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: c0a8a116-c7b6-4adc-849b-e0a4faa150d3] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ad:1c:c4', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '6d650b26-c3e7-4de7-98db-5e4b816d123a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'c210c30f-4f3c-4698-ab24-a6a81c78e658', 'vif_model': 'vmxnet3'}] {{(pid=61923) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1092.238763] env[61923]: DEBUG oslo.service.loopingcall [None req-affadfdd-ba30-4ce8-8d78-e18297fead5a tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61923) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1092.238972] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c0a8a116-c7b6-4adc-849b-e0a4faa150d3] Creating VM on the ESX host {{(pid=61923) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1092.239208] env[61923]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-755c5a04-0473-4cca-a3ad-8abd8d8f1eaf {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.260050] env[61923]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1092.260050] env[61923]: value = "task-1378021" [ 1092.260050] env[61923]: _type = "Task" [ 1092.260050] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1092.267475] env[61923]: DEBUG oslo_vmware.api [-] Task: {'id': task-1378021, 'name': CreateVM_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1092.310280] env[61923]: DEBUG nova.compute.manager [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] [instance: 87ae37dc-d0d0-4b76-afdd-0ba3e8f6ce0b] Instance has had 0 of 5 cleanup attempts {{(pid=61923) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1092.323653] env[61923]: DEBUG nova.virt.vmwareapi.volumeops [None req-717de34a-2011-46e6-8553-55b971b239b8 tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] [instance: 35bdbd45-5838-4b09-a232-7f4a02322ad7] Volume attach. Driver type: vmdk {{(pid=61923) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1092.324079] env[61923]: DEBUG nova.virt.vmwareapi.volumeops [None req-717de34a-2011-46e6-8553-55b971b239b8 tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] [instance: 35bdbd45-5838-4b09-a232-7f4a02322ad7] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-292754', 'volume_id': '8744225b-4483-40f9-b314-ebab57d963bc', 'name': 'volume-8744225b-4483-40f9-b314-ebab57d963bc', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '35bdbd45-5838-4b09-a232-7f4a02322ad7', 'attached_at': '', 'detached_at': '', 'volume_id': '8744225b-4483-40f9-b314-ebab57d963bc', 'serial': '8744225b-4483-40f9-b314-ebab57d963bc'} {{(pid=61923) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1092.325667] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99d6b335-763a-4920-8066-6391483a47cb {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.343242] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dda42cac-874b-4efb-93f0-11c7ee0edd93 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.369892] env[61923]: DEBUG nova.virt.vmwareapi.volumeops [None req-717de34a-2011-46e6-8553-55b971b239b8 tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] [instance: 35bdbd45-5838-4b09-a232-7f4a02322ad7] Reconfiguring VM instance instance-00000060 to attach disk [datastore2] volume-8744225b-4483-40f9-b314-ebab57d963bc/volume-8744225b-4483-40f9-b314-ebab57d963bc.vmdk or device None with type thin {{(pid=61923) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1092.370430] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b4947a6d-69bd-499a-96f2-cc750704e1b8 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.387962] env[61923]: DEBUG oslo_vmware.api [None req-717de34a-2011-46e6-8553-55b971b239b8 tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Waiting for the task: (returnval){ [ 1092.387962] env[61923]: value = "task-1378022" [ 1092.387962] env[61923]: _type = "Task" [ 1092.387962] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1092.395692] env[61923]: DEBUG oslo_vmware.api [None req-717de34a-2011-46e6-8553-55b971b239b8 tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Task: {'id': task-1378022, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1092.488964] env[61923]: DEBUG nova.compute.manager [req-886f5c3b-3b8f-4d4d-a5af-39f53c3e5b78 req-e9a62c60-044c-4318-a840-d58e1b941f29 service nova] [instance: c0a8a116-c7b6-4adc-849b-e0a4faa150d3] Received event network-changed-c210c30f-4f3c-4698-ab24-a6a81c78e658 {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1092.489270] env[61923]: DEBUG nova.compute.manager [req-886f5c3b-3b8f-4d4d-a5af-39f53c3e5b78 req-e9a62c60-044c-4318-a840-d58e1b941f29 service nova] [instance: c0a8a116-c7b6-4adc-849b-e0a4faa150d3] Refreshing instance network info cache due to event network-changed-c210c30f-4f3c-4698-ab24-a6a81c78e658. {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1092.489540] env[61923]: DEBUG oslo_concurrency.lockutils [req-886f5c3b-3b8f-4d4d-a5af-39f53c3e5b78 req-e9a62c60-044c-4318-a840-d58e1b941f29 service nova] Acquiring lock "refresh_cache-c0a8a116-c7b6-4adc-849b-e0a4faa150d3" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1092.489669] env[61923]: DEBUG oslo_concurrency.lockutils [req-886f5c3b-3b8f-4d4d-a5af-39f53c3e5b78 req-e9a62c60-044c-4318-a840-d58e1b941f29 service nova] Acquired lock "refresh_cache-c0a8a116-c7b6-4adc-849b-e0a4faa150d3" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1092.489852] env[61923]: DEBUG nova.network.neutron [req-886f5c3b-3b8f-4d4d-a5af-39f53c3e5b78 req-e9a62c60-044c-4318-a840-d58e1b941f29 service nova] [instance: c0a8a116-c7b6-4adc-849b-e0a4faa150d3] Refreshing network info cache for port c210c30f-4f3c-4698-ab24-a6a81c78e658 {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1092.770070] env[61923]: DEBUG oslo_vmware.api [-] Task: {'id': task-1378021, 'name': CreateVM_Task, 'duration_secs': 0.31217} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1092.770245] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c0a8a116-c7b6-4adc-849b-e0a4faa150d3] Created VM on the ESX host {{(pid=61923) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1092.770957] env[61923]: DEBUG oslo_concurrency.lockutils [None req-affadfdd-ba30-4ce8-8d78-e18297fead5a tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1092.771192] env[61923]: DEBUG oslo_concurrency.lockutils [None req-affadfdd-ba30-4ce8-8d78-e18297fead5a tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1092.771823] env[61923]: DEBUG oslo_concurrency.lockutils [None req-affadfdd-ba30-4ce8-8d78-e18297fead5a tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1092.772123] env[61923]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5120f8c0-951a-4344-a231-12737b46c22d {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.776715] env[61923]: DEBUG oslo_vmware.api [None req-affadfdd-ba30-4ce8-8d78-e18297fead5a tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Waiting for the task: (returnval){ [ 1092.776715] env[61923]: value = "session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]5236cf90-7a60-723d-2f30-c95c7aa2d1b5" [ 1092.776715] env[61923]: _type = "Task" [ 1092.776715] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1092.784615] env[61923]: DEBUG oslo_vmware.api [None req-affadfdd-ba30-4ce8-8d78-e18297fead5a tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]5236cf90-7a60-723d-2f30-c95c7aa2d1b5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1092.814043] env[61923]: DEBUG nova.compute.manager [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] [instance: af3c317a-4007-4cea-a060-1e7dde5ce49e] Instance has had 0 of 5 cleanup attempts {{(pid=61923) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1092.900049] env[61923]: DEBUG oslo_vmware.api [None req-717de34a-2011-46e6-8553-55b971b239b8 tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Task: {'id': task-1378022, 'name': ReconfigVM_Task, 'duration_secs': 0.337762} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1092.900375] env[61923]: DEBUG nova.virt.vmwareapi.volumeops [None req-717de34a-2011-46e6-8553-55b971b239b8 tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] [instance: 35bdbd45-5838-4b09-a232-7f4a02322ad7] Reconfigured VM instance instance-00000060 to attach disk [datastore2] volume-8744225b-4483-40f9-b314-ebab57d963bc/volume-8744225b-4483-40f9-b314-ebab57d963bc.vmdk or device None with type thin {{(pid=61923) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1092.905513] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8348fcea-a2d6-4bfc-aa37-80502dff88d7 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.921021] env[61923]: DEBUG oslo_vmware.api [None req-717de34a-2011-46e6-8553-55b971b239b8 tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Waiting for the task: (returnval){ [ 1092.921021] env[61923]: value = "task-1378023" [ 1092.921021] env[61923]: _type = "Task" [ 1092.921021] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1092.928922] env[61923]: DEBUG oslo_vmware.api [None req-717de34a-2011-46e6-8553-55b971b239b8 tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Task: {'id': task-1378023, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1093.251182] env[61923]: DEBUG nova.network.neutron [req-886f5c3b-3b8f-4d4d-a5af-39f53c3e5b78 req-e9a62c60-044c-4318-a840-d58e1b941f29 service nova] [instance: c0a8a116-c7b6-4adc-849b-e0a4faa150d3] Updated VIF entry in instance network info cache for port c210c30f-4f3c-4698-ab24-a6a81c78e658. {{(pid=61923) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1093.251590] env[61923]: DEBUG nova.network.neutron [req-886f5c3b-3b8f-4d4d-a5af-39f53c3e5b78 req-e9a62c60-044c-4318-a840-d58e1b941f29 service nova] [instance: c0a8a116-c7b6-4adc-849b-e0a4faa150d3] Updating instance_info_cache with network_info: [{"id": "c210c30f-4f3c-4698-ab24-a6a81c78e658", "address": "fa:16:3e:ad:1c:c4", "network": {"id": "9533e79d-fde2-4c10-86a0-86a36845a8cf", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1736743137-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "780d3541d9604417b977bb62390c4299", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6d650b26-c3e7-4de7-98db-5e4b816d123a", "external-id": "nsx-vlan-transportzone-757", "segmentation_id": 757, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc210c30f-4f", "ovs_interfaceid": "c210c30f-4f3c-4698-ab24-a6a81c78e658", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1093.287087] env[61923]: DEBUG oslo_vmware.api [None req-affadfdd-ba30-4ce8-8d78-e18297fead5a tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]5236cf90-7a60-723d-2f30-c95c7aa2d1b5, 'name': SearchDatastore_Task, 'duration_secs': 0.009246} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1093.287353] env[61923]: DEBUG oslo_concurrency.lockutils [None req-affadfdd-ba30-4ce8-8d78-e18297fead5a tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1093.287587] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-affadfdd-ba30-4ce8-8d78-e18297fead5a tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: c0a8a116-c7b6-4adc-849b-e0a4faa150d3] Processing image 0f153f63-ae0a-45b1-b7f5-7f9b673c947f {{(pid=61923) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1093.287820] env[61923]: DEBUG oslo_concurrency.lockutils [None req-affadfdd-ba30-4ce8-8d78-e18297fead5a tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1093.287969] env[61923]: DEBUG oslo_concurrency.lockutils [None req-affadfdd-ba30-4ce8-8d78-e18297fead5a tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1093.288164] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-affadfdd-ba30-4ce8-8d78-e18297fead5a tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61923) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1093.288418] env[61923]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-18997111-53b6-4479-ba8b-3d1f09415715 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.310622] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-affadfdd-ba30-4ce8-8d78-e18297fead5a tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61923) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1093.311046] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-affadfdd-ba30-4ce8-8d78-e18297fead5a tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61923) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1093.311876] env[61923]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9b8a6e70-4247-4d55-b94e-8703de113b16 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.316693] env[61923]: DEBUG nova.compute.manager [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] [instance: fc4fdf89-4615-4586-9dcd-ea90dc258361] Instance has had 0 of 5 cleanup attempts {{(pid=61923) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1093.319956] env[61923]: DEBUG oslo_vmware.api [None req-affadfdd-ba30-4ce8-8d78-e18297fead5a tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Waiting for the task: (returnval){ [ 1093.319956] env[61923]: value = "session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52f73789-ed35-b83d-fb1d-78b7a1f7035f" [ 1093.319956] env[61923]: _type = "Task" [ 1093.319956] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1093.327976] env[61923]: DEBUG oslo_vmware.api [None req-affadfdd-ba30-4ce8-8d78-e18297fead5a tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52f73789-ed35-b83d-fb1d-78b7a1f7035f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1093.430872] env[61923]: DEBUG oslo_vmware.api [None req-717de34a-2011-46e6-8553-55b971b239b8 tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Task: {'id': task-1378023, 'name': ReconfigVM_Task, 'duration_secs': 0.134828} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1093.431251] env[61923]: DEBUG nova.virt.vmwareapi.volumeops [None req-717de34a-2011-46e6-8553-55b971b239b8 tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] [instance: 35bdbd45-5838-4b09-a232-7f4a02322ad7] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-292754', 'volume_id': '8744225b-4483-40f9-b314-ebab57d963bc', 'name': 'volume-8744225b-4483-40f9-b314-ebab57d963bc', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '35bdbd45-5838-4b09-a232-7f4a02322ad7', 'attached_at': '', 'detached_at': '', 'volume_id': '8744225b-4483-40f9-b314-ebab57d963bc', 'serial': '8744225b-4483-40f9-b314-ebab57d963bc'} {{(pid=61923) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1093.457493] env[61923]: DEBUG oslo_concurrency.lockutils [None req-781874ef-b4d2-4025-bf94-9b496335d5ac tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Acquiring lock "ab3edc53-37a5-4119-84c5-fd408ed66b69" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1093.457759] env[61923]: DEBUG oslo_concurrency.lockutils [None req-781874ef-b4d2-4025-bf94-9b496335d5ac tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Lock "ab3edc53-37a5-4119-84c5-fd408ed66b69" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1093.457951] env[61923]: DEBUG nova.compute.manager [None req-781874ef-b4d2-4025-bf94-9b496335d5ac tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: ab3edc53-37a5-4119-84c5-fd408ed66b69] Going to confirm migration 1 {{(pid=61923) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:4783}} [ 1093.754451] env[61923]: DEBUG oslo_concurrency.lockutils [req-886f5c3b-3b8f-4d4d-a5af-39f53c3e5b78 req-e9a62c60-044c-4318-a840-d58e1b941f29 service nova] Releasing lock "refresh_cache-c0a8a116-c7b6-4adc-849b-e0a4faa150d3" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1093.821424] env[61923]: DEBUG nova.compute.manager [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] [instance: 59198f4d-4dde-4eaf-9f6c-a962cbe53c6e] Instance has had 0 of 5 cleanup attempts {{(pid=61923) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1093.832895] env[61923]: DEBUG oslo_vmware.api [None req-affadfdd-ba30-4ce8-8d78-e18297fead5a tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52f73789-ed35-b83d-fb1d-78b7a1f7035f, 'name': SearchDatastore_Task, 'duration_secs': 0.037502} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1093.833703] env[61923]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-da962d64-767c-479b-8403-b8f3735679ad {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1093.838543] env[61923]: DEBUG oslo_vmware.api [None req-affadfdd-ba30-4ce8-8d78-e18297fead5a tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Waiting for the task: (returnval){ [ 1093.838543] env[61923]: value = "session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52bfd7f4-fb27-0a37-9bcf-bcf01ab9d261" [ 1093.838543] env[61923]: _type = "Task" [ 1093.838543] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1093.845858] env[61923]: DEBUG oslo_vmware.api [None req-affadfdd-ba30-4ce8-8d78-e18297fead5a tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52bfd7f4-fb27-0a37-9bcf-bcf01ab9d261, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1094.014393] env[61923]: DEBUG oslo_concurrency.lockutils [None req-781874ef-b4d2-4025-bf94-9b496335d5ac tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Acquiring lock "refresh_cache-ab3edc53-37a5-4119-84c5-fd408ed66b69" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1094.014578] env[61923]: DEBUG oslo_concurrency.lockutils [None req-781874ef-b4d2-4025-bf94-9b496335d5ac tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Acquired lock "refresh_cache-ab3edc53-37a5-4119-84c5-fd408ed66b69" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1094.014756] env[61923]: DEBUG nova.network.neutron [None req-781874ef-b4d2-4025-bf94-9b496335d5ac tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: ab3edc53-37a5-4119-84c5-fd408ed66b69] Building network info cache for instance {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1094.014941] env[61923]: DEBUG nova.objects.instance [None req-781874ef-b4d2-4025-bf94-9b496335d5ac tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Lazy-loading 'info_cache' on Instance uuid ab3edc53-37a5-4119-84c5-fd408ed66b69 {{(pid=61923) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1094.327548] env[61923]: DEBUG nova.compute.manager [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] [instance: 73692517-1816-4e9b-ab2d-8265c683e83d] Instance has had 0 of 5 cleanup attempts {{(pid=61923) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1094.348531] env[61923]: DEBUG oslo_vmware.api [None req-affadfdd-ba30-4ce8-8d78-e18297fead5a tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52bfd7f4-fb27-0a37-9bcf-bcf01ab9d261, 'name': SearchDatastore_Task, 'duration_secs': 0.009964} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1094.349364] env[61923]: DEBUG oslo_concurrency.lockutils [None req-affadfdd-ba30-4ce8-8d78-e18297fead5a tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1094.349628] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-affadfdd-ba30-4ce8-8d78-e18297fead5a tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk to [datastore2] c0a8a116-c7b6-4adc-849b-e0a4faa150d3/c0a8a116-c7b6-4adc-849b-e0a4faa150d3.vmdk {{(pid=61923) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1094.349881] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-9ffa8c32-d5cc-4c90-9c2d-6e0c2a23baf7 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.355930] env[61923]: DEBUG oslo_vmware.api [None req-affadfdd-ba30-4ce8-8d78-e18297fead5a tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Waiting for the task: (returnval){ [ 1094.355930] env[61923]: value = "task-1378024" [ 1094.355930] env[61923]: _type = "Task" [ 1094.355930] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1094.363344] env[61923]: DEBUG oslo_vmware.api [None req-affadfdd-ba30-4ce8-8d78-e18297fead5a tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Task: {'id': task-1378024, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1094.467554] env[61923]: DEBUG nova.objects.instance [None req-717de34a-2011-46e6-8553-55b971b239b8 tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Lazy-loading 'flavor' on Instance uuid 35bdbd45-5838-4b09-a232-7f4a02322ad7 {{(pid=61923) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1094.831300] env[61923]: DEBUG nova.compute.manager [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] [instance: 422d06d6-6932-46c3-bb25-841e6f65c028] Instance has had 0 of 5 cleanup attempts {{(pid=61923) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1094.865576] env[61923]: DEBUG oslo_vmware.api [None req-affadfdd-ba30-4ce8-8d78-e18297fead5a tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Task: {'id': task-1378024, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.460399} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1094.865844] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-affadfdd-ba30-4ce8-8d78-e18297fead5a tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk to [datastore2] c0a8a116-c7b6-4adc-849b-e0a4faa150d3/c0a8a116-c7b6-4adc-849b-e0a4faa150d3.vmdk {{(pid=61923) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1094.866075] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-affadfdd-ba30-4ce8-8d78-e18297fead5a tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: c0a8a116-c7b6-4adc-849b-e0a4faa150d3] Extending root virtual disk to 1048576 {{(pid=61923) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1094.866337] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-43dfe5aa-3e53-4692-8c6c-e693b30bec55 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.872345] env[61923]: DEBUG oslo_vmware.api [None req-affadfdd-ba30-4ce8-8d78-e18297fead5a tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Waiting for the task: (returnval){ [ 1094.872345] env[61923]: value = "task-1378025" [ 1094.872345] env[61923]: _type = "Task" [ 1094.872345] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1094.880776] env[61923]: DEBUG oslo_vmware.api [None req-affadfdd-ba30-4ce8-8d78-e18297fead5a tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Task: {'id': task-1378025, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1094.973224] env[61923]: DEBUG oslo_concurrency.lockutils [None req-717de34a-2011-46e6-8553-55b971b239b8 tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Lock "35bdbd45-5838-4b09-a232-7f4a02322ad7" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.244s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1095.203670] env[61923]: DEBUG nova.network.neutron [None req-781874ef-b4d2-4025-bf94-9b496335d5ac tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: ab3edc53-37a5-4119-84c5-fd408ed66b69] Updating instance_info_cache with network_info: [{"id": "f2c777e2-f8f4-437b-b990-20e42dd2491b", "address": "fa:16:3e:b3:a4:a2", "network": {"id": "77aec6ca-d1be-4b39-9d51-08853a1183a1", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1813085414-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5a4db4f8aa194a219e80c538cc715d43", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "56b944d8-803d-43f2-945d-0f334ee4ea1c", "external-id": "nsx-vlan-transportzone-799", "segmentation_id": 799, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf2c777e2-f8", "ovs_interfaceid": "f2c777e2-f8f4-437b-b990-20e42dd2491b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1095.316843] env[61923]: DEBUG oslo_concurrency.lockutils [None req-9155cecc-8b54-46b6-9db8-ad98ae309cb1 tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Acquiring lock "35bdbd45-5838-4b09-a232-7f4a02322ad7" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1095.317187] env[61923]: DEBUG oslo_concurrency.lockutils [None req-9155cecc-8b54-46b6-9db8-ad98ae309cb1 tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Lock "35bdbd45-5838-4b09-a232-7f4a02322ad7" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1095.334832] env[61923]: DEBUG nova.compute.manager [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] [instance: 48825592-dcdc-41d5-9fbf-500d1f31017b] Instance has had 0 of 5 cleanup attempts {{(pid=61923) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1095.381811] env[61923]: DEBUG oslo_vmware.api [None req-affadfdd-ba30-4ce8-8d78-e18297fead5a tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Task: {'id': task-1378025, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.055773} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1095.382114] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-affadfdd-ba30-4ce8-8d78-e18297fead5a tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: c0a8a116-c7b6-4adc-849b-e0a4faa150d3] Extended root virtual disk {{(pid=61923) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1095.382904] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cfa9b6f6-361c-4c41-99e8-4a9be7751974 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.405738] env[61923]: DEBUG nova.virt.vmwareapi.volumeops [None req-affadfdd-ba30-4ce8-8d78-e18297fead5a tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: c0a8a116-c7b6-4adc-849b-e0a4faa150d3] Reconfiguring VM instance instance-00000065 to attach disk [datastore2] c0a8a116-c7b6-4adc-849b-e0a4faa150d3/c0a8a116-c7b6-4adc-849b-e0a4faa150d3.vmdk or device None with type sparse {{(pid=61923) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1095.405994] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1bfd51f1-a13b-4bb5-b551-1c2a3dd71114 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.424299] env[61923]: DEBUG oslo_vmware.api [None req-affadfdd-ba30-4ce8-8d78-e18297fead5a tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Waiting for the task: (returnval){ [ 1095.424299] env[61923]: value = "task-1378026" [ 1095.424299] env[61923]: _type = "Task" [ 1095.424299] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1095.432517] env[61923]: DEBUG oslo_vmware.api [None req-affadfdd-ba30-4ce8-8d78-e18297fead5a tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Task: {'id': task-1378026, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1095.706039] env[61923]: DEBUG oslo_concurrency.lockutils [None req-781874ef-b4d2-4025-bf94-9b496335d5ac tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Releasing lock "refresh_cache-ab3edc53-37a5-4119-84c5-fd408ed66b69" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1095.706298] env[61923]: DEBUG nova.objects.instance [None req-781874ef-b4d2-4025-bf94-9b496335d5ac tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Lazy-loading 'migration_context' on Instance uuid ab3edc53-37a5-4119-84c5-fd408ed66b69 {{(pid=61923) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1095.820518] env[61923]: INFO nova.compute.manager [None req-9155cecc-8b54-46b6-9db8-ad98ae309cb1 tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] [instance: 35bdbd45-5838-4b09-a232-7f4a02322ad7] Detaching volume f908034d-f063-4b8d-9877-08531afe6e62 [ 1095.838058] env[61923]: DEBUG nova.compute.manager [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] [instance: f1839f61-4314-48fe-8ab6-14b5e49d569d] Instance has had 0 of 5 cleanup attempts {{(pid=61923) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1095.850965] env[61923]: INFO nova.virt.block_device [None req-9155cecc-8b54-46b6-9db8-ad98ae309cb1 tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] [instance: 35bdbd45-5838-4b09-a232-7f4a02322ad7] Attempting to driver detach volume f908034d-f063-4b8d-9877-08531afe6e62 from mountpoint /dev/sdb [ 1095.852060] env[61923]: DEBUG nova.virt.vmwareapi.volumeops [None req-9155cecc-8b54-46b6-9db8-ad98ae309cb1 tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] [instance: 35bdbd45-5838-4b09-a232-7f4a02322ad7] Volume detach. Driver type: vmdk {{(pid=61923) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1095.852060] env[61923]: DEBUG nova.virt.vmwareapi.volumeops [None req-9155cecc-8b54-46b6-9db8-ad98ae309cb1 tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] [instance: 35bdbd45-5838-4b09-a232-7f4a02322ad7] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-292753', 'volume_id': 'f908034d-f063-4b8d-9877-08531afe6e62', 'name': 'volume-f908034d-f063-4b8d-9877-08531afe6e62', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '35bdbd45-5838-4b09-a232-7f4a02322ad7', 'attached_at': '', 'detached_at': '', 'volume_id': 'f908034d-f063-4b8d-9877-08531afe6e62', 'serial': 'f908034d-f063-4b8d-9877-08531afe6e62'} {{(pid=61923) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1095.852383] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5be1d05-28e5-4a5f-9581-54476d30f9d8 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.876239] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-740176f4-cab3-4fb3-80b8-7621668aa429 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.883289] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92088c03-8158-4ff3-8d98-262201162800 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.905350] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-456aa489-1c14-4bca-8ea5-fa7504228e8f {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.919006] env[61923]: DEBUG nova.virt.vmwareapi.volumeops [None req-9155cecc-8b54-46b6-9db8-ad98ae309cb1 tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] The volume has not been displaced from its original location: [datastore2] volume-f908034d-f063-4b8d-9877-08531afe6e62/volume-f908034d-f063-4b8d-9877-08531afe6e62.vmdk. No consolidation needed. {{(pid=61923) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1095.924128] env[61923]: DEBUG nova.virt.vmwareapi.volumeops [None req-9155cecc-8b54-46b6-9db8-ad98ae309cb1 tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] [instance: 35bdbd45-5838-4b09-a232-7f4a02322ad7] Reconfiguring VM instance instance-00000060 to detach disk 2001 {{(pid=61923) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1095.924606] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-106d4371-cbb4-485a-949d-afd3d54169f6 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.944096] env[61923]: DEBUG oslo_vmware.api [None req-affadfdd-ba30-4ce8-8d78-e18297fead5a tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Task: {'id': task-1378026, 'name': ReconfigVM_Task, 'duration_secs': 0.261644} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1095.945201] env[61923]: DEBUG nova.virt.vmwareapi.volumeops [None req-affadfdd-ba30-4ce8-8d78-e18297fead5a tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: c0a8a116-c7b6-4adc-849b-e0a4faa150d3] Reconfigured VM instance instance-00000065 to attach disk [datastore2] c0a8a116-c7b6-4adc-849b-e0a4faa150d3/c0a8a116-c7b6-4adc-849b-e0a4faa150d3.vmdk or device None with type sparse {{(pid=61923) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1095.945944] env[61923]: DEBUG oslo_vmware.api [None req-9155cecc-8b54-46b6-9db8-ad98ae309cb1 tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Waiting for the task: (returnval){ [ 1095.945944] env[61923]: value = "task-1378027" [ 1095.945944] env[61923]: _type = "Task" [ 1095.945944] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1095.946151] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-46a6233b-43b6-476b-9044-b6502a08abb1 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.956288] env[61923]: DEBUG oslo_vmware.api [None req-9155cecc-8b54-46b6-9db8-ad98ae309cb1 tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Task: {'id': task-1378027, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1095.957562] env[61923]: DEBUG oslo_vmware.api [None req-affadfdd-ba30-4ce8-8d78-e18297fead5a tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Waiting for the task: (returnval){ [ 1095.957562] env[61923]: value = "task-1378028" [ 1095.957562] env[61923]: _type = "Task" [ 1095.957562] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1095.964145] env[61923]: DEBUG oslo_vmware.api [None req-affadfdd-ba30-4ce8-8d78-e18297fead5a tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Task: {'id': task-1378028, 'name': Rename_Task} progress is 5%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1096.209571] env[61923]: DEBUG nova.objects.base [None req-781874ef-b4d2-4025-bf94-9b496335d5ac tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Object Instance lazy-loaded attributes: info_cache,migration_context {{(pid=61923) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 1096.210510] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b55e5a20-c5de-479b-bdb7-580465923646 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.229014] env[61923]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9123740c-eea1-4fd1-8ecb-a079872a0cb6 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.234337] env[61923]: DEBUG oslo_vmware.api [None req-781874ef-b4d2-4025-bf94-9b496335d5ac tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Waiting for the task: (returnval){ [ 1096.234337] env[61923]: value = "session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52caf272-5fc1-b807-d106-891b31fa169a" [ 1096.234337] env[61923]: _type = "Task" [ 1096.234337] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1096.241751] env[61923]: DEBUG oslo_vmware.api [None req-781874ef-b4d2-4025-bf94-9b496335d5ac tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52caf272-5fc1-b807-d106-891b31fa169a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1096.341553] env[61923]: DEBUG nova.compute.manager [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] [instance: 1d1b1dc4-c701-42b2-beba-344d622aef41] Instance has had 0 of 5 cleanup attempts {{(pid=61923) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1096.457158] env[61923]: DEBUG oslo_vmware.api [None req-9155cecc-8b54-46b6-9db8-ad98ae309cb1 tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Task: {'id': task-1378027, 'name': ReconfigVM_Task, 'duration_secs': 0.392121} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1096.457455] env[61923]: DEBUG nova.virt.vmwareapi.volumeops [None req-9155cecc-8b54-46b6-9db8-ad98ae309cb1 tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] [instance: 35bdbd45-5838-4b09-a232-7f4a02322ad7] Reconfigured VM instance instance-00000060 to detach disk 2001 {{(pid=61923) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1096.462054] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-87280af7-2e98-463d-ac3c-4aae923fc83c {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.480127] env[61923]: DEBUG oslo_vmware.api [None req-affadfdd-ba30-4ce8-8d78-e18297fead5a tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Task: {'id': task-1378028, 'name': Rename_Task, 'duration_secs': 0.130195} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1096.481253] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-affadfdd-ba30-4ce8-8d78-e18297fead5a tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: c0a8a116-c7b6-4adc-849b-e0a4faa150d3] Powering on the VM {{(pid=61923) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1096.481561] env[61923]: DEBUG oslo_vmware.api [None req-9155cecc-8b54-46b6-9db8-ad98ae309cb1 tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Waiting for the task: (returnval){ [ 1096.481561] env[61923]: value = "task-1378029" [ 1096.481561] env[61923]: _type = "Task" [ 1096.481561] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1096.481845] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b5f11d58-b15a-40fd-aa51-06d9218c10f6 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.491594] env[61923]: DEBUG oslo_vmware.api [None req-9155cecc-8b54-46b6-9db8-ad98ae309cb1 tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Task: {'id': task-1378029, 'name': ReconfigVM_Task} progress is 10%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1096.491864] env[61923]: DEBUG oslo_vmware.api [None req-affadfdd-ba30-4ce8-8d78-e18297fead5a tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Waiting for the task: (returnval){ [ 1096.491864] env[61923]: value = "task-1378030" [ 1096.491864] env[61923]: _type = "Task" [ 1096.491864] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1096.499181] env[61923]: DEBUG oslo_vmware.api [None req-affadfdd-ba30-4ce8-8d78-e18297fead5a tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Task: {'id': task-1378030, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1096.744757] env[61923]: DEBUG oslo_vmware.api [None req-781874ef-b4d2-4025-bf94-9b496335d5ac tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52caf272-5fc1-b807-d106-891b31fa169a, 'name': SearchDatastore_Task, 'duration_secs': 0.009027} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1096.745053] env[61923]: DEBUG oslo_concurrency.lockutils [None req-781874ef-b4d2-4025-bf94-9b496335d5ac tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1096.745316] env[61923]: DEBUG oslo_concurrency.lockutils [None req-781874ef-b4d2-4025-bf94-9b496335d5ac tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1096.844818] env[61923]: DEBUG nova.compute.manager [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] [instance: b8c60af7-3561-468c-a9a5-1c645080fc69] Instance has had 0 of 5 cleanup attempts {{(pid=61923) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1096.992595] env[61923]: DEBUG oslo_vmware.api [None req-9155cecc-8b54-46b6-9db8-ad98ae309cb1 tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Task: {'id': task-1378029, 'name': ReconfigVM_Task, 'duration_secs': 0.143584} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1096.995537] env[61923]: DEBUG nova.virt.vmwareapi.volumeops [None req-9155cecc-8b54-46b6-9db8-ad98ae309cb1 tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] [instance: 35bdbd45-5838-4b09-a232-7f4a02322ad7] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-292753', 'volume_id': 'f908034d-f063-4b8d-9877-08531afe6e62', 'name': 'volume-f908034d-f063-4b8d-9877-08531afe6e62', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '35bdbd45-5838-4b09-a232-7f4a02322ad7', 'attached_at': '', 'detached_at': '', 'volume_id': 'f908034d-f063-4b8d-9877-08531afe6e62', 'serial': 'f908034d-f063-4b8d-9877-08531afe6e62'} {{(pid=61923) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1097.001917] env[61923]: DEBUG oslo_vmware.api [None req-affadfdd-ba30-4ce8-8d78-e18297fead5a tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Task: {'id': task-1378030, 'name': PowerOnVM_Task, 'duration_secs': 0.423532} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1097.002202] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-affadfdd-ba30-4ce8-8d78-e18297fead5a tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: c0a8a116-c7b6-4adc-849b-e0a4faa150d3] Powered on the VM {{(pid=61923) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1097.002414] env[61923]: INFO nova.compute.manager [None req-affadfdd-ba30-4ce8-8d78-e18297fead5a tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: c0a8a116-c7b6-4adc-849b-e0a4faa150d3] Took 6.86 seconds to spawn the instance on the hypervisor. [ 1097.002665] env[61923]: DEBUG nova.compute.manager [None req-affadfdd-ba30-4ce8-8d78-e18297fead5a tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: c0a8a116-c7b6-4adc-849b-e0a4faa150d3] Checking state {{(pid=61923) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1097.003619] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29d6ff93-9dbb-4d6e-a7a5-db5f8683b21e {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.308465] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0f515f2-028f-42e2-a788-625695ae3371 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.317640] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7201e24-d1a3-44eb-8380-ad79f0b851db {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.346514] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3fb98f40-fc6f-42ad-a54d-4105bb229a7d {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.349020] env[61923]: DEBUG nova.compute.manager [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] [instance: ecedb77f-4a2a-4047-964f-80796954d847] Instance has had 0 of 5 cleanup attempts {{(pid=61923) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1097.356044] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a993eb0e-b636-4a24-be0a-07b30c8580a2 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.368626] env[61923]: DEBUG nova.compute.provider_tree [None req-781874ef-b4d2-4025-bf94-9b496335d5ac tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1097.519156] env[61923]: INFO nova.compute.manager [None req-affadfdd-ba30-4ce8-8d78-e18297fead5a tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: c0a8a116-c7b6-4adc-849b-e0a4faa150d3] Took 11.54 seconds to build instance. [ 1097.537227] env[61923]: DEBUG nova.objects.instance [None req-9155cecc-8b54-46b6-9db8-ad98ae309cb1 tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Lazy-loading 'flavor' on Instance uuid 35bdbd45-5838-4b09-a232-7f4a02322ad7 {{(pid=61923) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1097.852682] env[61923]: DEBUG nova.compute.manager [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] [instance: 655b01ae-76a5-4c09-9245-e629fec8fd4b] Instance has had 0 of 5 cleanup attempts {{(pid=61923) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1097.872034] env[61923]: DEBUG nova.scheduler.client.report [None req-781874ef-b4d2-4025-bf94-9b496335d5ac tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1098.021558] env[61923]: DEBUG oslo_concurrency.lockutils [None req-affadfdd-ba30-4ce8-8d78-e18297fead5a tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Lock "c0a8a116-c7b6-4adc-849b-e0a4faa150d3" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 13.054s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1098.543861] env[61923]: DEBUG oslo_concurrency.lockutils [None req-9155cecc-8b54-46b6-9db8-ad98ae309cb1 tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Lock "35bdbd45-5838-4b09-a232-7f4a02322ad7" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.227s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1098.586605] env[61923]: DEBUG oslo_concurrency.lockutils [None req-59f96632-fae4-4988-bf79-6cb89cabbbef tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Acquiring lock "35bdbd45-5838-4b09-a232-7f4a02322ad7" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1098.586868] env[61923]: DEBUG oslo_concurrency.lockutils [None req-59f96632-fae4-4988-bf79-6cb89cabbbef tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Lock "35bdbd45-5838-4b09-a232-7f4a02322ad7" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1098.882644] env[61923]: DEBUG oslo_concurrency.lockutils [None req-781874ef-b4d2-4025-bf94-9b496335d5ac tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.137s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1098.946878] env[61923]: DEBUG nova.compute.manager [None req-c54f3d94-4858-42ec-a33f-5691c52b7589 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: c0a8a116-c7b6-4adc-849b-e0a4faa150d3] Stashing vm_state: active {{(pid=61923) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 1099.089622] env[61923]: INFO nova.compute.manager [None req-59f96632-fae4-4988-bf79-6cb89cabbbef tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] [instance: 35bdbd45-5838-4b09-a232-7f4a02322ad7] Detaching volume 8744225b-4483-40f9-b314-ebab57d963bc [ 1099.126256] env[61923]: INFO nova.virt.block_device [None req-59f96632-fae4-4988-bf79-6cb89cabbbef tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] [instance: 35bdbd45-5838-4b09-a232-7f4a02322ad7] Attempting to driver detach volume 8744225b-4483-40f9-b314-ebab57d963bc from mountpoint /dev/sdc [ 1099.126610] env[61923]: DEBUG nova.virt.vmwareapi.volumeops [None req-59f96632-fae4-4988-bf79-6cb89cabbbef tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] [instance: 35bdbd45-5838-4b09-a232-7f4a02322ad7] Volume detach. Driver type: vmdk {{(pid=61923) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1099.126896] env[61923]: DEBUG nova.virt.vmwareapi.volumeops [None req-59f96632-fae4-4988-bf79-6cb89cabbbef tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] [instance: 35bdbd45-5838-4b09-a232-7f4a02322ad7] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-292754', 'volume_id': '8744225b-4483-40f9-b314-ebab57d963bc', 'name': 'volume-8744225b-4483-40f9-b314-ebab57d963bc', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '35bdbd45-5838-4b09-a232-7f4a02322ad7', 'attached_at': '', 'detached_at': '', 'volume_id': '8744225b-4483-40f9-b314-ebab57d963bc', 'serial': '8744225b-4483-40f9-b314-ebab57d963bc'} {{(pid=61923) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1099.128192] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-301ab315-5d28-4b16-882a-8a9a91980f15 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.158836] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf37dc9d-1c71-4158-85a5-5390cf8cf199 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.167560] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15b56dd5-8e51-479b-a05b-0be650a0e5d7 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.196416] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82faf416-8a0e-47af-94c1-466dc38ad5d8 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.213480] env[61923]: DEBUG nova.virt.vmwareapi.volumeops [None req-59f96632-fae4-4988-bf79-6cb89cabbbef tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] The volume has not been displaced from its original location: [datastore2] volume-8744225b-4483-40f9-b314-ebab57d963bc/volume-8744225b-4483-40f9-b314-ebab57d963bc.vmdk. No consolidation needed. {{(pid=61923) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1099.218757] env[61923]: DEBUG nova.virt.vmwareapi.volumeops [None req-59f96632-fae4-4988-bf79-6cb89cabbbef tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] [instance: 35bdbd45-5838-4b09-a232-7f4a02322ad7] Reconfiguring VM instance instance-00000060 to detach disk 2002 {{(pid=61923) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1099.219042] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-41865a26-c730-4a54-8547-a62779000f26 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.236304] env[61923]: DEBUG oslo_vmware.api [None req-59f96632-fae4-4988-bf79-6cb89cabbbef tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Waiting for the task: (returnval){ [ 1099.236304] env[61923]: value = "task-1378031" [ 1099.236304] env[61923]: _type = "Task" [ 1099.236304] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1099.243775] env[61923]: DEBUG oslo_vmware.api [None req-59f96632-fae4-4988-bf79-6cb89cabbbef tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Task: {'id': task-1378031, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1099.439926] env[61923]: INFO nova.scheduler.client.report [None req-781874ef-b4d2-4025-bf94-9b496335d5ac tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Deleted allocation for migration 2c6aa9f7-6c7b-47e6-8e98-c1574b78f4a1 [ 1099.462959] env[61923]: DEBUG oslo_concurrency.lockutils [None req-c54f3d94-4858-42ec-a33f-5691c52b7589 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1099.463291] env[61923]: DEBUG oslo_concurrency.lockutils [None req-c54f3d94-4858-42ec-a33f-5691c52b7589 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1099.746197] env[61923]: DEBUG oslo_vmware.api [None req-59f96632-fae4-4988-bf79-6cb89cabbbef tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Task: {'id': task-1378031, 'name': ReconfigVM_Task, 'duration_secs': 0.237144} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1099.746572] env[61923]: DEBUG nova.virt.vmwareapi.volumeops [None req-59f96632-fae4-4988-bf79-6cb89cabbbef tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] [instance: 35bdbd45-5838-4b09-a232-7f4a02322ad7] Reconfigured VM instance instance-00000060 to detach disk 2002 {{(pid=61923) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1099.751240] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0af0a0d9-3a52-41a9-9313-71bd147f49d5 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.765502] env[61923]: DEBUG oslo_vmware.api [None req-59f96632-fae4-4988-bf79-6cb89cabbbef tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Waiting for the task: (returnval){ [ 1099.765502] env[61923]: value = "task-1378032" [ 1099.765502] env[61923]: _type = "Task" [ 1099.765502] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1099.773605] env[61923]: DEBUG oslo_vmware.api [None req-59f96632-fae4-4988-bf79-6cb89cabbbef tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Task: {'id': task-1378032, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1099.945886] env[61923]: DEBUG oslo_concurrency.lockutils [None req-781874ef-b4d2-4025-bf94-9b496335d5ac tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Lock "ab3edc53-37a5-4119-84c5-fd408ed66b69" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 6.488s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1099.967645] env[61923]: INFO nova.compute.claims [None req-c54f3d94-4858-42ec-a33f-5691c52b7589 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: c0a8a116-c7b6-4adc-849b-e0a4faa150d3] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1100.275607] env[61923]: DEBUG oslo_vmware.api [None req-59f96632-fae4-4988-bf79-6cb89cabbbef tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Task: {'id': task-1378032, 'name': ReconfigVM_Task, 'duration_secs': 0.129138} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1100.275943] env[61923]: DEBUG nova.virt.vmwareapi.volumeops [None req-59f96632-fae4-4988-bf79-6cb89cabbbef tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] [instance: 35bdbd45-5838-4b09-a232-7f4a02322ad7] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-292754', 'volume_id': '8744225b-4483-40f9-b314-ebab57d963bc', 'name': 'volume-8744225b-4483-40f9-b314-ebab57d963bc', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '35bdbd45-5838-4b09-a232-7f4a02322ad7', 'attached_at': '', 'detached_at': '', 'volume_id': '8744225b-4483-40f9-b314-ebab57d963bc', 'serial': '8744225b-4483-40f9-b314-ebab57d963bc'} {{(pid=61923) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1100.473297] env[61923]: INFO nova.compute.resource_tracker [None req-c54f3d94-4858-42ec-a33f-5691c52b7589 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: c0a8a116-c7b6-4adc-849b-e0a4faa150d3] Updating resource usage from migration af6ff10e-1948-46aa-b5c5-bbe410bf3124 [ 1100.533687] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2826ca2-0183-4287-a68e-24ce7cfd38ac {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.541394] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2cac420b-549b-4f93-8650-e2b4cc5f7188 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.571509] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3265fa47-c870-41f9-8994-ec8df47329cb {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.578617] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-086e6be3-a5c8-46af-8580-8fbd008c1d33 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1100.591611] env[61923]: DEBUG nova.compute.provider_tree [None req-c54f3d94-4858-42ec-a33f-5691c52b7589 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1100.820121] env[61923]: DEBUG nova.objects.instance [None req-59f96632-fae4-4988-bf79-6cb89cabbbef tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Lazy-loading 'flavor' on Instance uuid 35bdbd45-5838-4b09-a232-7f4a02322ad7 {{(pid=61923) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1101.095293] env[61923]: DEBUG nova.scheduler.client.report [None req-c54f3d94-4858-42ec-a33f-5691c52b7589 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1101.222510] env[61923]: DEBUG oslo_concurrency.lockutils [None req-09014bcf-7023-44df-a13d-2100c4196634 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Acquiring lock "ab3edc53-37a5-4119-84c5-fd408ed66b69" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1101.222849] env[61923]: DEBUG oslo_concurrency.lockutils [None req-09014bcf-7023-44df-a13d-2100c4196634 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Lock "ab3edc53-37a5-4119-84c5-fd408ed66b69" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1101.223096] env[61923]: DEBUG oslo_concurrency.lockutils [None req-09014bcf-7023-44df-a13d-2100c4196634 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Acquiring lock "ab3edc53-37a5-4119-84c5-fd408ed66b69-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1101.223304] env[61923]: DEBUG oslo_concurrency.lockutils [None req-09014bcf-7023-44df-a13d-2100c4196634 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Lock "ab3edc53-37a5-4119-84c5-fd408ed66b69-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1101.223481] env[61923]: DEBUG oslo_concurrency.lockutils [None req-09014bcf-7023-44df-a13d-2100c4196634 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Lock "ab3edc53-37a5-4119-84c5-fd408ed66b69-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1101.227057] env[61923]: INFO nova.compute.manager [None req-09014bcf-7023-44df-a13d-2100c4196634 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: ab3edc53-37a5-4119-84c5-fd408ed66b69] Terminating instance [ 1101.227897] env[61923]: DEBUG nova.compute.manager [None req-09014bcf-7023-44df-a13d-2100c4196634 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: ab3edc53-37a5-4119-84c5-fd408ed66b69] Start destroying the instance on the hypervisor. {{(pid=61923) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1101.228105] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-09014bcf-7023-44df-a13d-2100c4196634 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: ab3edc53-37a5-4119-84c5-fd408ed66b69] Destroying instance {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1101.228996] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8dcc42f7-761d-4fce-b0e3-337b304f8df0 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.236773] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-09014bcf-7023-44df-a13d-2100c4196634 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: ab3edc53-37a5-4119-84c5-fd408ed66b69] Powering off the VM {{(pid=61923) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1101.237313] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-05a33df8-5326-48f9-a0eb-23646a517410 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.244156] env[61923]: DEBUG oslo_vmware.api [None req-09014bcf-7023-44df-a13d-2100c4196634 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Waiting for the task: (returnval){ [ 1101.244156] env[61923]: value = "task-1378033" [ 1101.244156] env[61923]: _type = "Task" [ 1101.244156] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1101.251932] env[61923]: DEBUG oslo_vmware.api [None req-09014bcf-7023-44df-a13d-2100c4196634 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Task: {'id': task-1378033, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1101.599777] env[61923]: DEBUG oslo_concurrency.lockutils [None req-c54f3d94-4858-42ec-a33f-5691c52b7589 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.136s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1101.600066] env[61923]: INFO nova.compute.manager [None req-c54f3d94-4858-42ec-a33f-5691c52b7589 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: c0a8a116-c7b6-4adc-849b-e0a4faa150d3] Migrating [ 1101.635515] env[61923]: DEBUG oslo_concurrency.lockutils [None req-22e03437-b9fc-405c-b5bb-a6bb5bdc3129 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Acquiring lock "6fddc45c-d589-4423-a990-4be24d9dc84e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1101.635753] env[61923]: DEBUG oslo_concurrency.lockutils [None req-22e03437-b9fc-405c-b5bb-a6bb5bdc3129 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Lock "6fddc45c-d589-4423-a990-4be24d9dc84e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1101.755184] env[61923]: DEBUG oslo_vmware.api [None req-09014bcf-7023-44df-a13d-2100c4196634 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Task: {'id': task-1378033, 'name': PowerOffVM_Task, 'duration_secs': 0.183655} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1101.755479] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-09014bcf-7023-44df-a13d-2100c4196634 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: ab3edc53-37a5-4119-84c5-fd408ed66b69] Powered off the VM {{(pid=61923) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1101.755656] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-09014bcf-7023-44df-a13d-2100c4196634 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: ab3edc53-37a5-4119-84c5-fd408ed66b69] Unregistering the VM {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1101.755901] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-ef16a6cd-9886-42ea-9529-eba27bcd765a {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.828236] env[61923]: DEBUG oslo_concurrency.lockutils [None req-59f96632-fae4-4988-bf79-6cb89cabbbef tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Lock "35bdbd45-5838-4b09-a232-7f4a02322ad7" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.241s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1101.830840] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-09014bcf-7023-44df-a13d-2100c4196634 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: ab3edc53-37a5-4119-84c5-fd408ed66b69] Unregistered the VM {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1101.831280] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-09014bcf-7023-44df-a13d-2100c4196634 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: ab3edc53-37a5-4119-84c5-fd408ed66b69] Deleting contents of the VM from datastore datastore1 {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1101.831508] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-09014bcf-7023-44df-a13d-2100c4196634 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Deleting the datastore file [datastore1] ab3edc53-37a5-4119-84c5-fd408ed66b69 {{(pid=61923) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1101.832031] env[61923]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-570630f3-ee5c-48f2-8b43-65a88007b024 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.839738] env[61923]: DEBUG oslo_vmware.api [None req-09014bcf-7023-44df-a13d-2100c4196634 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Waiting for the task: (returnval){ [ 1101.839738] env[61923]: value = "task-1378035" [ 1101.839738] env[61923]: _type = "Task" [ 1101.839738] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1101.848868] env[61923]: DEBUG oslo_vmware.api [None req-09014bcf-7023-44df-a13d-2100c4196634 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Task: {'id': task-1378035, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1102.115915] env[61923]: DEBUG oslo_concurrency.lockutils [None req-c54f3d94-4858-42ec-a33f-5691c52b7589 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Acquiring lock "refresh_cache-c0a8a116-c7b6-4adc-849b-e0a4faa150d3" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1102.116236] env[61923]: DEBUG oslo_concurrency.lockutils [None req-c54f3d94-4858-42ec-a33f-5691c52b7589 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Acquired lock "refresh_cache-c0a8a116-c7b6-4adc-849b-e0a4faa150d3" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1102.116333] env[61923]: DEBUG nova.network.neutron [None req-c54f3d94-4858-42ec-a33f-5691c52b7589 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: c0a8a116-c7b6-4adc-849b-e0a4faa150d3] Building network info cache for instance {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1102.138104] env[61923]: DEBUG nova.compute.manager [None req-22e03437-b9fc-405c-b5bb-a6bb5bdc3129 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: 6fddc45c-d589-4423-a990-4be24d9dc84e] Starting instance... {{(pid=61923) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1102.350066] env[61923]: DEBUG oslo_vmware.api [None req-09014bcf-7023-44df-a13d-2100c4196634 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Task: {'id': task-1378035, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.15487} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1102.350349] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-09014bcf-7023-44df-a13d-2100c4196634 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Deleted the datastore file {{(pid=61923) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1102.350538] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-09014bcf-7023-44df-a13d-2100c4196634 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: ab3edc53-37a5-4119-84c5-fd408ed66b69] Deleted contents of the VM from datastore datastore1 {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1102.350725] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-09014bcf-7023-44df-a13d-2100c4196634 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: ab3edc53-37a5-4119-84c5-fd408ed66b69] Instance destroyed {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1102.350909] env[61923]: INFO nova.compute.manager [None req-09014bcf-7023-44df-a13d-2100c4196634 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: ab3edc53-37a5-4119-84c5-fd408ed66b69] Took 1.12 seconds to destroy the instance on the hypervisor. [ 1102.351162] env[61923]: DEBUG oslo.service.loopingcall [None req-09014bcf-7023-44df-a13d-2100c4196634 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61923) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1102.351518] env[61923]: DEBUG oslo_service.periodic_task [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61923) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1102.351801] env[61923]: DEBUG nova.compute.manager [-] [instance: ab3edc53-37a5-4119-84c5-fd408ed66b69] Deallocating network for instance {{(pid=61923) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1102.351897] env[61923]: DEBUG nova.network.neutron [-] [instance: ab3edc53-37a5-4119-84c5-fd408ed66b69] deallocate_for_instance() {{(pid=61923) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1102.353539] env[61923]: DEBUG oslo_service.periodic_task [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=61923) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1102.664527] env[61923]: DEBUG oslo_concurrency.lockutils [None req-22e03437-b9fc-405c-b5bb-a6bb5bdc3129 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1102.664625] env[61923]: DEBUG oslo_concurrency.lockutils [None req-22e03437-b9fc-405c-b5bb-a6bb5bdc3129 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1102.666828] env[61923]: INFO nova.compute.claims [None req-22e03437-b9fc-405c-b5bb-a6bb5bdc3129 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: 6fddc45c-d589-4423-a990-4be24d9dc84e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1102.670859] env[61923]: DEBUG nova.compute.manager [req-bc1e082f-053c-419e-b8fe-bc010bffcc9f req-0f1c3195-9643-4610-b95f-f9c176faef5d service nova] [instance: ab3edc53-37a5-4119-84c5-fd408ed66b69] Received event network-vif-deleted-f2c777e2-f8f4-437b-b990-20e42dd2491b {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1102.670859] env[61923]: INFO nova.compute.manager [req-bc1e082f-053c-419e-b8fe-bc010bffcc9f req-0f1c3195-9643-4610-b95f-f9c176faef5d service nova] [instance: ab3edc53-37a5-4119-84c5-fd408ed66b69] Neutron deleted interface f2c777e2-f8f4-437b-b990-20e42dd2491b; detaching it from the instance and deleting it from the info cache [ 1102.670859] env[61923]: DEBUG nova.network.neutron [req-bc1e082f-053c-419e-b8fe-bc010bffcc9f req-0f1c3195-9643-4610-b95f-f9c176faef5d service nova] [instance: ab3edc53-37a5-4119-84c5-fd408ed66b69] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1102.807359] env[61923]: DEBUG nova.network.neutron [None req-c54f3d94-4858-42ec-a33f-5691c52b7589 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: c0a8a116-c7b6-4adc-849b-e0a4faa150d3] Updating instance_info_cache with network_info: [{"id": "c210c30f-4f3c-4698-ab24-a6a81c78e658", "address": "fa:16:3e:ad:1c:c4", "network": {"id": "9533e79d-fde2-4c10-86a0-86a36845a8cf", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1736743137-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "780d3541d9604417b977bb62390c4299", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6d650b26-c3e7-4de7-98db-5e4b816d123a", "external-id": "nsx-vlan-transportzone-757", "segmentation_id": 757, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc210c30f-4f", "ovs_interfaceid": "c210c30f-4f3c-4698-ab24-a6a81c78e658", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1102.858472] env[61923]: DEBUG oslo_service.periodic_task [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Running periodic task ComputeManager._sync_power_states {{(pid=61923) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1102.977842] env[61923]: DEBUG oslo_concurrency.lockutils [None req-048a195a-ccb9-4ef0-aba8-3142327ae2bc tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Acquiring lock "35bdbd45-5838-4b09-a232-7f4a02322ad7" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1102.978209] env[61923]: DEBUG oslo_concurrency.lockutils [None req-048a195a-ccb9-4ef0-aba8-3142327ae2bc tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Lock "35bdbd45-5838-4b09-a232-7f4a02322ad7" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1102.978442] env[61923]: DEBUG oslo_concurrency.lockutils [None req-048a195a-ccb9-4ef0-aba8-3142327ae2bc tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Acquiring lock "35bdbd45-5838-4b09-a232-7f4a02322ad7-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1102.978644] env[61923]: DEBUG oslo_concurrency.lockutils [None req-048a195a-ccb9-4ef0-aba8-3142327ae2bc tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Lock "35bdbd45-5838-4b09-a232-7f4a02322ad7-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1102.978846] env[61923]: DEBUG oslo_concurrency.lockutils [None req-048a195a-ccb9-4ef0-aba8-3142327ae2bc tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Lock "35bdbd45-5838-4b09-a232-7f4a02322ad7-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1102.980947] env[61923]: INFO nova.compute.manager [None req-048a195a-ccb9-4ef0-aba8-3142327ae2bc tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] [instance: 35bdbd45-5838-4b09-a232-7f4a02322ad7] Terminating instance [ 1102.982732] env[61923]: DEBUG nova.compute.manager [None req-048a195a-ccb9-4ef0-aba8-3142327ae2bc tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] [instance: 35bdbd45-5838-4b09-a232-7f4a02322ad7] Start destroying the instance on the hypervisor. {{(pid=61923) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1102.982924] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-048a195a-ccb9-4ef0-aba8-3142327ae2bc tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] [instance: 35bdbd45-5838-4b09-a232-7f4a02322ad7] Destroying instance {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1102.983758] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ab423cd-7feb-4522-abd1-0d1934f98364 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1102.992373] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-048a195a-ccb9-4ef0-aba8-3142327ae2bc tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] [instance: 35bdbd45-5838-4b09-a232-7f4a02322ad7] Powering off the VM {{(pid=61923) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1102.992599] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-fac36af0-a9b5-4d59-bbd7-3bc467eb0fd7 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.000953] env[61923]: DEBUG oslo_vmware.api [None req-048a195a-ccb9-4ef0-aba8-3142327ae2bc tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Waiting for the task: (returnval){ [ 1103.000953] env[61923]: value = "task-1378036" [ 1103.000953] env[61923]: _type = "Task" [ 1103.000953] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1103.008597] env[61923]: DEBUG oslo_vmware.api [None req-048a195a-ccb9-4ef0-aba8-3142327ae2bc tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Task: {'id': task-1378036, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1103.146815] env[61923]: DEBUG nova.network.neutron [-] [instance: ab3edc53-37a5-4119-84c5-fd408ed66b69] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1103.174858] env[61923]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-602d34d1-70f0-4e54-9b18-280b51b47d30 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.183952] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ddab6793-7e09-4f1e-9053-bdbda078ce25 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.209149] env[61923]: DEBUG nova.compute.manager [req-bc1e082f-053c-419e-b8fe-bc010bffcc9f req-0f1c3195-9643-4610-b95f-f9c176faef5d service nova] [instance: ab3edc53-37a5-4119-84c5-fd408ed66b69] Detach interface failed, port_id=f2c777e2-f8f4-437b-b990-20e42dd2491b, reason: Instance ab3edc53-37a5-4119-84c5-fd408ed66b69 could not be found. {{(pid=61923) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1103.310101] env[61923]: DEBUG oslo_concurrency.lockutils [None req-c54f3d94-4858-42ec-a33f-5691c52b7589 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Releasing lock "refresh_cache-c0a8a116-c7b6-4adc-849b-e0a4faa150d3" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1103.361468] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Getting list of instances from cluster (obj){ [ 1103.361468] env[61923]: value = "domain-c8" [ 1103.361468] env[61923]: _type = "ClusterComputeResource" [ 1103.361468] env[61923]: } {{(pid=61923) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2116}} [ 1103.362547] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8199572-9833-432e-a844-5413bdeffc6f {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.374643] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Got total of 2 instances {{(pid=61923) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2125}} [ 1103.374820] env[61923]: WARNING nova.compute.manager [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] While synchronizing instance power states, found 4 instances in the database and 2 instances on the hypervisor. [ 1103.374961] env[61923]: DEBUG nova.compute.manager [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Triggering sync for uuid 35bdbd45-5838-4b09-a232-7f4a02322ad7 {{(pid=61923) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10338}} [ 1103.375170] env[61923]: DEBUG nova.compute.manager [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Triggering sync for uuid ab3edc53-37a5-4119-84c5-fd408ed66b69 {{(pid=61923) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10338}} [ 1103.375357] env[61923]: DEBUG nova.compute.manager [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Triggering sync for uuid c0a8a116-c7b6-4adc-849b-e0a4faa150d3 {{(pid=61923) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10338}} [ 1103.375481] env[61923]: DEBUG nova.compute.manager [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Triggering sync for uuid 6fddc45c-d589-4423-a990-4be24d9dc84e {{(pid=61923) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10338}} [ 1103.375788] env[61923]: DEBUG oslo_concurrency.lockutils [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Acquiring lock "35bdbd45-5838-4b09-a232-7f4a02322ad7" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1103.376037] env[61923]: DEBUG oslo_concurrency.lockutils [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Acquiring lock "ab3edc53-37a5-4119-84c5-fd408ed66b69" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1103.376242] env[61923]: DEBUG oslo_concurrency.lockutils [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Acquiring lock "c0a8a116-c7b6-4adc-849b-e0a4faa150d3" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1103.376434] env[61923]: DEBUG oslo_concurrency.lockutils [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Lock "c0a8a116-c7b6-4adc-849b-e0a4faa150d3" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1103.376654] env[61923]: INFO nova.compute.manager [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] [instance: c0a8a116-c7b6-4adc-849b-e0a4faa150d3] During sync_power_state the instance has a pending task (resize_prep). Skip. [ 1103.376833] env[61923]: DEBUG oslo_concurrency.lockutils [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Lock "c0a8a116-c7b6-4adc-849b-e0a4faa150d3" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1103.377039] env[61923]: DEBUG oslo_concurrency.lockutils [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Acquiring lock "6fddc45c-d589-4423-a990-4be24d9dc84e" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1103.510784] env[61923]: DEBUG oslo_vmware.api [None req-048a195a-ccb9-4ef0-aba8-3142327ae2bc tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Task: {'id': task-1378036, 'name': PowerOffVM_Task, 'duration_secs': 0.215467} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1103.511079] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-048a195a-ccb9-4ef0-aba8-3142327ae2bc tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] [instance: 35bdbd45-5838-4b09-a232-7f4a02322ad7] Powered off the VM {{(pid=61923) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1103.511266] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-048a195a-ccb9-4ef0-aba8-3142327ae2bc tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] [instance: 35bdbd45-5838-4b09-a232-7f4a02322ad7] Unregistering the VM {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1103.511554] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-ca4be5fa-c3f8-486b-b4c9-9a101fb98367 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.573728] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-048a195a-ccb9-4ef0-aba8-3142327ae2bc tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] [instance: 35bdbd45-5838-4b09-a232-7f4a02322ad7] Unregistered the VM {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1103.573990] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-048a195a-ccb9-4ef0-aba8-3142327ae2bc tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] [instance: 35bdbd45-5838-4b09-a232-7f4a02322ad7] Deleting contents of the VM from datastore datastore2 {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1103.574197] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-048a195a-ccb9-4ef0-aba8-3142327ae2bc tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Deleting the datastore file [datastore2] 35bdbd45-5838-4b09-a232-7f4a02322ad7 {{(pid=61923) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1103.574472] env[61923]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-eef4094d-fdcf-468a-98b0-313cf40f5310 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.580637] env[61923]: DEBUG oslo_vmware.api [None req-048a195a-ccb9-4ef0-aba8-3142327ae2bc tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Waiting for the task: (returnval){ [ 1103.580637] env[61923]: value = "task-1378038" [ 1103.580637] env[61923]: _type = "Task" [ 1103.580637] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1103.588338] env[61923]: DEBUG oslo_vmware.api [None req-048a195a-ccb9-4ef0-aba8-3142327ae2bc tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Task: {'id': task-1378038, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1103.650486] env[61923]: INFO nova.compute.manager [-] [instance: ab3edc53-37a5-4119-84c5-fd408ed66b69] Took 1.30 seconds to deallocate network for instance. [ 1103.748364] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-808fa522-34dd-4266-baee-d9a8521677f4 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.754061] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c310c0fb-dcfe-4d8e-92e8-80eaa7a85b8e {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.784296] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-abe97dcb-2df8-447f-acee-0b0143012486 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.791110] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ba32f81-82b7-4a96-b8fb-2744ad770cfc {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.803972] env[61923]: DEBUG nova.compute.provider_tree [None req-22e03437-b9fc-405c-b5bb-a6bb5bdc3129 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1104.090601] env[61923]: DEBUG oslo_vmware.api [None req-048a195a-ccb9-4ef0-aba8-3142327ae2bc tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Task: {'id': task-1378038, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.132353} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1104.091156] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-048a195a-ccb9-4ef0-aba8-3142327ae2bc tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Deleted the datastore file {{(pid=61923) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1104.091512] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-048a195a-ccb9-4ef0-aba8-3142327ae2bc tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] [instance: 35bdbd45-5838-4b09-a232-7f4a02322ad7] Deleted contents of the VM from datastore datastore2 {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1104.091851] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-048a195a-ccb9-4ef0-aba8-3142327ae2bc tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] [instance: 35bdbd45-5838-4b09-a232-7f4a02322ad7] Instance destroyed {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1104.092175] env[61923]: INFO nova.compute.manager [None req-048a195a-ccb9-4ef0-aba8-3142327ae2bc tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] [instance: 35bdbd45-5838-4b09-a232-7f4a02322ad7] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1104.092575] env[61923]: DEBUG oslo.service.loopingcall [None req-048a195a-ccb9-4ef0-aba8-3142327ae2bc tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61923) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1104.092909] env[61923]: DEBUG nova.compute.manager [-] [instance: 35bdbd45-5838-4b09-a232-7f4a02322ad7] Deallocating network for instance {{(pid=61923) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1104.093143] env[61923]: DEBUG nova.network.neutron [-] [instance: 35bdbd45-5838-4b09-a232-7f4a02322ad7] deallocate_for_instance() {{(pid=61923) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1104.160463] env[61923]: DEBUG oslo_concurrency.lockutils [None req-09014bcf-7023-44df-a13d-2100c4196634 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1104.308415] env[61923]: DEBUG nova.scheduler.client.report [None req-22e03437-b9fc-405c-b5bb-a6bb5bdc3129 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1104.566436] env[61923]: DEBUG nova.compute.manager [req-529d12f4-5acb-4fc0-882b-9c682bc8f948 req-495aa599-1c99-44d8-bc21-b3da53cc0475 service nova] [instance: 35bdbd45-5838-4b09-a232-7f4a02322ad7] Received event network-vif-deleted-42e3ee29-941c-43d0-a7f3-81fc7755f64b {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1104.566624] env[61923]: INFO nova.compute.manager [req-529d12f4-5acb-4fc0-882b-9c682bc8f948 req-495aa599-1c99-44d8-bc21-b3da53cc0475 service nova] [instance: 35bdbd45-5838-4b09-a232-7f4a02322ad7] Neutron deleted interface 42e3ee29-941c-43d0-a7f3-81fc7755f64b; detaching it from the instance and deleting it from the info cache [ 1104.567083] env[61923]: DEBUG nova.network.neutron [req-529d12f4-5acb-4fc0-882b-9c682bc8f948 req-495aa599-1c99-44d8-bc21-b3da53cc0475 service nova] [instance: 35bdbd45-5838-4b09-a232-7f4a02322ad7] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1104.813390] env[61923]: DEBUG oslo_concurrency.lockutils [None req-22e03437-b9fc-405c-b5bb-a6bb5bdc3129 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.149s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1104.813957] env[61923]: DEBUG nova.compute.manager [None req-22e03437-b9fc-405c-b5bb-a6bb5bdc3129 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: 6fddc45c-d589-4423-a990-4be24d9dc84e] Start building networks asynchronously for instance. {{(pid=61923) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1104.816622] env[61923]: DEBUG oslo_concurrency.lockutils [None req-09014bcf-7023-44df-a13d-2100c4196634 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.656s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1104.816849] env[61923]: DEBUG oslo_concurrency.lockutils [None req-09014bcf-7023-44df-a13d-2100c4196634 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1104.824624] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88247095-8436-4926-9351-a16fa8f2b00b {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.847179] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-c54f3d94-4858-42ec-a33f-5691c52b7589 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: c0a8a116-c7b6-4adc-849b-e0a4faa150d3] Updating instance 'c0a8a116-c7b6-4adc-849b-e0a4faa150d3' progress to 0 {{(pid=61923) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1104.853740] env[61923]: INFO nova.scheduler.client.report [None req-09014bcf-7023-44df-a13d-2100c4196634 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Deleted allocations for instance ab3edc53-37a5-4119-84c5-fd408ed66b69 [ 1105.036364] env[61923]: DEBUG nova.network.neutron [-] [instance: 35bdbd45-5838-4b09-a232-7f4a02322ad7] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1105.069665] env[61923]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-343f3e2f-9614-4e74-886d-f153fa24a664 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.079478] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8b55101-e059-42a3-b157-94d2b002f6c7 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.102885] env[61923]: DEBUG nova.compute.manager [req-529d12f4-5acb-4fc0-882b-9c682bc8f948 req-495aa599-1c99-44d8-bc21-b3da53cc0475 service nova] [instance: 35bdbd45-5838-4b09-a232-7f4a02322ad7] Detach interface failed, port_id=42e3ee29-941c-43d0-a7f3-81fc7755f64b, reason: Instance 35bdbd45-5838-4b09-a232-7f4a02322ad7 could not be found. {{(pid=61923) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1105.321246] env[61923]: DEBUG nova.compute.utils [None req-22e03437-b9fc-405c-b5bb-a6bb5bdc3129 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Using /dev/sd instead of None {{(pid=61923) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1105.323279] env[61923]: DEBUG nova.compute.manager [None req-22e03437-b9fc-405c-b5bb-a6bb5bdc3129 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: 6fddc45c-d589-4423-a990-4be24d9dc84e] Allocating IP information in the background. {{(pid=61923) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1105.323546] env[61923]: DEBUG nova.network.neutron [None req-22e03437-b9fc-405c-b5bb-a6bb5bdc3129 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: 6fddc45c-d589-4423-a990-4be24d9dc84e] allocate_for_instance() {{(pid=61923) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1105.353530] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-c54f3d94-4858-42ec-a33f-5691c52b7589 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: c0a8a116-c7b6-4adc-849b-e0a4faa150d3] Powering off the VM {{(pid=61923) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1105.353815] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-27a2a4ba-d849-4687-8085-7062652078a5 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.363032] env[61923]: DEBUG oslo_vmware.api [None req-c54f3d94-4858-42ec-a33f-5691c52b7589 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Waiting for the task: (returnval){ [ 1105.363032] env[61923]: value = "task-1378039" [ 1105.363032] env[61923]: _type = "Task" [ 1105.363032] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1105.363032] env[61923]: DEBUG oslo_concurrency.lockutils [None req-09014bcf-7023-44df-a13d-2100c4196634 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Lock "ab3edc53-37a5-4119-84c5-fd408ed66b69" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 4.140s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1105.364334] env[61923]: DEBUG oslo_concurrency.lockutils [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Lock "ab3edc53-37a5-4119-84c5-fd408ed66b69" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 1.988s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1105.365022] env[61923]: INFO nova.compute.manager [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] [instance: ab3edc53-37a5-4119-84c5-fd408ed66b69] During sync_power_state the instance has a pending task (deleting). Skip. [ 1105.365022] env[61923]: DEBUG oslo_concurrency.lockutils [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Lock "ab3edc53-37a5-4119-84c5-fd408ed66b69" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1105.374014] env[61923]: DEBUG oslo_vmware.api [None req-c54f3d94-4858-42ec-a33f-5691c52b7589 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Task: {'id': task-1378039, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1105.374879] env[61923]: DEBUG nova.policy [None req-22e03437-b9fc-405c-b5bb-a6bb5bdc3129 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6087a385c42e4383aa26a7f586cca87f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5a4db4f8aa194a219e80c538cc715d43', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61923) authorize /opt/stack/nova/nova/policy.py:201}} [ 1105.539151] env[61923]: INFO nova.compute.manager [-] [instance: 35bdbd45-5838-4b09-a232-7f4a02322ad7] Took 1.45 seconds to deallocate network for instance. [ 1105.630183] env[61923]: DEBUG nova.network.neutron [None req-22e03437-b9fc-405c-b5bb-a6bb5bdc3129 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: 6fddc45c-d589-4423-a990-4be24d9dc84e] Successfully created port: cdf08c60-c555-4f7e-9a38-0b43dc6bd520 {{(pid=61923) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1105.824257] env[61923]: DEBUG nova.compute.manager [None req-22e03437-b9fc-405c-b5bb-a6bb5bdc3129 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: 6fddc45c-d589-4423-a990-4be24d9dc84e] Start building block device mappings for instance. {{(pid=61923) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1105.874827] env[61923]: DEBUG oslo_vmware.api [None req-c54f3d94-4858-42ec-a33f-5691c52b7589 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Task: {'id': task-1378039, 'name': PowerOffVM_Task, 'duration_secs': 0.214638} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1105.875534] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-c54f3d94-4858-42ec-a33f-5691c52b7589 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: c0a8a116-c7b6-4adc-849b-e0a4faa150d3] Powered off the VM {{(pid=61923) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1105.875791] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-c54f3d94-4858-42ec-a33f-5691c52b7589 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: c0a8a116-c7b6-4adc-849b-e0a4faa150d3] Updating instance 'c0a8a116-c7b6-4adc-849b-e0a4faa150d3' progress to 17 {{(pid=61923) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1106.046261] env[61923]: DEBUG oslo_concurrency.lockutils [None req-048a195a-ccb9-4ef0-aba8-3142327ae2bc tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1106.046554] env[61923]: DEBUG oslo_concurrency.lockutils [None req-048a195a-ccb9-4ef0-aba8-3142327ae2bc tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1106.046781] env[61923]: DEBUG nova.objects.instance [None req-048a195a-ccb9-4ef0-aba8-3142327ae2bc tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Lazy-loading 'resources' on Instance uuid 35bdbd45-5838-4b09-a232-7f4a02322ad7 {{(pid=61923) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1106.384771] env[61923]: DEBUG nova.virt.hardware [None req-c54f3d94-4858-42ec-a33f-5691c52b7589 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-29T20:07:54Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=0f153f63-ae0a-45b1-b7f5-7f9b673c947f,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1106.385166] env[61923]: DEBUG nova.virt.hardware [None req-c54f3d94-4858-42ec-a33f-5691c52b7589 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Flavor limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1106.385207] env[61923]: DEBUG nova.virt.hardware [None req-c54f3d94-4858-42ec-a33f-5691c52b7589 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Image limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1106.385386] env[61923]: DEBUG nova.virt.hardware [None req-c54f3d94-4858-42ec-a33f-5691c52b7589 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Flavor pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1106.385539] env[61923]: DEBUG nova.virt.hardware [None req-c54f3d94-4858-42ec-a33f-5691c52b7589 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Image pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1106.385689] env[61923]: DEBUG nova.virt.hardware [None req-c54f3d94-4858-42ec-a33f-5691c52b7589 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1106.385915] env[61923]: DEBUG nova.virt.hardware [None req-c54f3d94-4858-42ec-a33f-5691c52b7589 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1106.386112] env[61923]: DEBUG nova.virt.hardware [None req-c54f3d94-4858-42ec-a33f-5691c52b7589 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1106.386305] env[61923]: DEBUG nova.virt.hardware [None req-c54f3d94-4858-42ec-a33f-5691c52b7589 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Got 1 possible topologies {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1106.386526] env[61923]: DEBUG nova.virt.hardware [None req-c54f3d94-4858-42ec-a33f-5691c52b7589 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1106.386712] env[61923]: DEBUG nova.virt.hardware [None req-c54f3d94-4858-42ec-a33f-5691c52b7589 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1106.391755] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-36a198e3-2ed9-4f6f-b959-dd1541394fd8 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.407426] env[61923]: DEBUG oslo_vmware.api [None req-c54f3d94-4858-42ec-a33f-5691c52b7589 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Waiting for the task: (returnval){ [ 1106.407426] env[61923]: value = "task-1378040" [ 1106.407426] env[61923]: _type = "Task" [ 1106.407426] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1106.416352] env[61923]: DEBUG oslo_vmware.api [None req-c54f3d94-4858-42ec-a33f-5691c52b7589 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Task: {'id': task-1378040, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1106.613650] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fcc0b007-d76a-46e4-b0d7-f2d92303c598 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.620791] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c83a5319-a229-4251-8eee-e283b42999e5 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.650472] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3071e414-d25f-47e4-98ae-8e88e16af261 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.658309] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-815024dd-870f-42fd-879f-e319396c2031 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.671769] env[61923]: DEBUG nova.compute.provider_tree [None req-048a195a-ccb9-4ef0-aba8-3142327ae2bc tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1106.833211] env[61923]: DEBUG nova.compute.manager [None req-22e03437-b9fc-405c-b5bb-a6bb5bdc3129 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: 6fddc45c-d589-4423-a990-4be24d9dc84e] Start spawning the instance on the hypervisor. {{(pid=61923) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1106.857854] env[61923]: DEBUG nova.virt.hardware [None req-22e03437-b9fc-405c-b5bb-a6bb5bdc3129 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-29T20:07:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-29T20:07:35Z,direct_url=,disk_format='vmdk',id=0f153f63-ae0a-45b1-b7f5-7f9b673c947f,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='4e7b5e3b3c3443c8bd5c70f8a15739f5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-29T20:07:36Z,virtual_size=,visibility=), allow threads: False {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1106.858118] env[61923]: DEBUG nova.virt.hardware [None req-22e03437-b9fc-405c-b5bb-a6bb5bdc3129 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Flavor limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1106.858282] env[61923]: DEBUG nova.virt.hardware [None req-22e03437-b9fc-405c-b5bb-a6bb5bdc3129 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Image limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1106.858467] env[61923]: DEBUG nova.virt.hardware [None req-22e03437-b9fc-405c-b5bb-a6bb5bdc3129 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Flavor pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1106.858723] env[61923]: DEBUG nova.virt.hardware [None req-22e03437-b9fc-405c-b5bb-a6bb5bdc3129 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Image pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1106.858938] env[61923]: DEBUG nova.virt.hardware [None req-22e03437-b9fc-405c-b5bb-a6bb5bdc3129 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1106.859179] env[61923]: DEBUG nova.virt.hardware [None req-22e03437-b9fc-405c-b5bb-a6bb5bdc3129 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1106.859444] env[61923]: DEBUG nova.virt.hardware [None req-22e03437-b9fc-405c-b5bb-a6bb5bdc3129 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1106.859650] env[61923]: DEBUG nova.virt.hardware [None req-22e03437-b9fc-405c-b5bb-a6bb5bdc3129 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Got 1 possible topologies {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1106.859823] env[61923]: DEBUG nova.virt.hardware [None req-22e03437-b9fc-405c-b5bb-a6bb5bdc3129 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1106.859998] env[61923]: DEBUG nova.virt.hardware [None req-22e03437-b9fc-405c-b5bb-a6bb5bdc3129 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1106.860855] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b9c86c5-58d2-4cae-9e76-67905fa186d9 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.869116] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd400fc3-7759-4341-a67b-797a461f0211 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.916895] env[61923]: DEBUG oslo_vmware.api [None req-c54f3d94-4858-42ec-a33f-5691c52b7589 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Task: {'id': task-1378040, 'name': ReconfigVM_Task, 'duration_secs': 0.143155} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1106.917252] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-c54f3d94-4858-42ec-a33f-5691c52b7589 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: c0a8a116-c7b6-4adc-849b-e0a4faa150d3] Updating instance 'c0a8a116-c7b6-4adc-849b-e0a4faa150d3' progress to 33 {{(pid=61923) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1107.013966] env[61923]: DEBUG nova.compute.manager [req-b32a113e-d259-4d63-b433-38e59bc8ed85 req-16559f1a-3c5e-43d9-9a0e-dabeedf41813 service nova] [instance: 6fddc45c-d589-4423-a990-4be24d9dc84e] Received event network-vif-plugged-cdf08c60-c555-4f7e-9a38-0b43dc6bd520 {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1107.014133] env[61923]: DEBUG oslo_concurrency.lockutils [req-b32a113e-d259-4d63-b433-38e59bc8ed85 req-16559f1a-3c5e-43d9-9a0e-dabeedf41813 service nova] Acquiring lock "6fddc45c-d589-4423-a990-4be24d9dc84e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1107.014334] env[61923]: DEBUG oslo_concurrency.lockutils [req-b32a113e-d259-4d63-b433-38e59bc8ed85 req-16559f1a-3c5e-43d9-9a0e-dabeedf41813 service nova] Lock "6fddc45c-d589-4423-a990-4be24d9dc84e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1107.014515] env[61923]: DEBUG oslo_concurrency.lockutils [req-b32a113e-d259-4d63-b433-38e59bc8ed85 req-16559f1a-3c5e-43d9-9a0e-dabeedf41813 service nova] Lock "6fddc45c-d589-4423-a990-4be24d9dc84e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1107.014685] env[61923]: DEBUG nova.compute.manager [req-b32a113e-d259-4d63-b433-38e59bc8ed85 req-16559f1a-3c5e-43d9-9a0e-dabeedf41813 service nova] [instance: 6fddc45c-d589-4423-a990-4be24d9dc84e] No waiting events found dispatching network-vif-plugged-cdf08c60-c555-4f7e-9a38-0b43dc6bd520 {{(pid=61923) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1107.014854] env[61923]: WARNING nova.compute.manager [req-b32a113e-d259-4d63-b433-38e59bc8ed85 req-16559f1a-3c5e-43d9-9a0e-dabeedf41813 service nova] [instance: 6fddc45c-d589-4423-a990-4be24d9dc84e] Received unexpected event network-vif-plugged-cdf08c60-c555-4f7e-9a38-0b43dc6bd520 for instance with vm_state building and task_state spawning. [ 1107.096058] env[61923]: DEBUG nova.network.neutron [None req-22e03437-b9fc-405c-b5bb-a6bb5bdc3129 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: 6fddc45c-d589-4423-a990-4be24d9dc84e] Successfully updated port: cdf08c60-c555-4f7e-9a38-0b43dc6bd520 {{(pid=61923) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1107.174696] env[61923]: DEBUG nova.scheduler.client.report [None req-048a195a-ccb9-4ef0-aba8-3142327ae2bc tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1107.423777] env[61923]: DEBUG nova.virt.hardware [None req-c54f3d94-4858-42ec-a33f-5691c52b7589 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-29T20:07:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=0f153f63-ae0a-45b1-b7f5-7f9b673c947f,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1107.424202] env[61923]: DEBUG nova.virt.hardware [None req-c54f3d94-4858-42ec-a33f-5691c52b7589 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Flavor limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1107.424202] env[61923]: DEBUG nova.virt.hardware [None req-c54f3d94-4858-42ec-a33f-5691c52b7589 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Image limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1107.424380] env[61923]: DEBUG nova.virt.hardware [None req-c54f3d94-4858-42ec-a33f-5691c52b7589 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Flavor pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1107.424530] env[61923]: DEBUG nova.virt.hardware [None req-c54f3d94-4858-42ec-a33f-5691c52b7589 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Image pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1107.424680] env[61923]: DEBUG nova.virt.hardware [None req-c54f3d94-4858-42ec-a33f-5691c52b7589 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1107.424912] env[61923]: DEBUG nova.virt.hardware [None req-c54f3d94-4858-42ec-a33f-5691c52b7589 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1107.425052] env[61923]: DEBUG nova.virt.hardware [None req-c54f3d94-4858-42ec-a33f-5691c52b7589 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1107.425229] env[61923]: DEBUG nova.virt.hardware [None req-c54f3d94-4858-42ec-a33f-5691c52b7589 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Got 1 possible topologies {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1107.425390] env[61923]: DEBUG nova.virt.hardware [None req-c54f3d94-4858-42ec-a33f-5691c52b7589 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1107.425565] env[61923]: DEBUG nova.virt.hardware [None req-c54f3d94-4858-42ec-a33f-5691c52b7589 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1107.430813] env[61923]: DEBUG nova.virt.vmwareapi.volumeops [None req-c54f3d94-4858-42ec-a33f-5691c52b7589 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: c0a8a116-c7b6-4adc-849b-e0a4faa150d3] Reconfiguring VM instance instance-00000065 to detach disk 2000 {{(pid=61923) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1107.431112] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-381da983-0ecd-406a-9a75-165649d72948 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1107.450261] env[61923]: DEBUG oslo_vmware.api [None req-c54f3d94-4858-42ec-a33f-5691c52b7589 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Waiting for the task: (returnval){ [ 1107.450261] env[61923]: value = "task-1378041" [ 1107.450261] env[61923]: _type = "Task" [ 1107.450261] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1107.458016] env[61923]: DEBUG oslo_vmware.api [None req-c54f3d94-4858-42ec-a33f-5691c52b7589 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Task: {'id': task-1378041, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1107.598701] env[61923]: DEBUG oslo_concurrency.lockutils [None req-22e03437-b9fc-405c-b5bb-a6bb5bdc3129 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Acquiring lock "refresh_cache-6fddc45c-d589-4423-a990-4be24d9dc84e" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1107.598899] env[61923]: DEBUG oslo_concurrency.lockutils [None req-22e03437-b9fc-405c-b5bb-a6bb5bdc3129 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Acquired lock "refresh_cache-6fddc45c-d589-4423-a990-4be24d9dc84e" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1107.599040] env[61923]: DEBUG nova.network.neutron [None req-22e03437-b9fc-405c-b5bb-a6bb5bdc3129 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: 6fddc45c-d589-4423-a990-4be24d9dc84e] Building network info cache for instance {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1107.679363] env[61923]: DEBUG oslo_concurrency.lockutils [None req-048a195a-ccb9-4ef0-aba8-3142327ae2bc tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.633s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1107.697377] env[61923]: INFO nova.scheduler.client.report [None req-048a195a-ccb9-4ef0-aba8-3142327ae2bc tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Deleted allocations for instance 35bdbd45-5838-4b09-a232-7f4a02322ad7 [ 1107.960546] env[61923]: DEBUG oslo_vmware.api [None req-c54f3d94-4858-42ec-a33f-5691c52b7589 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Task: {'id': task-1378041, 'name': ReconfigVM_Task, 'duration_secs': 0.147605} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1107.960848] env[61923]: DEBUG nova.virt.vmwareapi.volumeops [None req-c54f3d94-4858-42ec-a33f-5691c52b7589 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: c0a8a116-c7b6-4adc-849b-e0a4faa150d3] Reconfigured VM instance instance-00000065 to detach disk 2000 {{(pid=61923) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1107.961699] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84dcf4e2-431b-499c-9ebb-8cd03b80e700 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1107.984777] env[61923]: DEBUG nova.virt.vmwareapi.volumeops [None req-c54f3d94-4858-42ec-a33f-5691c52b7589 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: c0a8a116-c7b6-4adc-849b-e0a4faa150d3] Reconfiguring VM instance instance-00000065 to attach disk [datastore2] c0a8a116-c7b6-4adc-849b-e0a4faa150d3/c0a8a116-c7b6-4adc-849b-e0a4faa150d3.vmdk or device None with type thin {{(pid=61923) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1107.984777] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0ffb3a55-290e-420e-8f37-f7e96492ddb9 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.003227] env[61923]: DEBUG oslo_vmware.api [None req-c54f3d94-4858-42ec-a33f-5691c52b7589 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Waiting for the task: (returnval){ [ 1108.003227] env[61923]: value = "task-1378042" [ 1108.003227] env[61923]: _type = "Task" [ 1108.003227] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1108.010905] env[61923]: DEBUG oslo_vmware.api [None req-c54f3d94-4858-42ec-a33f-5691c52b7589 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Task: {'id': task-1378042, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1108.140743] env[61923]: DEBUG nova.network.neutron [None req-22e03437-b9fc-405c-b5bb-a6bb5bdc3129 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: 6fddc45c-d589-4423-a990-4be24d9dc84e] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1108.204871] env[61923]: DEBUG oslo_concurrency.lockutils [None req-048a195a-ccb9-4ef0-aba8-3142327ae2bc tempest-AttachVolumeTestJSON-2111165375 tempest-AttachVolumeTestJSON-2111165375-project-member] Lock "35bdbd45-5838-4b09-a232-7f4a02322ad7" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.227s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1108.205794] env[61923]: DEBUG oslo_concurrency.lockutils [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Lock "35bdbd45-5838-4b09-a232-7f4a02322ad7" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 4.830s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1108.205984] env[61923]: INFO nova.compute.manager [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] [instance: 35bdbd45-5838-4b09-a232-7f4a02322ad7] During sync_power_state the instance has a pending task (deleting). Skip. [ 1108.206172] env[61923]: DEBUG oslo_concurrency.lockutils [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Lock "35bdbd45-5838-4b09-a232-7f4a02322ad7" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1108.339168] env[61923]: DEBUG nova.network.neutron [None req-22e03437-b9fc-405c-b5bb-a6bb5bdc3129 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: 6fddc45c-d589-4423-a990-4be24d9dc84e] Updating instance_info_cache with network_info: [{"id": "cdf08c60-c555-4f7e-9a38-0b43dc6bd520", "address": "fa:16:3e:50:f7:08", "network": {"id": "77aec6ca-d1be-4b39-9d51-08853a1183a1", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1813085414-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5a4db4f8aa194a219e80c538cc715d43", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "56b944d8-803d-43f2-945d-0f334ee4ea1c", "external-id": "nsx-vlan-transportzone-799", "segmentation_id": 799, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcdf08c60-c5", "ovs_interfaceid": "cdf08c60-c555-4f7e-9a38-0b43dc6bd520", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1108.512894] env[61923]: DEBUG oslo_vmware.api [None req-c54f3d94-4858-42ec-a33f-5691c52b7589 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Task: {'id': task-1378042, 'name': ReconfigVM_Task, 'duration_secs': 0.252187} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1108.513227] env[61923]: DEBUG nova.virt.vmwareapi.volumeops [None req-c54f3d94-4858-42ec-a33f-5691c52b7589 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: c0a8a116-c7b6-4adc-849b-e0a4faa150d3] Reconfigured VM instance instance-00000065 to attach disk [datastore2] c0a8a116-c7b6-4adc-849b-e0a4faa150d3/c0a8a116-c7b6-4adc-849b-e0a4faa150d3.vmdk or device None with type thin {{(pid=61923) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1108.513443] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-c54f3d94-4858-42ec-a33f-5691c52b7589 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: c0a8a116-c7b6-4adc-849b-e0a4faa150d3] Updating instance 'c0a8a116-c7b6-4adc-849b-e0a4faa150d3' progress to 50 {{(pid=61923) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1108.841826] env[61923]: DEBUG oslo_concurrency.lockutils [None req-22e03437-b9fc-405c-b5bb-a6bb5bdc3129 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Releasing lock "refresh_cache-6fddc45c-d589-4423-a990-4be24d9dc84e" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1108.842557] env[61923]: DEBUG nova.compute.manager [None req-22e03437-b9fc-405c-b5bb-a6bb5bdc3129 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: 6fddc45c-d589-4423-a990-4be24d9dc84e] Instance network_info: |[{"id": "cdf08c60-c555-4f7e-9a38-0b43dc6bd520", "address": "fa:16:3e:50:f7:08", "network": {"id": "77aec6ca-d1be-4b39-9d51-08853a1183a1", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1813085414-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5a4db4f8aa194a219e80c538cc715d43", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "56b944d8-803d-43f2-945d-0f334ee4ea1c", "external-id": "nsx-vlan-transportzone-799", "segmentation_id": 799, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcdf08c60-c5", "ovs_interfaceid": "cdf08c60-c555-4f7e-9a38-0b43dc6bd520", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61923) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1108.842743] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-22e03437-b9fc-405c-b5bb-a6bb5bdc3129 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: 6fddc45c-d589-4423-a990-4be24d9dc84e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:50:f7:08', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '56b944d8-803d-43f2-945d-0f334ee4ea1c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'cdf08c60-c555-4f7e-9a38-0b43dc6bd520', 'vif_model': 'vmxnet3'}] {{(pid=61923) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1108.851229] env[61923]: DEBUG oslo.service.loopingcall [None req-22e03437-b9fc-405c-b5bb-a6bb5bdc3129 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61923) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1108.851524] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6fddc45c-d589-4423-a990-4be24d9dc84e] Creating VM on the ESX host {{(pid=61923) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1108.851815] env[61923]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-8d293bf3-f727-4b42-9472-567d319fc4f4 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.873049] env[61923]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1108.873049] env[61923]: value = "task-1378044" [ 1108.873049] env[61923]: _type = "Task" [ 1108.873049] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1108.880386] env[61923]: DEBUG oslo_vmware.api [-] Task: {'id': task-1378044, 'name': CreateVM_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1109.020221] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9469fbde-11f5-4551-970b-244b53add954 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.039719] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1cc037a-50db-46b6-9742-f208093809c9 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.043189] env[61923]: DEBUG nova.compute.manager [req-95ca3baf-a96f-4970-8116-1048678727cd req-523ca1ef-fd8b-4e77-9a98-db6cb4817231 service nova] [instance: 6fddc45c-d589-4423-a990-4be24d9dc84e] Received event network-changed-cdf08c60-c555-4f7e-9a38-0b43dc6bd520 {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1109.043385] env[61923]: DEBUG nova.compute.manager [req-95ca3baf-a96f-4970-8116-1048678727cd req-523ca1ef-fd8b-4e77-9a98-db6cb4817231 service nova] [instance: 6fddc45c-d589-4423-a990-4be24d9dc84e] Refreshing instance network info cache due to event network-changed-cdf08c60-c555-4f7e-9a38-0b43dc6bd520. {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1109.043604] env[61923]: DEBUG oslo_concurrency.lockutils [req-95ca3baf-a96f-4970-8116-1048678727cd req-523ca1ef-fd8b-4e77-9a98-db6cb4817231 service nova] Acquiring lock "refresh_cache-6fddc45c-d589-4423-a990-4be24d9dc84e" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1109.043749] env[61923]: DEBUG oslo_concurrency.lockutils [req-95ca3baf-a96f-4970-8116-1048678727cd req-523ca1ef-fd8b-4e77-9a98-db6cb4817231 service nova] Acquired lock "refresh_cache-6fddc45c-d589-4423-a990-4be24d9dc84e" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1109.043911] env[61923]: DEBUG nova.network.neutron [req-95ca3baf-a96f-4970-8116-1048678727cd req-523ca1ef-fd8b-4e77-9a98-db6cb4817231 service nova] [instance: 6fddc45c-d589-4423-a990-4be24d9dc84e] Refreshing network info cache for port cdf08c60-c555-4f7e-9a38-0b43dc6bd520 {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1109.060146] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-c54f3d94-4858-42ec-a33f-5691c52b7589 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: c0a8a116-c7b6-4adc-849b-e0a4faa150d3] Updating instance 'c0a8a116-c7b6-4adc-849b-e0a4faa150d3' progress to 67 {{(pid=61923) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1109.382699] env[61923]: DEBUG oslo_vmware.api [-] Task: {'id': task-1378044, 'name': CreateVM_Task, 'duration_secs': 0.307398} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1109.382880] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6fddc45c-d589-4423-a990-4be24d9dc84e] Created VM on the ESX host {{(pid=61923) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1109.383568] env[61923]: DEBUG oslo_concurrency.lockutils [None req-22e03437-b9fc-405c-b5bb-a6bb5bdc3129 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1109.383771] env[61923]: DEBUG oslo_concurrency.lockutils [None req-22e03437-b9fc-405c-b5bb-a6bb5bdc3129 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1109.384121] env[61923]: DEBUG oslo_concurrency.lockutils [None req-22e03437-b9fc-405c-b5bb-a6bb5bdc3129 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1109.384374] env[61923]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ddb60e4d-7511-4f79-bc70-bd8f4aecf289 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.388384] env[61923]: DEBUG oslo_vmware.api [None req-22e03437-b9fc-405c-b5bb-a6bb5bdc3129 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Waiting for the task: (returnval){ [ 1109.388384] env[61923]: value = "session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]521c30bc-e1f9-4533-d119-591928e82c05" [ 1109.388384] env[61923]: _type = "Task" [ 1109.388384] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1109.395419] env[61923]: DEBUG oslo_vmware.api [None req-22e03437-b9fc-405c-b5bb-a6bb5bdc3129 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]521c30bc-e1f9-4533-d119-591928e82c05, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1109.599320] env[61923]: DEBUG nova.network.neutron [None req-c54f3d94-4858-42ec-a33f-5691c52b7589 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: c0a8a116-c7b6-4adc-849b-e0a4faa150d3] Port c210c30f-4f3c-4698-ab24-a6a81c78e658 binding to destination host cpu-1 is already ACTIVE {{(pid=61923) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 1109.898334] env[61923]: DEBUG oslo_vmware.api [None req-22e03437-b9fc-405c-b5bb-a6bb5bdc3129 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]521c30bc-e1f9-4533-d119-591928e82c05, 'name': SearchDatastore_Task, 'duration_secs': 0.008753} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1109.898640] env[61923]: DEBUG oslo_concurrency.lockutils [None req-22e03437-b9fc-405c-b5bb-a6bb5bdc3129 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1109.898886] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-22e03437-b9fc-405c-b5bb-a6bb5bdc3129 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: 6fddc45c-d589-4423-a990-4be24d9dc84e] Processing image 0f153f63-ae0a-45b1-b7f5-7f9b673c947f {{(pid=61923) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1109.899138] env[61923]: DEBUG oslo_concurrency.lockutils [None req-22e03437-b9fc-405c-b5bb-a6bb5bdc3129 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1109.899292] env[61923]: DEBUG oslo_concurrency.lockutils [None req-22e03437-b9fc-405c-b5bb-a6bb5bdc3129 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1109.899474] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-22e03437-b9fc-405c-b5bb-a6bb5bdc3129 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61923) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1109.899739] env[61923]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-cddb114f-eb55-45a2-a63f-d36ab2594b13 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.907371] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-22e03437-b9fc-405c-b5bb-a6bb5bdc3129 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61923) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1109.907557] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-22e03437-b9fc-405c-b5bb-a6bb5bdc3129 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61923) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1109.908271] env[61923]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ac7c8973-a142-4b5a-93ed-be135733386c {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.910859] env[61923]: DEBUG nova.network.neutron [req-95ca3baf-a96f-4970-8116-1048678727cd req-523ca1ef-fd8b-4e77-9a98-db6cb4817231 service nova] [instance: 6fddc45c-d589-4423-a990-4be24d9dc84e] Updated VIF entry in instance network info cache for port cdf08c60-c555-4f7e-9a38-0b43dc6bd520. {{(pid=61923) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1109.911198] env[61923]: DEBUG nova.network.neutron [req-95ca3baf-a96f-4970-8116-1048678727cd req-523ca1ef-fd8b-4e77-9a98-db6cb4817231 service nova] [instance: 6fddc45c-d589-4423-a990-4be24d9dc84e] Updating instance_info_cache with network_info: [{"id": "cdf08c60-c555-4f7e-9a38-0b43dc6bd520", "address": "fa:16:3e:50:f7:08", "network": {"id": "77aec6ca-d1be-4b39-9d51-08853a1183a1", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1813085414-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5a4db4f8aa194a219e80c538cc715d43", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "56b944d8-803d-43f2-945d-0f334ee4ea1c", "external-id": "nsx-vlan-transportzone-799", "segmentation_id": 799, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcdf08c60-c5", "ovs_interfaceid": "cdf08c60-c555-4f7e-9a38-0b43dc6bd520", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1109.915034] env[61923]: DEBUG oslo_vmware.api [None req-22e03437-b9fc-405c-b5bb-a6bb5bdc3129 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Waiting for the task: (returnval){ [ 1109.915034] env[61923]: value = "session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]5220817a-4531-9a64-5791-bfa8928464f8" [ 1109.915034] env[61923]: _type = "Task" [ 1109.915034] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1109.924186] env[61923]: DEBUG oslo_vmware.api [None req-22e03437-b9fc-405c-b5bb-a6bb5bdc3129 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]5220817a-4531-9a64-5791-bfa8928464f8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1110.413720] env[61923]: DEBUG oslo_concurrency.lockutils [req-95ca3baf-a96f-4970-8116-1048678727cd req-523ca1ef-fd8b-4e77-9a98-db6cb4817231 service nova] Releasing lock "refresh_cache-6fddc45c-d589-4423-a990-4be24d9dc84e" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1110.425335] env[61923]: DEBUG oslo_vmware.api [None req-22e03437-b9fc-405c-b5bb-a6bb5bdc3129 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]5220817a-4531-9a64-5791-bfa8928464f8, 'name': SearchDatastore_Task, 'duration_secs': 0.008262} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1110.426072] env[61923]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6a7a102f-3888-4570-9913-26d1c124f20a {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.430773] env[61923]: DEBUG oslo_vmware.api [None req-22e03437-b9fc-405c-b5bb-a6bb5bdc3129 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Waiting for the task: (returnval){ [ 1110.430773] env[61923]: value = "session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]526c8cb8-186c-3f2e-7398-636512f378f6" [ 1110.430773] env[61923]: _type = "Task" [ 1110.430773] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1110.437893] env[61923]: DEBUG oslo_vmware.api [None req-22e03437-b9fc-405c-b5bb-a6bb5bdc3129 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]526c8cb8-186c-3f2e-7398-636512f378f6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1110.621345] env[61923]: DEBUG oslo_concurrency.lockutils [None req-c54f3d94-4858-42ec-a33f-5691c52b7589 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Acquiring lock "c0a8a116-c7b6-4adc-849b-e0a4faa150d3-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1110.621701] env[61923]: DEBUG oslo_concurrency.lockutils [None req-c54f3d94-4858-42ec-a33f-5691c52b7589 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Lock "c0a8a116-c7b6-4adc-849b-e0a4faa150d3-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1110.621767] env[61923]: DEBUG oslo_concurrency.lockutils [None req-c54f3d94-4858-42ec-a33f-5691c52b7589 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Lock "c0a8a116-c7b6-4adc-849b-e0a4faa150d3-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1110.940670] env[61923]: DEBUG oslo_vmware.api [None req-22e03437-b9fc-405c-b5bb-a6bb5bdc3129 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]526c8cb8-186c-3f2e-7398-636512f378f6, 'name': SearchDatastore_Task, 'duration_secs': 0.008826} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1110.940930] env[61923]: DEBUG oslo_concurrency.lockutils [None req-22e03437-b9fc-405c-b5bb-a6bb5bdc3129 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1110.941209] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-22e03437-b9fc-405c-b5bb-a6bb5bdc3129 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk to [datastore1] 6fddc45c-d589-4423-a990-4be24d9dc84e/6fddc45c-d589-4423-a990-4be24d9dc84e.vmdk {{(pid=61923) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1110.941506] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-85ac1e13-6b7b-439a-a385-566c03b50d21 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1110.948072] env[61923]: DEBUG oslo_vmware.api [None req-22e03437-b9fc-405c-b5bb-a6bb5bdc3129 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Waiting for the task: (returnval){ [ 1110.948072] env[61923]: value = "task-1378045" [ 1110.948072] env[61923]: _type = "Task" [ 1110.948072] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1110.955391] env[61923]: DEBUG oslo_vmware.api [None req-22e03437-b9fc-405c-b5bb-a6bb5bdc3129 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Task: {'id': task-1378045, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1111.457254] env[61923]: DEBUG oslo_vmware.api [None req-22e03437-b9fc-405c-b5bb-a6bb5bdc3129 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Task: {'id': task-1378045, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.463319} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1111.457621] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-22e03437-b9fc-405c-b5bb-a6bb5bdc3129 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk to [datastore1] 6fddc45c-d589-4423-a990-4be24d9dc84e/6fddc45c-d589-4423-a990-4be24d9dc84e.vmdk {{(pid=61923) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1111.457922] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-22e03437-b9fc-405c-b5bb-a6bb5bdc3129 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: 6fddc45c-d589-4423-a990-4be24d9dc84e] Extending root virtual disk to 1048576 {{(pid=61923) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1111.458236] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-3a74cb26-144d-47ff-b71c-375771b24239 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.464320] env[61923]: DEBUG oslo_vmware.api [None req-22e03437-b9fc-405c-b5bb-a6bb5bdc3129 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Waiting for the task: (returnval){ [ 1111.464320] env[61923]: value = "task-1378046" [ 1111.464320] env[61923]: _type = "Task" [ 1111.464320] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1111.471206] env[61923]: DEBUG oslo_vmware.api [None req-22e03437-b9fc-405c-b5bb-a6bb5bdc3129 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Task: {'id': task-1378046, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1111.656669] env[61923]: DEBUG oslo_concurrency.lockutils [None req-c54f3d94-4858-42ec-a33f-5691c52b7589 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Acquiring lock "refresh_cache-c0a8a116-c7b6-4adc-849b-e0a4faa150d3" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1111.656911] env[61923]: DEBUG oslo_concurrency.lockutils [None req-c54f3d94-4858-42ec-a33f-5691c52b7589 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Acquired lock "refresh_cache-c0a8a116-c7b6-4adc-849b-e0a4faa150d3" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1111.657045] env[61923]: DEBUG nova.network.neutron [None req-c54f3d94-4858-42ec-a33f-5691c52b7589 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: c0a8a116-c7b6-4adc-849b-e0a4faa150d3] Building network info cache for instance {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1111.973935] env[61923]: DEBUG oslo_vmware.api [None req-22e03437-b9fc-405c-b5bb-a6bb5bdc3129 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Task: {'id': task-1378046, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.091754} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1111.974237] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-22e03437-b9fc-405c-b5bb-a6bb5bdc3129 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: 6fddc45c-d589-4423-a990-4be24d9dc84e] Extended root virtual disk {{(pid=61923) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1111.974971] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a5df1c9-2722-4b16-ba93-4254e4c51947 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.997047] env[61923]: DEBUG nova.virt.vmwareapi.volumeops [None req-22e03437-b9fc-405c-b5bb-a6bb5bdc3129 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: 6fddc45c-d589-4423-a990-4be24d9dc84e] Reconfiguring VM instance instance-00000066 to attach disk [datastore1] 6fddc45c-d589-4423-a990-4be24d9dc84e/6fddc45c-d589-4423-a990-4be24d9dc84e.vmdk or device None with type sparse {{(pid=61923) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1111.997275] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-43dba962-ff6c-4b41-97e1-ba9a47cd2dbf {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.016157] env[61923]: DEBUG oslo_vmware.api [None req-22e03437-b9fc-405c-b5bb-a6bb5bdc3129 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Waiting for the task: (returnval){ [ 1112.016157] env[61923]: value = "task-1378048" [ 1112.016157] env[61923]: _type = "Task" [ 1112.016157] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1112.023160] env[61923]: DEBUG oslo_vmware.api [None req-22e03437-b9fc-405c-b5bb-a6bb5bdc3129 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Task: {'id': task-1378048, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1112.355550] env[61923]: DEBUG nova.network.neutron [None req-c54f3d94-4858-42ec-a33f-5691c52b7589 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: c0a8a116-c7b6-4adc-849b-e0a4faa150d3] Updating instance_info_cache with network_info: [{"id": "c210c30f-4f3c-4698-ab24-a6a81c78e658", "address": "fa:16:3e:ad:1c:c4", "network": {"id": "9533e79d-fde2-4c10-86a0-86a36845a8cf", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1736743137-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "780d3541d9604417b977bb62390c4299", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6d650b26-c3e7-4de7-98db-5e4b816d123a", "external-id": "nsx-vlan-transportzone-757", "segmentation_id": 757, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc210c30f-4f", "ovs_interfaceid": "c210c30f-4f3c-4698-ab24-a6a81c78e658", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1112.526229] env[61923]: DEBUG oslo_vmware.api [None req-22e03437-b9fc-405c-b5bb-a6bb5bdc3129 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Task: {'id': task-1378048, 'name': ReconfigVM_Task, 'duration_secs': 0.254348} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1112.526579] env[61923]: DEBUG nova.virt.vmwareapi.volumeops [None req-22e03437-b9fc-405c-b5bb-a6bb5bdc3129 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: 6fddc45c-d589-4423-a990-4be24d9dc84e] Reconfigured VM instance instance-00000066 to attach disk [datastore1] 6fddc45c-d589-4423-a990-4be24d9dc84e/6fddc45c-d589-4423-a990-4be24d9dc84e.vmdk or device None with type sparse {{(pid=61923) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1112.527306] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-876b6f08-1812-4bde-835d-3e54c570f7b8 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.532889] env[61923]: DEBUG oslo_vmware.api [None req-22e03437-b9fc-405c-b5bb-a6bb5bdc3129 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Waiting for the task: (returnval){ [ 1112.532889] env[61923]: value = "task-1378049" [ 1112.532889] env[61923]: _type = "Task" [ 1112.532889] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1112.540189] env[61923]: DEBUG oslo_vmware.api [None req-22e03437-b9fc-405c-b5bb-a6bb5bdc3129 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Task: {'id': task-1378049, 'name': Rename_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1112.858445] env[61923]: DEBUG oslo_concurrency.lockutils [None req-c54f3d94-4858-42ec-a33f-5691c52b7589 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Releasing lock "refresh_cache-c0a8a116-c7b6-4adc-849b-e0a4faa150d3" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1113.042869] env[61923]: DEBUG oslo_vmware.api [None req-22e03437-b9fc-405c-b5bb-a6bb5bdc3129 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Task: {'id': task-1378049, 'name': Rename_Task, 'duration_secs': 0.131875} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1113.043165] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-22e03437-b9fc-405c-b5bb-a6bb5bdc3129 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: 6fddc45c-d589-4423-a990-4be24d9dc84e] Powering on the VM {{(pid=61923) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1113.043411] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-76483d26-99dc-4a70-b194-2de31c8c4b6a {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.050387] env[61923]: DEBUG oslo_vmware.api [None req-22e03437-b9fc-405c-b5bb-a6bb5bdc3129 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Waiting for the task: (returnval){ [ 1113.050387] env[61923]: value = "task-1378050" [ 1113.050387] env[61923]: _type = "Task" [ 1113.050387] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1113.057490] env[61923]: DEBUG oslo_vmware.api [None req-22e03437-b9fc-405c-b5bb-a6bb5bdc3129 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Task: {'id': task-1378050, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1113.386385] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c87fa99-49f6-426e-8971-f52327377539 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.406961] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-104fd251-1679-4c30-90bb-a63c87124a69 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.414200] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-c54f3d94-4858-42ec-a33f-5691c52b7589 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: c0a8a116-c7b6-4adc-849b-e0a4faa150d3] Updating instance 'c0a8a116-c7b6-4adc-849b-e0a4faa150d3' progress to 83 {{(pid=61923) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1113.559979] env[61923]: DEBUG oslo_vmware.api [None req-22e03437-b9fc-405c-b5bb-a6bb5bdc3129 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Task: {'id': task-1378050, 'name': PowerOnVM_Task, 'duration_secs': 0.428229} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1113.560221] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-22e03437-b9fc-405c-b5bb-a6bb5bdc3129 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: 6fddc45c-d589-4423-a990-4be24d9dc84e] Powered on the VM {{(pid=61923) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1113.560421] env[61923]: INFO nova.compute.manager [None req-22e03437-b9fc-405c-b5bb-a6bb5bdc3129 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: 6fddc45c-d589-4423-a990-4be24d9dc84e] Took 6.73 seconds to spawn the instance on the hypervisor. [ 1113.560603] env[61923]: DEBUG nova.compute.manager [None req-22e03437-b9fc-405c-b5bb-a6bb5bdc3129 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: 6fddc45c-d589-4423-a990-4be24d9dc84e] Checking state {{(pid=61923) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1113.561361] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-abd19df5-fa1e-4132-8f6b-e4fb71ea1193 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.920887] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-c54f3d94-4858-42ec-a33f-5691c52b7589 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: c0a8a116-c7b6-4adc-849b-e0a4faa150d3] Powering on the VM {{(pid=61923) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1113.921257] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6c0c2632-e1f2-4f83-86d9-6e5e68dcfdb3 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.929194] env[61923]: DEBUG oslo_vmware.api [None req-c54f3d94-4858-42ec-a33f-5691c52b7589 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Waiting for the task: (returnval){ [ 1113.929194] env[61923]: value = "task-1378051" [ 1113.929194] env[61923]: _type = "Task" [ 1113.929194] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1113.938293] env[61923]: DEBUG oslo_vmware.api [None req-c54f3d94-4858-42ec-a33f-5691c52b7589 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Task: {'id': task-1378051, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1114.079027] env[61923]: INFO nova.compute.manager [None req-22e03437-b9fc-405c-b5bb-a6bb5bdc3129 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: 6fddc45c-d589-4423-a990-4be24d9dc84e] Took 11.44 seconds to build instance. [ 1114.440279] env[61923]: DEBUG oslo_vmware.api [None req-c54f3d94-4858-42ec-a33f-5691c52b7589 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Task: {'id': task-1378051, 'name': PowerOnVM_Task, 'duration_secs': 0.353958} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1114.440983] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-c54f3d94-4858-42ec-a33f-5691c52b7589 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: c0a8a116-c7b6-4adc-849b-e0a4faa150d3] Powered on the VM {{(pid=61923) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1114.440983] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-c54f3d94-4858-42ec-a33f-5691c52b7589 tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: c0a8a116-c7b6-4adc-849b-e0a4faa150d3] Updating instance 'c0a8a116-c7b6-4adc-849b-e0a4faa150d3' progress to 100 {{(pid=61923) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1114.582058] env[61923]: DEBUG oslo_concurrency.lockutils [None req-22e03437-b9fc-405c-b5bb-a6bb5bdc3129 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Lock "6fddc45c-d589-4423-a990-4be24d9dc84e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 12.945s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1114.582058] env[61923]: DEBUG oslo_concurrency.lockutils [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Lock "6fddc45c-d589-4423-a990-4be24d9dc84e" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 11.205s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1114.582674] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21b773cc-329d-4d9d-acd5-abc09c2fec97 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1115.092080] env[61923]: INFO nova.compute.manager [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] [instance: 6fddc45c-d589-4423-a990-4be24d9dc84e] During sync_power_state the instance has a pending task (resize_prep). Skip. [ 1115.092428] env[61923]: DEBUG oslo_concurrency.lockutils [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Lock "6fddc45c-d589-4423-a990-4be24d9dc84e" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.511s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1115.540451] env[61923]: DEBUG nova.compute.manager [None req-1b9fefdd-4058-437d-ac48-92ab2d363b47 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: 6fddc45c-d589-4423-a990-4be24d9dc84e] Stashing vm_state: active {{(pid=61923) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 1116.062134] env[61923]: DEBUG oslo_concurrency.lockutils [None req-1b9fefdd-4058-437d-ac48-92ab2d363b47 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1116.062492] env[61923]: DEBUG oslo_concurrency.lockutils [None req-1b9fefdd-4058-437d-ac48-92ab2d363b47 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 0.001s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1116.193062] env[61923]: DEBUG oslo_concurrency.lockutils [None req-6c4e6468-04fb-4b3f-95a4-880b7a5cf18d tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Acquiring lock "c0a8a116-c7b6-4adc-849b-e0a4faa150d3" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1116.193435] env[61923]: DEBUG oslo_concurrency.lockutils [None req-6c4e6468-04fb-4b3f-95a4-880b7a5cf18d tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Lock "c0a8a116-c7b6-4adc-849b-e0a4faa150d3" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.001s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1116.193563] env[61923]: DEBUG nova.compute.manager [None req-6c4e6468-04fb-4b3f-95a4-880b7a5cf18d tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: c0a8a116-c7b6-4adc-849b-e0a4faa150d3] Going to confirm migration 2 {{(pid=61923) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:4783}} [ 1116.567812] env[61923]: INFO nova.compute.claims [None req-1b9fefdd-4058-437d-ac48-92ab2d363b47 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: 6fddc45c-d589-4423-a990-4be24d9dc84e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1116.848341] env[61923]: DEBUG oslo_concurrency.lockutils [None req-6c4e6468-04fb-4b3f-95a4-880b7a5cf18d tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Acquiring lock "refresh_cache-c0a8a116-c7b6-4adc-849b-e0a4faa150d3" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1116.848577] env[61923]: DEBUG oslo_concurrency.lockutils [None req-6c4e6468-04fb-4b3f-95a4-880b7a5cf18d tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Acquired lock "refresh_cache-c0a8a116-c7b6-4adc-849b-e0a4faa150d3" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1116.848794] env[61923]: DEBUG nova.network.neutron [None req-6c4e6468-04fb-4b3f-95a4-880b7a5cf18d tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: c0a8a116-c7b6-4adc-849b-e0a4faa150d3] Building network info cache for instance {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1116.849040] env[61923]: DEBUG nova.objects.instance [None req-6c4e6468-04fb-4b3f-95a4-880b7a5cf18d tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Lazy-loading 'info_cache' on Instance uuid c0a8a116-c7b6-4adc-849b-e0a4faa150d3 {{(pid=61923) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1117.073991] env[61923]: INFO nova.compute.resource_tracker [None req-1b9fefdd-4058-437d-ac48-92ab2d363b47 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: 6fddc45c-d589-4423-a990-4be24d9dc84e] Updating resource usage from migration 63ebdf72-ea86-44f9-bf69-a031c88ccb35 [ 1117.130159] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4486d6fe-26e2-414d-8522-019ab9b3a52e {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.137786] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf525d13-0ccc-422e-873a-3d01dce45cba {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.166423] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1747b71e-07e0-429a-b467-1f108d4c124a {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.173473] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd2e08b0-71ed-4097-ac0e-bed9d8a8763b {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.186313] env[61923]: DEBUG nova.compute.provider_tree [None req-1b9fefdd-4058-437d-ac48-92ab2d363b47 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1117.689702] env[61923]: DEBUG nova.scheduler.client.report [None req-1b9fefdd-4058-437d-ac48-92ab2d363b47 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1118.046282] env[61923]: DEBUG nova.network.neutron [None req-6c4e6468-04fb-4b3f-95a4-880b7a5cf18d tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: c0a8a116-c7b6-4adc-849b-e0a4faa150d3] Updating instance_info_cache with network_info: [{"id": "c210c30f-4f3c-4698-ab24-a6a81c78e658", "address": "fa:16:3e:ad:1c:c4", "network": {"id": "9533e79d-fde2-4c10-86a0-86a36845a8cf", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-1736743137-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "780d3541d9604417b977bb62390c4299", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6d650b26-c3e7-4de7-98db-5e4b816d123a", "external-id": "nsx-vlan-transportzone-757", "segmentation_id": 757, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc210c30f-4f", "ovs_interfaceid": "c210c30f-4f3c-4698-ab24-a6a81c78e658", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1118.196653] env[61923]: DEBUG oslo_concurrency.lockutils [None req-1b9fefdd-4058-437d-ac48-92ab2d363b47 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.134s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1118.196866] env[61923]: INFO nova.compute.manager [None req-1b9fefdd-4058-437d-ac48-92ab2d363b47 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: 6fddc45c-d589-4423-a990-4be24d9dc84e] Migrating [ 1118.548732] env[61923]: DEBUG oslo_concurrency.lockutils [None req-6c4e6468-04fb-4b3f-95a4-880b7a5cf18d tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Releasing lock "refresh_cache-c0a8a116-c7b6-4adc-849b-e0a4faa150d3" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1118.549036] env[61923]: DEBUG nova.objects.instance [None req-6c4e6468-04fb-4b3f-95a4-880b7a5cf18d tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Lazy-loading 'migration_context' on Instance uuid c0a8a116-c7b6-4adc-849b-e0a4faa150d3 {{(pid=61923) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1118.711316] env[61923]: DEBUG oslo_concurrency.lockutils [None req-1b9fefdd-4058-437d-ac48-92ab2d363b47 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Acquiring lock "refresh_cache-6fddc45c-d589-4423-a990-4be24d9dc84e" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1118.711699] env[61923]: DEBUG oslo_concurrency.lockutils [None req-1b9fefdd-4058-437d-ac48-92ab2d363b47 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Acquired lock "refresh_cache-6fddc45c-d589-4423-a990-4be24d9dc84e" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1118.711804] env[61923]: DEBUG nova.network.neutron [None req-1b9fefdd-4058-437d-ac48-92ab2d363b47 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: 6fddc45c-d589-4423-a990-4be24d9dc84e] Building network info cache for instance {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1119.053150] env[61923]: DEBUG nova.objects.base [None req-6c4e6468-04fb-4b3f-95a4-880b7a5cf18d tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Object Instance lazy-loaded attributes: info_cache,migration_context {{(pid=61923) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 1119.054176] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a71d5a0-6541-4087-a922-60e5dca3cd2c {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1119.079344] env[61923]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-64f77b84-ccdc-47e7-aba3-2ea324f534f7 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1119.085122] env[61923]: DEBUG oslo_vmware.api [None req-6c4e6468-04fb-4b3f-95a4-880b7a5cf18d tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Waiting for the task: (returnval){ [ 1119.085122] env[61923]: value = "session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]527e0c38-bc78-a169-2333-388cb76b9da3" [ 1119.085122] env[61923]: _type = "Task" [ 1119.085122] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1119.092972] env[61923]: DEBUG oslo_vmware.api [None req-6c4e6468-04fb-4b3f-95a4-880b7a5cf18d tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]527e0c38-bc78-a169-2333-388cb76b9da3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1119.434100] env[61923]: DEBUG nova.network.neutron [None req-1b9fefdd-4058-437d-ac48-92ab2d363b47 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: 6fddc45c-d589-4423-a990-4be24d9dc84e] Updating instance_info_cache with network_info: [{"id": "cdf08c60-c555-4f7e-9a38-0b43dc6bd520", "address": "fa:16:3e:50:f7:08", "network": {"id": "77aec6ca-d1be-4b39-9d51-08853a1183a1", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1813085414-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5a4db4f8aa194a219e80c538cc715d43", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "56b944d8-803d-43f2-945d-0f334ee4ea1c", "external-id": "nsx-vlan-transportzone-799", "segmentation_id": 799, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcdf08c60-c5", "ovs_interfaceid": "cdf08c60-c555-4f7e-9a38-0b43dc6bd520", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1119.595699] env[61923]: DEBUG oslo_vmware.api [None req-6c4e6468-04fb-4b3f-95a4-880b7a5cf18d tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]527e0c38-bc78-a169-2333-388cb76b9da3, 'name': SearchDatastore_Task, 'duration_secs': 0.00928} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1119.595861] env[61923]: DEBUG oslo_concurrency.lockutils [None req-6c4e6468-04fb-4b3f-95a4-880b7a5cf18d tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1119.596014] env[61923]: DEBUG oslo_concurrency.lockutils [None req-6c4e6468-04fb-4b3f-95a4-880b7a5cf18d tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1119.937373] env[61923]: DEBUG oslo_concurrency.lockutils [None req-1b9fefdd-4058-437d-ac48-92ab2d363b47 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Releasing lock "refresh_cache-6fddc45c-d589-4423-a990-4be24d9dc84e" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1120.158261] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a37d0849-fc40-4e64-b1d7-c9f11e2cfb74 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.165647] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4e9c8df-f238-4213-8d95-c085406b8235 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.195632] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31863270-1e14-4111-99cd-0c349a3f21fd {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.202203] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17de3003-a117-4543-94d7-cb030f3102d9 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.215107] env[61923]: DEBUG nova.compute.provider_tree [None req-6c4e6468-04fb-4b3f-95a4-880b7a5cf18d tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1120.718135] env[61923]: DEBUG nova.scheduler.client.report [None req-6c4e6468-04fb-4b3f-95a4-880b7a5cf18d tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1121.452165] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f4ff33e1-07d2-4f74-93ef-a8978e3dc0bd {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.472015] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-1b9fefdd-4058-437d-ac48-92ab2d363b47 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: 6fddc45c-d589-4423-a990-4be24d9dc84e] Updating instance '6fddc45c-d589-4423-a990-4be24d9dc84e' progress to 0 {{(pid=61923) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1121.728343] env[61923]: DEBUG oslo_concurrency.lockutils [None req-6c4e6468-04fb-4b3f-95a4-880b7a5cf18d tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.132s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1121.978254] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-1b9fefdd-4058-437d-ac48-92ab2d363b47 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: 6fddc45c-d589-4423-a990-4be24d9dc84e] Powering off the VM {{(pid=61923) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1121.978637] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9f5d5251-8e95-4041-ba86-3aa1ce0b69f0 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1121.985992] env[61923]: DEBUG oslo_vmware.api [None req-1b9fefdd-4058-437d-ac48-92ab2d363b47 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Waiting for the task: (returnval){ [ 1121.985992] env[61923]: value = "task-1378053" [ 1121.985992] env[61923]: _type = "Task" [ 1121.985992] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1121.993882] env[61923]: DEBUG oslo_vmware.api [None req-1b9fefdd-4058-437d-ac48-92ab2d363b47 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Task: {'id': task-1378053, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1122.290820] env[61923]: INFO nova.scheduler.client.report [None req-6c4e6468-04fb-4b3f-95a4-880b7a5cf18d tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Deleted allocation for migration af6ff10e-1948-46aa-b5c5-bbe410bf3124 [ 1122.496505] env[61923]: DEBUG oslo_vmware.api [None req-1b9fefdd-4058-437d-ac48-92ab2d363b47 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Task: {'id': task-1378053, 'name': PowerOffVM_Task, 'duration_secs': 0.193622} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1122.496858] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-1b9fefdd-4058-437d-ac48-92ab2d363b47 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: 6fddc45c-d589-4423-a990-4be24d9dc84e] Powered off the VM {{(pid=61923) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1122.496946] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-1b9fefdd-4058-437d-ac48-92ab2d363b47 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: 6fddc45c-d589-4423-a990-4be24d9dc84e] Updating instance '6fddc45c-d589-4423-a990-4be24d9dc84e' progress to 17 {{(pid=61923) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1122.797249] env[61923]: DEBUG oslo_concurrency.lockutils [None req-6c4e6468-04fb-4b3f-95a4-880b7a5cf18d tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Lock "c0a8a116-c7b6-4adc-849b-e0a4faa150d3" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 6.604s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1123.003964] env[61923]: DEBUG nova.virt.hardware [None req-1b9fefdd-4058-437d-ac48-92ab2d363b47 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-29T20:07:54Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=0f153f63-ae0a-45b1-b7f5-7f9b673c947f,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1123.004249] env[61923]: DEBUG nova.virt.hardware [None req-1b9fefdd-4058-437d-ac48-92ab2d363b47 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Flavor limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1123.004415] env[61923]: DEBUG nova.virt.hardware [None req-1b9fefdd-4058-437d-ac48-92ab2d363b47 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Image limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1123.004601] env[61923]: DEBUG nova.virt.hardware [None req-1b9fefdd-4058-437d-ac48-92ab2d363b47 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Flavor pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1123.004751] env[61923]: DEBUG nova.virt.hardware [None req-1b9fefdd-4058-437d-ac48-92ab2d363b47 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Image pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1123.004899] env[61923]: DEBUG nova.virt.hardware [None req-1b9fefdd-4058-437d-ac48-92ab2d363b47 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1123.005114] env[61923]: DEBUG nova.virt.hardware [None req-1b9fefdd-4058-437d-ac48-92ab2d363b47 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1123.005280] env[61923]: DEBUG nova.virt.hardware [None req-1b9fefdd-4058-437d-ac48-92ab2d363b47 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1123.005447] env[61923]: DEBUG nova.virt.hardware [None req-1b9fefdd-4058-437d-ac48-92ab2d363b47 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Got 1 possible topologies {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1123.005612] env[61923]: DEBUG nova.virt.hardware [None req-1b9fefdd-4058-437d-ac48-92ab2d363b47 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1123.005785] env[61923]: DEBUG nova.virt.hardware [None req-1b9fefdd-4058-437d-ac48-92ab2d363b47 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1123.010752] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-73586ec4-e2d5-496a-add5-f930c8af814b {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1123.028077] env[61923]: DEBUG oslo_vmware.api [None req-1b9fefdd-4058-437d-ac48-92ab2d363b47 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Waiting for the task: (returnval){ [ 1123.028077] env[61923]: value = "task-1378054" [ 1123.028077] env[61923]: _type = "Task" [ 1123.028077] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1123.035612] env[61923]: DEBUG oslo_vmware.api [None req-1b9fefdd-4058-437d-ac48-92ab2d363b47 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Task: {'id': task-1378054, 'name': ReconfigVM_Task} progress is 6%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1123.538189] env[61923]: DEBUG oslo_vmware.api [None req-1b9fefdd-4058-437d-ac48-92ab2d363b47 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Task: {'id': task-1378054, 'name': ReconfigVM_Task, 'duration_secs': 0.124462} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1123.538506] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-1b9fefdd-4058-437d-ac48-92ab2d363b47 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: 6fddc45c-d589-4423-a990-4be24d9dc84e] Updating instance '6fddc45c-d589-4423-a990-4be24d9dc84e' progress to 33 {{(pid=61923) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1124.044702] env[61923]: DEBUG nova.virt.hardware [None req-1b9fefdd-4058-437d-ac48-92ab2d363b47 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-29T20:07:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=0f153f63-ae0a-45b1-b7f5-7f9b673c947f,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1124.045031] env[61923]: DEBUG nova.virt.hardware [None req-1b9fefdd-4058-437d-ac48-92ab2d363b47 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Flavor limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1124.045120] env[61923]: DEBUG nova.virt.hardware [None req-1b9fefdd-4058-437d-ac48-92ab2d363b47 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Image limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1124.045310] env[61923]: DEBUG nova.virt.hardware [None req-1b9fefdd-4058-437d-ac48-92ab2d363b47 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Flavor pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1124.045539] env[61923]: DEBUG nova.virt.hardware [None req-1b9fefdd-4058-437d-ac48-92ab2d363b47 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Image pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1124.045736] env[61923]: DEBUG nova.virt.hardware [None req-1b9fefdd-4058-437d-ac48-92ab2d363b47 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1124.045958] env[61923]: DEBUG nova.virt.hardware [None req-1b9fefdd-4058-437d-ac48-92ab2d363b47 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1124.046139] env[61923]: DEBUG nova.virt.hardware [None req-1b9fefdd-4058-437d-ac48-92ab2d363b47 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1124.046314] env[61923]: DEBUG nova.virt.hardware [None req-1b9fefdd-4058-437d-ac48-92ab2d363b47 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Got 1 possible topologies {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1124.046481] env[61923]: DEBUG nova.virt.hardware [None req-1b9fefdd-4058-437d-ac48-92ab2d363b47 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1124.046658] env[61923]: DEBUG nova.virt.hardware [None req-1b9fefdd-4058-437d-ac48-92ab2d363b47 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1124.052176] env[61923]: DEBUG nova.virt.vmwareapi.volumeops [None req-1b9fefdd-4058-437d-ac48-92ab2d363b47 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: 6fddc45c-d589-4423-a990-4be24d9dc84e] Reconfiguring VM instance instance-00000066 to detach disk 2000 {{(pid=61923) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1124.052470] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e35e1a3b-85fc-4698-bafb-434a7d9a8ef4 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.071204] env[61923]: DEBUG oslo_vmware.api [None req-1b9fefdd-4058-437d-ac48-92ab2d363b47 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Waiting for the task: (returnval){ [ 1124.071204] env[61923]: value = "task-1378055" [ 1124.071204] env[61923]: _type = "Task" [ 1124.071204] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1124.079069] env[61923]: DEBUG oslo_vmware.api [None req-1b9fefdd-4058-437d-ac48-92ab2d363b47 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Task: {'id': task-1378055, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1124.213944] env[61923]: DEBUG oslo_concurrency.lockutils [None req-6c4e6468-04fb-4b3f-95a4-880b7a5cf18d tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Acquiring lock "c0a8a116-c7b6-4adc-849b-e0a4faa150d3" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1124.214308] env[61923]: DEBUG oslo_concurrency.lockutils [None req-6c4e6468-04fb-4b3f-95a4-880b7a5cf18d tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Lock "c0a8a116-c7b6-4adc-849b-e0a4faa150d3" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1124.214532] env[61923]: DEBUG oslo_concurrency.lockutils [None req-6c4e6468-04fb-4b3f-95a4-880b7a5cf18d tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Acquiring lock "c0a8a116-c7b6-4adc-849b-e0a4faa150d3-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1124.214724] env[61923]: DEBUG oslo_concurrency.lockutils [None req-6c4e6468-04fb-4b3f-95a4-880b7a5cf18d tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Lock "c0a8a116-c7b6-4adc-849b-e0a4faa150d3-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1124.214904] env[61923]: DEBUG oslo_concurrency.lockutils [None req-6c4e6468-04fb-4b3f-95a4-880b7a5cf18d tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Lock "c0a8a116-c7b6-4adc-849b-e0a4faa150d3-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1124.217342] env[61923]: INFO nova.compute.manager [None req-6c4e6468-04fb-4b3f-95a4-880b7a5cf18d tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: c0a8a116-c7b6-4adc-849b-e0a4faa150d3] Terminating instance [ 1124.219265] env[61923]: DEBUG nova.compute.manager [None req-6c4e6468-04fb-4b3f-95a4-880b7a5cf18d tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: c0a8a116-c7b6-4adc-849b-e0a4faa150d3] Start destroying the instance on the hypervisor. {{(pid=61923) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1124.219469] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-6c4e6468-04fb-4b3f-95a4-880b7a5cf18d tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: c0a8a116-c7b6-4adc-849b-e0a4faa150d3] Destroying instance {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1124.220335] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e53e83fd-5573-4723-b894-052c9a62048a {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.227278] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-6c4e6468-04fb-4b3f-95a4-880b7a5cf18d tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: c0a8a116-c7b6-4adc-849b-e0a4faa150d3] Powering off the VM {{(pid=61923) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1124.227512] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-72e60752-6132-40d3-a123-2a0ce7ed0b2c {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.233708] env[61923]: DEBUG oslo_vmware.api [None req-6c4e6468-04fb-4b3f-95a4-880b7a5cf18d tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Waiting for the task: (returnval){ [ 1124.233708] env[61923]: value = "task-1378056" [ 1124.233708] env[61923]: _type = "Task" [ 1124.233708] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1124.243692] env[61923]: DEBUG oslo_vmware.api [None req-6c4e6468-04fb-4b3f-95a4-880b7a5cf18d tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Task: {'id': task-1378056, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1124.580010] env[61923]: DEBUG oslo_vmware.api [None req-1b9fefdd-4058-437d-ac48-92ab2d363b47 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Task: {'id': task-1378055, 'name': ReconfigVM_Task, 'duration_secs': 0.163019} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1124.580327] env[61923]: DEBUG nova.virt.vmwareapi.volumeops [None req-1b9fefdd-4058-437d-ac48-92ab2d363b47 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: 6fddc45c-d589-4423-a990-4be24d9dc84e] Reconfigured VM instance instance-00000066 to detach disk 2000 {{(pid=61923) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1124.581093] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-103f2be7-fda8-4905-82fd-5cc8ce628245 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.602410] env[61923]: DEBUG nova.virt.vmwareapi.volumeops [None req-1b9fefdd-4058-437d-ac48-92ab2d363b47 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: 6fddc45c-d589-4423-a990-4be24d9dc84e] Reconfiguring VM instance instance-00000066 to attach disk [datastore1] 6fddc45c-d589-4423-a990-4be24d9dc84e/6fddc45c-d589-4423-a990-4be24d9dc84e.vmdk or device None with type thin {{(pid=61923) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1124.602651] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-032980f5-9027-4ffe-8a59-d2b774775d02 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.619343] env[61923]: DEBUG oslo_vmware.api [None req-1b9fefdd-4058-437d-ac48-92ab2d363b47 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Waiting for the task: (returnval){ [ 1124.619343] env[61923]: value = "task-1378057" [ 1124.619343] env[61923]: _type = "Task" [ 1124.619343] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1124.626450] env[61923]: DEBUG oslo_vmware.api [None req-1b9fefdd-4058-437d-ac48-92ab2d363b47 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Task: {'id': task-1378057, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1124.743958] env[61923]: DEBUG oslo_vmware.api [None req-6c4e6468-04fb-4b3f-95a4-880b7a5cf18d tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Task: {'id': task-1378056, 'name': PowerOffVM_Task, 'duration_secs': 0.188444} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1124.744206] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-6c4e6468-04fb-4b3f-95a4-880b7a5cf18d tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: c0a8a116-c7b6-4adc-849b-e0a4faa150d3] Powered off the VM {{(pid=61923) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1124.744382] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-6c4e6468-04fb-4b3f-95a4-880b7a5cf18d tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: c0a8a116-c7b6-4adc-849b-e0a4faa150d3] Unregistering the VM {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1124.744632] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e31137c8-c90f-4013-976f-6f7c8b995c9a {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.810548] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-6c4e6468-04fb-4b3f-95a4-880b7a5cf18d tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: c0a8a116-c7b6-4adc-849b-e0a4faa150d3] Unregistered the VM {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1124.810777] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-6c4e6468-04fb-4b3f-95a4-880b7a5cf18d tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: c0a8a116-c7b6-4adc-849b-e0a4faa150d3] Deleting contents of the VM from datastore datastore2 {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1124.810982] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-6c4e6468-04fb-4b3f-95a4-880b7a5cf18d tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Deleting the datastore file [datastore2] c0a8a116-c7b6-4adc-849b-e0a4faa150d3 {{(pid=61923) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1124.811256] env[61923]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-79ce8a0e-018b-4947-bc27-58c99698f40f {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.817744] env[61923]: DEBUG oslo_vmware.api [None req-6c4e6468-04fb-4b3f-95a4-880b7a5cf18d tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Waiting for the task: (returnval){ [ 1124.817744] env[61923]: value = "task-1378059" [ 1124.817744] env[61923]: _type = "Task" [ 1124.817744] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1124.825215] env[61923]: DEBUG oslo_vmware.api [None req-6c4e6468-04fb-4b3f-95a4-880b7a5cf18d tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Task: {'id': task-1378059, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1125.129074] env[61923]: DEBUG oslo_vmware.api [None req-1b9fefdd-4058-437d-ac48-92ab2d363b47 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Task: {'id': task-1378057, 'name': ReconfigVM_Task, 'duration_secs': 0.251521} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1125.129371] env[61923]: DEBUG nova.virt.vmwareapi.volumeops [None req-1b9fefdd-4058-437d-ac48-92ab2d363b47 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: 6fddc45c-d589-4423-a990-4be24d9dc84e] Reconfigured VM instance instance-00000066 to attach disk [datastore1] 6fddc45c-d589-4423-a990-4be24d9dc84e/6fddc45c-d589-4423-a990-4be24d9dc84e.vmdk or device None with type thin {{(pid=61923) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1125.129653] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-1b9fefdd-4058-437d-ac48-92ab2d363b47 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: 6fddc45c-d589-4423-a990-4be24d9dc84e] Updating instance '6fddc45c-d589-4423-a990-4be24d9dc84e' progress to 50 {{(pid=61923) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1125.327986] env[61923]: DEBUG oslo_vmware.api [None req-6c4e6468-04fb-4b3f-95a4-880b7a5cf18d tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Task: {'id': task-1378059, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.140213} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1125.328234] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-6c4e6468-04fb-4b3f-95a4-880b7a5cf18d tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Deleted the datastore file {{(pid=61923) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1125.328427] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-6c4e6468-04fb-4b3f-95a4-880b7a5cf18d tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: c0a8a116-c7b6-4adc-849b-e0a4faa150d3] Deleted contents of the VM from datastore datastore2 {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1125.328608] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-6c4e6468-04fb-4b3f-95a4-880b7a5cf18d tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: c0a8a116-c7b6-4adc-849b-e0a4faa150d3] Instance destroyed {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1125.328790] env[61923]: INFO nova.compute.manager [None req-6c4e6468-04fb-4b3f-95a4-880b7a5cf18d tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] [instance: c0a8a116-c7b6-4adc-849b-e0a4faa150d3] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1125.329050] env[61923]: DEBUG oslo.service.loopingcall [None req-6c4e6468-04fb-4b3f-95a4-880b7a5cf18d tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61923) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1125.329247] env[61923]: DEBUG nova.compute.manager [-] [instance: c0a8a116-c7b6-4adc-849b-e0a4faa150d3] Deallocating network for instance {{(pid=61923) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1125.329341] env[61923]: DEBUG nova.network.neutron [-] [instance: c0a8a116-c7b6-4adc-849b-e0a4faa150d3] deallocate_for_instance() {{(pid=61923) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1125.592141] env[61923]: DEBUG nova.compute.manager [req-ff171d76-068a-4587-b764-8b4e5af7ae74 req-30bf4610-c506-4c1b-8fb1-56d2fd9eb579 service nova] [instance: c0a8a116-c7b6-4adc-849b-e0a4faa150d3] Received event network-vif-deleted-c210c30f-4f3c-4698-ab24-a6a81c78e658 {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1125.592377] env[61923]: INFO nova.compute.manager [req-ff171d76-068a-4587-b764-8b4e5af7ae74 req-30bf4610-c506-4c1b-8fb1-56d2fd9eb579 service nova] [instance: c0a8a116-c7b6-4adc-849b-e0a4faa150d3] Neutron deleted interface c210c30f-4f3c-4698-ab24-a6a81c78e658; detaching it from the instance and deleting it from the info cache [ 1125.592538] env[61923]: DEBUG nova.network.neutron [req-ff171d76-068a-4587-b764-8b4e5af7ae74 req-30bf4610-c506-4c1b-8fb1-56d2fd9eb579 service nova] [instance: c0a8a116-c7b6-4adc-849b-e0a4faa150d3] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1125.636426] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bad2d632-1194-42bf-95a8-d169237e7ddf {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.654624] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f3b444b-5058-4208-876b-b36ad1ac46d6 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.672828] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-1b9fefdd-4058-437d-ac48-92ab2d363b47 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: 6fddc45c-d589-4423-a990-4be24d9dc84e] Updating instance '6fddc45c-d589-4423-a990-4be24d9dc84e' progress to 67 {{(pid=61923) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1126.063435] env[61923]: DEBUG nova.network.neutron [-] [instance: c0a8a116-c7b6-4adc-849b-e0a4faa150d3] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1126.095473] env[61923]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-997fd36e-1b78-4b11-b1b9-0edec86ee65e {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.105246] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa1819f9-dd4f-41a3-8f38-7f803f301f84 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.127367] env[61923]: DEBUG nova.compute.manager [req-ff171d76-068a-4587-b764-8b4e5af7ae74 req-30bf4610-c506-4c1b-8fb1-56d2fd9eb579 service nova] [instance: c0a8a116-c7b6-4adc-849b-e0a4faa150d3] Detach interface failed, port_id=c210c30f-4f3c-4698-ab24-a6a81c78e658, reason: Instance c0a8a116-c7b6-4adc-849b-e0a4faa150d3 could not be found. {{(pid=61923) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1126.221892] env[61923]: DEBUG nova.network.neutron [None req-1b9fefdd-4058-437d-ac48-92ab2d363b47 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: 6fddc45c-d589-4423-a990-4be24d9dc84e] Port cdf08c60-c555-4f7e-9a38-0b43dc6bd520 binding to destination host cpu-1 is already ACTIVE {{(pid=61923) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 1126.566599] env[61923]: INFO nova.compute.manager [-] [instance: c0a8a116-c7b6-4adc-849b-e0a4faa150d3] Took 1.24 seconds to deallocate network for instance. [ 1127.073647] env[61923]: DEBUG oslo_concurrency.lockutils [None req-6c4e6468-04fb-4b3f-95a4-880b7a5cf18d tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1127.073925] env[61923]: DEBUG oslo_concurrency.lockutils [None req-6c4e6468-04fb-4b3f-95a4-880b7a5cf18d tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1127.074135] env[61923]: DEBUG oslo_concurrency.lockutils [None req-6c4e6468-04fb-4b3f-95a4-880b7a5cf18d tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1127.094972] env[61923]: INFO nova.scheduler.client.report [None req-6c4e6468-04fb-4b3f-95a4-880b7a5cf18d tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Deleted allocations for instance c0a8a116-c7b6-4adc-849b-e0a4faa150d3 [ 1127.242585] env[61923]: DEBUG oslo_concurrency.lockutils [None req-1b9fefdd-4058-437d-ac48-92ab2d363b47 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Acquiring lock "6fddc45c-d589-4423-a990-4be24d9dc84e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1127.242775] env[61923]: DEBUG oslo_concurrency.lockutils [None req-1b9fefdd-4058-437d-ac48-92ab2d363b47 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Lock "6fddc45c-d589-4423-a990-4be24d9dc84e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1127.242939] env[61923]: DEBUG oslo_concurrency.lockutils [None req-1b9fefdd-4058-437d-ac48-92ab2d363b47 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Lock "6fddc45c-d589-4423-a990-4be24d9dc84e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1127.604600] env[61923]: DEBUG oslo_concurrency.lockutils [None req-6c4e6468-04fb-4b3f-95a4-880b7a5cf18d tempest-DeleteServersTestJSON-2016794872 tempest-DeleteServersTestJSON-2016794872-project-member] Lock "c0a8a116-c7b6-4adc-849b-e0a4faa150d3" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 3.390s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1128.279975] env[61923]: DEBUG oslo_concurrency.lockutils [None req-1b9fefdd-4058-437d-ac48-92ab2d363b47 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Acquiring lock "refresh_cache-6fddc45c-d589-4423-a990-4be24d9dc84e" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1128.280191] env[61923]: DEBUG oslo_concurrency.lockutils [None req-1b9fefdd-4058-437d-ac48-92ab2d363b47 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Acquired lock "refresh_cache-6fddc45c-d589-4423-a990-4be24d9dc84e" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1128.280361] env[61923]: DEBUG nova.network.neutron [None req-1b9fefdd-4058-437d-ac48-92ab2d363b47 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: 6fddc45c-d589-4423-a990-4be24d9dc84e] Building network info cache for instance {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1128.992193] env[61923]: DEBUG nova.network.neutron [None req-1b9fefdd-4058-437d-ac48-92ab2d363b47 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: 6fddc45c-d589-4423-a990-4be24d9dc84e] Updating instance_info_cache with network_info: [{"id": "cdf08c60-c555-4f7e-9a38-0b43dc6bd520", "address": "fa:16:3e:50:f7:08", "network": {"id": "77aec6ca-d1be-4b39-9d51-08853a1183a1", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1813085414-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5a4db4f8aa194a219e80c538cc715d43", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "56b944d8-803d-43f2-945d-0f334ee4ea1c", "external-id": "nsx-vlan-transportzone-799", "segmentation_id": 799, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcdf08c60-c5", "ovs_interfaceid": "cdf08c60-c555-4f7e-9a38-0b43dc6bd520", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1129.495128] env[61923]: DEBUG oslo_concurrency.lockutils [None req-1b9fefdd-4058-437d-ac48-92ab2d363b47 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Releasing lock "refresh_cache-6fddc45c-d589-4423-a990-4be24d9dc84e" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1130.019475] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98ab6bcf-e0b4-4fa7-b9e7-3d55493d6b6a {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.037890] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e929ceb7-5819-4b34-bc04-40635b3d7b08 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.044432] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-1b9fefdd-4058-437d-ac48-92ab2d363b47 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: 6fddc45c-d589-4423-a990-4be24d9dc84e] Updating instance '6fddc45c-d589-4423-a990-4be24d9dc84e' progress to 83 {{(pid=61923) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1130.551009] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-1b9fefdd-4058-437d-ac48-92ab2d363b47 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: 6fddc45c-d589-4423-a990-4be24d9dc84e] Powering on the VM {{(pid=61923) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1130.551342] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-41480b97-4397-4dbe-b56f-c57cdabf2669 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1130.558813] env[61923]: DEBUG oslo_vmware.api [None req-1b9fefdd-4058-437d-ac48-92ab2d363b47 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Waiting for the task: (returnval){ [ 1130.558813] env[61923]: value = "task-1378061" [ 1130.558813] env[61923]: _type = "Task" [ 1130.558813] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1130.566079] env[61923]: DEBUG oslo_vmware.api [None req-1b9fefdd-4058-437d-ac48-92ab2d363b47 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Task: {'id': task-1378061, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1131.068088] env[61923]: DEBUG oslo_vmware.api [None req-1b9fefdd-4058-437d-ac48-92ab2d363b47 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Task: {'id': task-1378061, 'name': PowerOnVM_Task, 'duration_secs': 0.358304} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1131.068377] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-1b9fefdd-4058-437d-ac48-92ab2d363b47 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: 6fddc45c-d589-4423-a990-4be24d9dc84e] Powered on the VM {{(pid=61923) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1131.068568] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-1b9fefdd-4058-437d-ac48-92ab2d363b47 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: 6fddc45c-d589-4423-a990-4be24d9dc84e] Updating instance '6fddc45c-d589-4423-a990-4be24d9dc84e' progress to 100 {{(pid=61923) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1132.771439] env[61923]: DEBUG oslo_concurrency.lockutils [None req-8370692d-cdea-49c1-b897-10ea8a6e5367 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Acquiring lock "6fddc45c-d589-4423-a990-4be24d9dc84e" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1132.771748] env[61923]: DEBUG oslo_concurrency.lockutils [None req-8370692d-cdea-49c1-b897-10ea8a6e5367 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Lock "6fddc45c-d589-4423-a990-4be24d9dc84e" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.001s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1132.771748] env[61923]: DEBUG nova.compute.manager [None req-8370692d-cdea-49c1-b897-10ea8a6e5367 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: 6fddc45c-d589-4423-a990-4be24d9dc84e] Going to confirm migration 3 {{(pid=61923) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:4783}} [ 1132.874045] env[61923]: DEBUG oslo_service.periodic_task [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61923) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1132.874371] env[61923]: DEBUG nova.compute.manager [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Starting heal instance info cache {{(pid=61923) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 1133.351917] env[61923]: DEBUG oslo_concurrency.lockutils [None req-8370692d-cdea-49c1-b897-10ea8a6e5367 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Acquiring lock "refresh_cache-6fddc45c-d589-4423-a990-4be24d9dc84e" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1133.352286] env[61923]: DEBUG oslo_concurrency.lockutils [None req-8370692d-cdea-49c1-b897-10ea8a6e5367 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Acquired lock "refresh_cache-6fddc45c-d589-4423-a990-4be24d9dc84e" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1133.352503] env[61923]: DEBUG nova.network.neutron [None req-8370692d-cdea-49c1-b897-10ea8a6e5367 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: 6fddc45c-d589-4423-a990-4be24d9dc84e] Building network info cache for instance {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1133.352703] env[61923]: DEBUG nova.objects.instance [None req-8370692d-cdea-49c1-b897-10ea8a6e5367 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Lazy-loading 'info_cache' on Instance uuid 6fddc45c-d589-4423-a990-4be24d9dc84e {{(pid=61923) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1133.377906] env[61923]: DEBUG nova.compute.manager [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Didn't find any instances for network info cache update. {{(pid=61923) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10014}} [ 1133.377998] env[61923]: DEBUG oslo_service.periodic_task [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61923) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1133.378292] env[61923]: DEBUG oslo_service.periodic_task [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61923) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1133.378524] env[61923]: DEBUG oslo_service.periodic_task [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61923) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1133.378742] env[61923]: DEBUG nova.compute.manager [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61923) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 1134.355281] env[61923]: DEBUG oslo_service.periodic_task [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61923) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1134.569887] env[61923]: DEBUG nova.network.neutron [None req-8370692d-cdea-49c1-b897-10ea8a6e5367 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: 6fddc45c-d589-4423-a990-4be24d9dc84e] Updating instance_info_cache with network_info: [{"id": "cdf08c60-c555-4f7e-9a38-0b43dc6bd520", "address": "fa:16:3e:50:f7:08", "network": {"id": "77aec6ca-d1be-4b39-9d51-08853a1183a1", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1813085414-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5a4db4f8aa194a219e80c538cc715d43", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "56b944d8-803d-43f2-945d-0f334ee4ea1c", "external-id": "nsx-vlan-transportzone-799", "segmentation_id": 799, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapcdf08c60-c5", "ovs_interfaceid": "cdf08c60-c555-4f7e-9a38-0b43dc6bd520", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1135.074398] env[61923]: DEBUG oslo_concurrency.lockutils [None req-8370692d-cdea-49c1-b897-10ea8a6e5367 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Releasing lock "refresh_cache-6fddc45c-d589-4423-a990-4be24d9dc84e" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1135.074667] env[61923]: DEBUG nova.objects.instance [None req-8370692d-cdea-49c1-b897-10ea8a6e5367 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Lazy-loading 'migration_context' on Instance uuid 6fddc45c-d589-4423-a990-4be24d9dc84e {{(pid=61923) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1135.355339] env[61923]: DEBUG oslo_service.periodic_task [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61923) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1135.355720] env[61923]: DEBUG oslo_service.periodic_task [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61923) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1135.577232] env[61923]: DEBUG nova.objects.base [None req-8370692d-cdea-49c1-b897-10ea8a6e5367 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Object Instance<6fddc45c-d589-4423-a990-4be24d9dc84e> lazy-loaded attributes: info_cache,migration_context {{(pid=61923) wrapper /opt/stack/nova/nova/objects/base.py:136}} [ 1135.578238] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fcc9a70e-4296-4edb-90b4-8faf598f479a {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.603275] env[61923]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9d657d19-9f1b-49ee-8756-51d60e4be13f {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.608608] env[61923]: DEBUG oslo_vmware.api [None req-8370692d-cdea-49c1-b897-10ea8a6e5367 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Waiting for the task: (returnval){ [ 1135.608608] env[61923]: value = "session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52f7ea6b-edf3-02d8-a81b-1485d0dfdb86" [ 1135.608608] env[61923]: _type = "Task" [ 1135.608608] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1135.616296] env[61923]: DEBUG oslo_vmware.api [None req-8370692d-cdea-49c1-b897-10ea8a6e5367 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52f7ea6b-edf3-02d8-a81b-1485d0dfdb86, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1135.931611] env[61923]: DEBUG oslo_concurrency.lockutils [None req-a84572db-25cb-45bf-8bab-bc3f62d5ec23 tempest-ServerMetadataTestJSON-1639770617 tempest-ServerMetadataTestJSON-1639770617-project-member] Acquiring lock "b82a241e-db60-40bc-a738-b961c3e38b43" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1135.931874] env[61923]: DEBUG oslo_concurrency.lockutils [None req-a84572db-25cb-45bf-8bab-bc3f62d5ec23 tempest-ServerMetadataTestJSON-1639770617 tempest-ServerMetadataTestJSON-1639770617-project-member] Lock "b82a241e-db60-40bc-a738-b961c3e38b43" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1136.119661] env[61923]: DEBUG oslo_vmware.api [None req-8370692d-cdea-49c1-b897-10ea8a6e5367 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52f7ea6b-edf3-02d8-a81b-1485d0dfdb86, 'name': SearchDatastore_Task, 'duration_secs': 0.006894} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1136.119987] env[61923]: DEBUG oslo_concurrency.lockutils [None req-8370692d-cdea-49c1-b897-10ea8a6e5367 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1136.120254] env[61923]: DEBUG oslo_concurrency.lockutils [None req-8370692d-cdea-49c1-b897-10ea8a6e5367 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1136.354783] env[61923]: DEBUG oslo_service.periodic_task [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Running periodic task ComputeManager.update_available_resource {{(pid=61923) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1136.434341] env[61923]: DEBUG nova.compute.manager [None req-a84572db-25cb-45bf-8bab-bc3f62d5ec23 tempest-ServerMetadataTestJSON-1639770617 tempest-ServerMetadataTestJSON-1639770617-project-member] [instance: b82a241e-db60-40bc-a738-b961c3e38b43] Starting instance... {{(pid=61923) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1136.682818] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce8a0b84-55eb-4c31-9201-8a8ccfe0591b {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1136.690097] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b8d0844-b531-4ac3-8ec6-0f8142dd44b4 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1136.719083] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da3160ac-c432-4d67-97e4-fa6e60de27ef {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1136.726418] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81c869fe-79e5-48cb-b01e-55946785d2c1 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1136.740394] env[61923]: DEBUG nova.compute.provider_tree [None req-8370692d-cdea-49c1-b897-10ea8a6e5367 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1136.857743] env[61923]: DEBUG oslo_concurrency.lockutils [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1136.952508] env[61923]: DEBUG oslo_concurrency.lockutils [None req-a84572db-25cb-45bf-8bab-bc3f62d5ec23 tempest-ServerMetadataTestJSON-1639770617 tempest-ServerMetadataTestJSON-1639770617-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1137.243557] env[61923]: DEBUG nova.scheduler.client.report [None req-8370692d-cdea-49c1-b897-10ea8a6e5367 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1138.254013] env[61923]: DEBUG oslo_concurrency.lockutils [None req-8370692d-cdea-49c1-b897-10ea8a6e5367 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.133s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1138.257221] env[61923]: DEBUG oslo_concurrency.lockutils [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 1.400s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1138.257370] env[61923]: DEBUG oslo_concurrency.lockutils [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1138.257575] env[61923]: DEBUG nova.compute.resource_tracker [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61923) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1138.257966] env[61923]: DEBUG oslo_concurrency.lockutils [None req-a84572db-25cb-45bf-8bab-bc3f62d5ec23 tempest-ServerMetadataTestJSON-1639770617 tempest-ServerMetadataTestJSON-1639770617-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.306s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1138.259276] env[61923]: INFO nova.compute.claims [None req-a84572db-25cb-45bf-8bab-bc3f62d5ec23 tempest-ServerMetadataTestJSON-1639770617 tempest-ServerMetadataTestJSON-1639770617-project-member] [instance: b82a241e-db60-40bc-a738-b961c3e38b43] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1138.262243] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d7f7357-eb7f-4a9b-b0cf-1b65bcd071bf {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1138.270421] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a9d98b6-65da-4f3a-9ba5-dec764ba6aba {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1138.284225] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1da22d8a-88af-4d46-9aa7-010e1d1015e2 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1138.290610] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5be0befa-9ede-429c-af94-06c77a6bb70b {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1138.320350] env[61923]: DEBUG nova.compute.resource_tracker [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181168MB free_disk=178GB free_vcpus=48 pci_devices=None {{(pid=61923) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1138.320509] env[61923]: DEBUG oslo_concurrency.lockutils [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1138.820677] env[61923]: INFO nova.scheduler.client.report [None req-8370692d-cdea-49c1-b897-10ea8a6e5367 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Deleted allocation for migration 63ebdf72-ea86-44f9-bf69-a031c88ccb35 [ 1139.307140] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1468c0e-d3b1-4aaa-abf9-0a843df95fe4 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.314615] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59b28100-f4eb-48a7-b0a6-81aa2b76d5df {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.346302] env[61923]: DEBUG oslo_concurrency.lockutils [None req-8370692d-cdea-49c1-b897-10ea8a6e5367 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Lock "6fddc45c-d589-4423-a990-4be24d9dc84e" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 6.575s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1139.349497] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7945fc8c-5fd2-410f-a800-b7a8d46d7e2d {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.361163] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26aece49-e726-47bd-bbbf-4cda7a532c5c {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.375882] env[61923]: DEBUG nova.compute.provider_tree [None req-a84572db-25cb-45bf-8bab-bc3f62d5ec23 tempest-ServerMetadataTestJSON-1639770617 tempest-ServerMetadataTestJSON-1639770617-project-member] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1139.855817] env[61923]: DEBUG oslo_concurrency.lockutils [None req-72bb36a2-c5ce-4bb4-b8a6-7d89be6fdabb tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Acquiring lock "6fddc45c-d589-4423-a990-4be24d9dc84e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1139.856079] env[61923]: DEBUG oslo_concurrency.lockutils [None req-72bb36a2-c5ce-4bb4-b8a6-7d89be6fdabb tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Lock "6fddc45c-d589-4423-a990-4be24d9dc84e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1139.856318] env[61923]: DEBUG oslo_concurrency.lockutils [None req-72bb36a2-c5ce-4bb4-b8a6-7d89be6fdabb tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Acquiring lock "6fddc45c-d589-4423-a990-4be24d9dc84e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1139.856525] env[61923]: DEBUG oslo_concurrency.lockutils [None req-72bb36a2-c5ce-4bb4-b8a6-7d89be6fdabb tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Lock "6fddc45c-d589-4423-a990-4be24d9dc84e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1139.856751] env[61923]: DEBUG oslo_concurrency.lockutils [None req-72bb36a2-c5ce-4bb4-b8a6-7d89be6fdabb tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Lock "6fddc45c-d589-4423-a990-4be24d9dc84e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1139.859296] env[61923]: INFO nova.compute.manager [None req-72bb36a2-c5ce-4bb4-b8a6-7d89be6fdabb tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: 6fddc45c-d589-4423-a990-4be24d9dc84e] Terminating instance [ 1139.861283] env[61923]: DEBUG nova.compute.manager [None req-72bb36a2-c5ce-4bb4-b8a6-7d89be6fdabb tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: 6fddc45c-d589-4423-a990-4be24d9dc84e] Start destroying the instance on the hypervisor. {{(pid=61923) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1139.861491] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-72bb36a2-c5ce-4bb4-b8a6-7d89be6fdabb tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: 6fddc45c-d589-4423-a990-4be24d9dc84e] Destroying instance {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1139.862420] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-684aaedc-f194-48cb-9b84-baa284bc1854 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.870387] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-72bb36a2-c5ce-4bb4-b8a6-7d89be6fdabb tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: 6fddc45c-d589-4423-a990-4be24d9dc84e] Powering off the VM {{(pid=61923) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1139.870636] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2003e771-cf61-49bc-9b8e-3c7c2bff04b8 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1139.878469] env[61923]: DEBUG nova.scheduler.client.report [None req-a84572db-25cb-45bf-8bab-bc3f62d5ec23 tempest-ServerMetadataTestJSON-1639770617 tempest-ServerMetadataTestJSON-1639770617-project-member] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1139.881559] env[61923]: DEBUG oslo_vmware.api [None req-72bb36a2-c5ce-4bb4-b8a6-7d89be6fdabb tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Waiting for the task: (returnval){ [ 1139.881559] env[61923]: value = "task-1378062" [ 1139.881559] env[61923]: _type = "Task" [ 1139.881559] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1139.889797] env[61923]: DEBUG oslo_vmware.api [None req-72bb36a2-c5ce-4bb4-b8a6-7d89be6fdabb tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Task: {'id': task-1378062, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1140.273190] env[61923]: DEBUG oslo_concurrency.lockutils [None req-53588979-1128-4b38-bd80-7fa4534341ad tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Acquiring lock "5dfbd79e-d43e-4769-b9f1-e8fdc09de31b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1140.273524] env[61923]: DEBUG oslo_concurrency.lockutils [None req-53588979-1128-4b38-bd80-7fa4534341ad tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Lock "5dfbd79e-d43e-4769-b9f1-e8fdc09de31b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1140.383271] env[61923]: DEBUG oslo_concurrency.lockutils [None req-a84572db-25cb-45bf-8bab-bc3f62d5ec23 tempest-ServerMetadataTestJSON-1639770617 tempest-ServerMetadataTestJSON-1639770617-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.125s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1140.383836] env[61923]: DEBUG nova.compute.manager [None req-a84572db-25cb-45bf-8bab-bc3f62d5ec23 tempest-ServerMetadataTestJSON-1639770617 tempest-ServerMetadataTestJSON-1639770617-project-member] [instance: b82a241e-db60-40bc-a738-b961c3e38b43] Start building networks asynchronously for instance. {{(pid=61923) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1140.386506] env[61923]: DEBUG oslo_concurrency.lockutils [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 2.066s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1140.397872] env[61923]: DEBUG oslo_vmware.api [None req-72bb36a2-c5ce-4bb4-b8a6-7d89be6fdabb tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Task: {'id': task-1378062, 'name': PowerOffVM_Task, 'duration_secs': 0.188573} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1140.398171] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-72bb36a2-c5ce-4bb4-b8a6-7d89be6fdabb tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: 6fddc45c-d589-4423-a990-4be24d9dc84e] Powered off the VM {{(pid=61923) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1140.398351] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-72bb36a2-c5ce-4bb4-b8a6-7d89be6fdabb tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: 6fddc45c-d589-4423-a990-4be24d9dc84e] Unregistering the VM {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1140.398631] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9e504ad8-9b6b-4689-a9e3-536859a26ddb {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1140.578859] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-72bb36a2-c5ce-4bb4-b8a6-7d89be6fdabb tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: 6fddc45c-d589-4423-a990-4be24d9dc84e] Unregistered the VM {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1140.579099] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-72bb36a2-c5ce-4bb4-b8a6-7d89be6fdabb tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: 6fddc45c-d589-4423-a990-4be24d9dc84e] Deleting contents of the VM from datastore datastore1 {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1140.579293] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-72bb36a2-c5ce-4bb4-b8a6-7d89be6fdabb tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Deleting the datastore file [datastore1] 6fddc45c-d589-4423-a990-4be24d9dc84e {{(pid=61923) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1140.579576] env[61923]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-0f8db8ce-9ca8-4632-b48b-2b9474be99fa {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1140.585842] env[61923]: DEBUG oslo_vmware.api [None req-72bb36a2-c5ce-4bb4-b8a6-7d89be6fdabb tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Waiting for the task: (returnval){ [ 1140.585842] env[61923]: value = "task-1378064" [ 1140.585842] env[61923]: _type = "Task" [ 1140.585842] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1140.594428] env[61923]: DEBUG oslo_vmware.api [None req-72bb36a2-c5ce-4bb4-b8a6-7d89be6fdabb tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Task: {'id': task-1378064, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1140.775761] env[61923]: DEBUG nova.compute.manager [None req-53588979-1128-4b38-bd80-7fa4534341ad tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: 5dfbd79e-d43e-4769-b9f1-e8fdc09de31b] Starting instance... {{(pid=61923) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1140.897284] env[61923]: DEBUG nova.compute.utils [None req-a84572db-25cb-45bf-8bab-bc3f62d5ec23 tempest-ServerMetadataTestJSON-1639770617 tempest-ServerMetadataTestJSON-1639770617-project-member] Using /dev/sd instead of None {{(pid=61923) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1140.899622] env[61923]: DEBUG nova.compute.manager [None req-a84572db-25cb-45bf-8bab-bc3f62d5ec23 tempest-ServerMetadataTestJSON-1639770617 tempest-ServerMetadataTestJSON-1639770617-project-member] [instance: b82a241e-db60-40bc-a738-b961c3e38b43] Allocating IP information in the background. {{(pid=61923) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1140.899922] env[61923]: DEBUG nova.network.neutron [None req-a84572db-25cb-45bf-8bab-bc3f62d5ec23 tempest-ServerMetadataTestJSON-1639770617 tempest-ServerMetadataTestJSON-1639770617-project-member] [instance: b82a241e-db60-40bc-a738-b961c3e38b43] allocate_for_instance() {{(pid=61923) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1140.953850] env[61923]: DEBUG nova.policy [None req-a84572db-25cb-45bf-8bab-bc3f62d5ec23 tempest-ServerMetadataTestJSON-1639770617 tempest-ServerMetadataTestJSON-1639770617-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0b8af8057ee44a41a055a25133398a96', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'dc12ba52164c4faca8b018c0862df979', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61923) authorize /opt/stack/nova/nova/policy.py:201}} [ 1141.099504] env[61923]: DEBUG oslo_vmware.api [None req-72bb36a2-c5ce-4bb4-b8a6-7d89be6fdabb tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Task: {'id': task-1378064, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.14655} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1141.099940] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-72bb36a2-c5ce-4bb4-b8a6-7d89be6fdabb tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Deleted the datastore file {{(pid=61923) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1141.100268] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-72bb36a2-c5ce-4bb4-b8a6-7d89be6fdabb tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: 6fddc45c-d589-4423-a990-4be24d9dc84e] Deleted contents of the VM from datastore datastore1 {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1141.100602] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-72bb36a2-c5ce-4bb4-b8a6-7d89be6fdabb tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: 6fddc45c-d589-4423-a990-4be24d9dc84e] Instance destroyed {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1141.100832] env[61923]: INFO nova.compute.manager [None req-72bb36a2-c5ce-4bb4-b8a6-7d89be6fdabb tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: 6fddc45c-d589-4423-a990-4be24d9dc84e] Took 1.24 seconds to destroy the instance on the hypervisor. [ 1141.101111] env[61923]: DEBUG oslo.service.loopingcall [None req-72bb36a2-c5ce-4bb4-b8a6-7d89be6fdabb tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61923) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1141.101349] env[61923]: DEBUG nova.compute.manager [-] [instance: 6fddc45c-d589-4423-a990-4be24d9dc84e] Deallocating network for instance {{(pid=61923) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1141.101447] env[61923]: DEBUG nova.network.neutron [-] [instance: 6fddc45c-d589-4423-a990-4be24d9dc84e] deallocate_for_instance() {{(pid=61923) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1141.221043] env[61923]: DEBUG nova.network.neutron [None req-a84572db-25cb-45bf-8bab-bc3f62d5ec23 tempest-ServerMetadataTestJSON-1639770617 tempest-ServerMetadataTestJSON-1639770617-project-member] [instance: b82a241e-db60-40bc-a738-b961c3e38b43] Successfully created port: 919e54ef-c0f6-4dde-b5e4-be987b7e3823 {{(pid=61923) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1141.305410] env[61923]: DEBUG oslo_concurrency.lockutils [None req-53588979-1128-4b38-bd80-7fa4534341ad tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1141.406300] env[61923]: DEBUG nova.compute.manager [None req-a84572db-25cb-45bf-8bab-bc3f62d5ec23 tempest-ServerMetadataTestJSON-1639770617 tempest-ServerMetadataTestJSON-1639770617-project-member] [instance: b82a241e-db60-40bc-a738-b961c3e38b43] Start building block device mappings for instance. {{(pid=61923) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1141.414413] env[61923]: DEBUG nova.compute.manager [req-6806dfde-e7e0-4813-9ace-a51f191bcbf1 req-0af9f937-293b-4d85-acb5-f8da3ace6a8f service nova] [instance: 6fddc45c-d589-4423-a990-4be24d9dc84e] Received event network-vif-deleted-cdf08c60-c555-4f7e-9a38-0b43dc6bd520 {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1141.415327] env[61923]: INFO nova.compute.manager [req-6806dfde-e7e0-4813-9ace-a51f191bcbf1 req-0af9f937-293b-4d85-acb5-f8da3ace6a8f service nova] [instance: 6fddc45c-d589-4423-a990-4be24d9dc84e] Neutron deleted interface cdf08c60-c555-4f7e-9a38-0b43dc6bd520; detaching it from the instance and deleting it from the info cache [ 1141.415327] env[61923]: DEBUG nova.network.neutron [req-6806dfde-e7e0-4813-9ace-a51f191bcbf1 req-0af9f937-293b-4d85-acb5-f8da3ace6a8f service nova] [instance: 6fddc45c-d589-4423-a990-4be24d9dc84e] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1141.423910] env[61923]: DEBUG nova.compute.resource_tracker [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Instance 6fddc45c-d589-4423-a990-4be24d9dc84e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=61923) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1141.423991] env[61923]: DEBUG nova.compute.resource_tracker [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Instance b82a241e-db60-40bc-a738-b961c3e38b43 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61923) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1141.859349] env[61923]: DEBUG nova.network.neutron [-] [instance: 6fddc45c-d589-4423-a990-4be24d9dc84e] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1141.918936] env[61923]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-6fc81080-90e9-47f8-8a0f-98c5d251df5e {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1141.927528] env[61923]: DEBUG nova.compute.resource_tracker [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Instance 5dfbd79e-d43e-4769-b9f1-e8fdc09de31b has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=61923) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 1141.927860] env[61923]: DEBUG nova.compute.resource_tracker [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Total usable vcpus: 48, total allocated vcpus: 2 {{(pid=61923) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1141.928124] env[61923]: DEBUG nova.compute.resource_tracker [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=960MB phys_disk=200GB used_disk=2GB total_vcpus=48 used_vcpus=2 pci_stats=[] {{(pid=61923) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1141.935955] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9853446-f94a-4013-8fb0-917a5091d909 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1141.973745] env[61923]: DEBUG nova.compute.manager [req-6806dfde-e7e0-4813-9ace-a51f191bcbf1 req-0af9f937-293b-4d85-acb5-f8da3ace6a8f service nova] [instance: 6fddc45c-d589-4423-a990-4be24d9dc84e] Detach interface failed, port_id=cdf08c60-c555-4f7e-9a38-0b43dc6bd520, reason: Instance 6fddc45c-d589-4423-a990-4be24d9dc84e could not be found. {{(pid=61923) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1141.997559] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ba38091-d4ad-45f6-9ded-4530f0702468 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1142.004975] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9fe94a60-5e2b-49b0-923f-e120298bedab {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1142.034409] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed6a3a5e-d6a2-4409-8a24-f9b24bed937d {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1142.041445] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3263000f-be38-40e2-bc11-2a8d37049572 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1142.054436] env[61923]: DEBUG nova.compute.provider_tree [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1142.362345] env[61923]: INFO nova.compute.manager [-] [instance: 6fddc45c-d589-4423-a990-4be24d9dc84e] Took 1.26 seconds to deallocate network for instance. [ 1142.419937] env[61923]: DEBUG nova.compute.manager [None req-a84572db-25cb-45bf-8bab-bc3f62d5ec23 tempest-ServerMetadataTestJSON-1639770617 tempest-ServerMetadataTestJSON-1639770617-project-member] [instance: b82a241e-db60-40bc-a738-b961c3e38b43] Start spawning the instance on the hypervisor. {{(pid=61923) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1142.445042] env[61923]: DEBUG nova.virt.hardware [None req-a84572db-25cb-45bf-8bab-bc3f62d5ec23 tempest-ServerMetadataTestJSON-1639770617 tempest-ServerMetadataTestJSON-1639770617-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-29T20:07:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-29T20:07:35Z,direct_url=,disk_format='vmdk',id=0f153f63-ae0a-45b1-b7f5-7f9b673c947f,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='4e7b5e3b3c3443c8bd5c70f8a15739f5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-29T20:07:36Z,virtual_size=,visibility=), allow threads: False {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1142.445330] env[61923]: DEBUG nova.virt.hardware [None req-a84572db-25cb-45bf-8bab-bc3f62d5ec23 tempest-ServerMetadataTestJSON-1639770617 tempest-ServerMetadataTestJSON-1639770617-project-member] Flavor limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1142.445489] env[61923]: DEBUG nova.virt.hardware [None req-a84572db-25cb-45bf-8bab-bc3f62d5ec23 tempest-ServerMetadataTestJSON-1639770617 tempest-ServerMetadataTestJSON-1639770617-project-member] Image limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1142.445674] env[61923]: DEBUG nova.virt.hardware [None req-a84572db-25cb-45bf-8bab-bc3f62d5ec23 tempest-ServerMetadataTestJSON-1639770617 tempest-ServerMetadataTestJSON-1639770617-project-member] Flavor pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1142.445823] env[61923]: DEBUG nova.virt.hardware [None req-a84572db-25cb-45bf-8bab-bc3f62d5ec23 tempest-ServerMetadataTestJSON-1639770617 tempest-ServerMetadataTestJSON-1639770617-project-member] Image pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1142.446102] env[61923]: DEBUG nova.virt.hardware [None req-a84572db-25cb-45bf-8bab-bc3f62d5ec23 tempest-ServerMetadataTestJSON-1639770617 tempest-ServerMetadataTestJSON-1639770617-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1142.446354] env[61923]: DEBUG nova.virt.hardware [None req-a84572db-25cb-45bf-8bab-bc3f62d5ec23 tempest-ServerMetadataTestJSON-1639770617 tempest-ServerMetadataTestJSON-1639770617-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1142.446517] env[61923]: DEBUG nova.virt.hardware [None req-a84572db-25cb-45bf-8bab-bc3f62d5ec23 tempest-ServerMetadataTestJSON-1639770617 tempest-ServerMetadataTestJSON-1639770617-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1142.446684] env[61923]: DEBUG nova.virt.hardware [None req-a84572db-25cb-45bf-8bab-bc3f62d5ec23 tempest-ServerMetadataTestJSON-1639770617 tempest-ServerMetadataTestJSON-1639770617-project-member] Got 1 possible topologies {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1142.446848] env[61923]: DEBUG nova.virt.hardware [None req-a84572db-25cb-45bf-8bab-bc3f62d5ec23 tempest-ServerMetadataTestJSON-1639770617 tempest-ServerMetadataTestJSON-1639770617-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1142.447115] env[61923]: DEBUG nova.virt.hardware [None req-a84572db-25cb-45bf-8bab-bc3f62d5ec23 tempest-ServerMetadataTestJSON-1639770617 tempest-ServerMetadataTestJSON-1639770617-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1142.447987] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18bbc3f6-5e04-463e-860f-6be10b62a76b {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1142.456639] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a245e1e-83a8-4cac-9b90-fe69467a4768 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1142.557190] env[61923]: DEBUG nova.scheduler.client.report [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1142.660806] env[61923]: DEBUG nova.compute.manager [req-43415c6e-4333-4ba8-8b40-4f59c7291ed9 req-1fca3c22-4bf8-4722-9c75-371f90a78047 service nova] [instance: b82a241e-db60-40bc-a738-b961c3e38b43] Received event network-vif-plugged-919e54ef-c0f6-4dde-b5e4-be987b7e3823 {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1142.661065] env[61923]: DEBUG oslo_concurrency.lockutils [req-43415c6e-4333-4ba8-8b40-4f59c7291ed9 req-1fca3c22-4bf8-4722-9c75-371f90a78047 service nova] Acquiring lock "b82a241e-db60-40bc-a738-b961c3e38b43-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1142.661684] env[61923]: DEBUG oslo_concurrency.lockutils [req-43415c6e-4333-4ba8-8b40-4f59c7291ed9 req-1fca3c22-4bf8-4722-9c75-371f90a78047 service nova] Lock "b82a241e-db60-40bc-a738-b961c3e38b43-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1142.661930] env[61923]: DEBUG oslo_concurrency.lockutils [req-43415c6e-4333-4ba8-8b40-4f59c7291ed9 req-1fca3c22-4bf8-4722-9c75-371f90a78047 service nova] Lock "b82a241e-db60-40bc-a738-b961c3e38b43-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1142.662141] env[61923]: DEBUG nova.compute.manager [req-43415c6e-4333-4ba8-8b40-4f59c7291ed9 req-1fca3c22-4bf8-4722-9c75-371f90a78047 service nova] [instance: b82a241e-db60-40bc-a738-b961c3e38b43] No waiting events found dispatching network-vif-plugged-919e54ef-c0f6-4dde-b5e4-be987b7e3823 {{(pid=61923) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1142.662314] env[61923]: WARNING nova.compute.manager [req-43415c6e-4333-4ba8-8b40-4f59c7291ed9 req-1fca3c22-4bf8-4722-9c75-371f90a78047 service nova] [instance: b82a241e-db60-40bc-a738-b961c3e38b43] Received unexpected event network-vif-plugged-919e54ef-c0f6-4dde-b5e4-be987b7e3823 for instance with vm_state building and task_state spawning. [ 1142.717157] env[61923]: DEBUG nova.network.neutron [None req-a84572db-25cb-45bf-8bab-bc3f62d5ec23 tempest-ServerMetadataTestJSON-1639770617 tempest-ServerMetadataTestJSON-1639770617-project-member] [instance: b82a241e-db60-40bc-a738-b961c3e38b43] Successfully updated port: 919e54ef-c0f6-4dde-b5e4-be987b7e3823 {{(pid=61923) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1142.869421] env[61923]: DEBUG oslo_concurrency.lockutils [None req-72bb36a2-c5ce-4bb4-b8a6-7d89be6fdabb tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1143.061661] env[61923]: DEBUG nova.compute.resource_tracker [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61923) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1143.061948] env[61923]: DEBUG oslo_concurrency.lockutils [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.675s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1143.062301] env[61923]: DEBUG oslo_concurrency.lockutils [None req-53588979-1128-4b38-bd80-7fa4534341ad tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.757s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1143.063992] env[61923]: INFO nova.compute.claims [None req-53588979-1128-4b38-bd80-7fa4534341ad tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: 5dfbd79e-d43e-4769-b9f1-e8fdc09de31b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1143.219668] env[61923]: DEBUG oslo_concurrency.lockutils [None req-a84572db-25cb-45bf-8bab-bc3f62d5ec23 tempest-ServerMetadataTestJSON-1639770617 tempest-ServerMetadataTestJSON-1639770617-project-member] Acquiring lock "refresh_cache-b82a241e-db60-40bc-a738-b961c3e38b43" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1143.219959] env[61923]: DEBUG oslo_concurrency.lockutils [None req-a84572db-25cb-45bf-8bab-bc3f62d5ec23 tempest-ServerMetadataTestJSON-1639770617 tempest-ServerMetadataTestJSON-1639770617-project-member] Acquired lock "refresh_cache-b82a241e-db60-40bc-a738-b961c3e38b43" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1143.220168] env[61923]: DEBUG nova.network.neutron [None req-a84572db-25cb-45bf-8bab-bc3f62d5ec23 tempest-ServerMetadataTestJSON-1639770617 tempest-ServerMetadataTestJSON-1639770617-project-member] [instance: b82a241e-db60-40bc-a738-b961c3e38b43] Building network info cache for instance {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1143.750491] env[61923]: DEBUG nova.network.neutron [None req-a84572db-25cb-45bf-8bab-bc3f62d5ec23 tempest-ServerMetadataTestJSON-1639770617 tempest-ServerMetadataTestJSON-1639770617-project-member] [instance: b82a241e-db60-40bc-a738-b961c3e38b43] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1143.877201] env[61923]: DEBUG nova.network.neutron [None req-a84572db-25cb-45bf-8bab-bc3f62d5ec23 tempest-ServerMetadataTestJSON-1639770617 tempest-ServerMetadataTestJSON-1639770617-project-member] [instance: b82a241e-db60-40bc-a738-b961c3e38b43] Updating instance_info_cache with network_info: [{"id": "919e54ef-c0f6-4dde-b5e4-be987b7e3823", "address": "fa:16:3e:d1:74:b7", "network": {"id": "5371fb9f-d5d8-4ece-b7cc-4b97b7c6da43", "bridge": "br-int", "label": "tempest-ServerMetadataTestJSON-368422525-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "dc12ba52164c4faca8b018c0862df979", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "aa410d21-2141-45bb-8d0b-16c77304605f", "external-id": "nsx-vlan-transportzone-886", "segmentation_id": 886, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap919e54ef-c0", "ovs_interfaceid": "919e54ef-c0f6-4dde-b5e4-be987b7e3823", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1144.062433] env[61923]: DEBUG oslo_service.periodic_task [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61923) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1144.134969] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bed21eb3-e0d9-4b07-8111-c9da963e11c0 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1144.142342] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-716440da-2494-4a5b-9992-001ecad68cfd {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1144.170908] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-434adf53-b848-41f7-a2fa-24378d6293ea {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1144.177877] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-662e732a-eb68-44a9-8f58-8be94702b590 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1144.191791] env[61923]: DEBUG nova.compute.provider_tree [None req-53588979-1128-4b38-bd80-7fa4534341ad tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1144.379022] env[61923]: DEBUG oslo_concurrency.lockutils [None req-a84572db-25cb-45bf-8bab-bc3f62d5ec23 tempest-ServerMetadataTestJSON-1639770617 tempest-ServerMetadataTestJSON-1639770617-project-member] Releasing lock "refresh_cache-b82a241e-db60-40bc-a738-b961c3e38b43" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1144.379351] env[61923]: DEBUG nova.compute.manager [None req-a84572db-25cb-45bf-8bab-bc3f62d5ec23 tempest-ServerMetadataTestJSON-1639770617 tempest-ServerMetadataTestJSON-1639770617-project-member] [instance: b82a241e-db60-40bc-a738-b961c3e38b43] Instance network_info: |[{"id": "919e54ef-c0f6-4dde-b5e4-be987b7e3823", "address": "fa:16:3e:d1:74:b7", "network": {"id": "5371fb9f-d5d8-4ece-b7cc-4b97b7c6da43", "bridge": "br-int", "label": "tempest-ServerMetadataTestJSON-368422525-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "dc12ba52164c4faca8b018c0862df979", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "aa410d21-2141-45bb-8d0b-16c77304605f", "external-id": "nsx-vlan-transportzone-886", "segmentation_id": 886, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap919e54ef-c0", "ovs_interfaceid": "919e54ef-c0f6-4dde-b5e4-be987b7e3823", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61923) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1144.379845] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-a84572db-25cb-45bf-8bab-bc3f62d5ec23 tempest-ServerMetadataTestJSON-1639770617 tempest-ServerMetadataTestJSON-1639770617-project-member] [instance: b82a241e-db60-40bc-a738-b961c3e38b43] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:d1:74:b7', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'aa410d21-2141-45bb-8d0b-16c77304605f', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '919e54ef-c0f6-4dde-b5e4-be987b7e3823', 'vif_model': 'vmxnet3'}] {{(pid=61923) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1144.387199] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-a84572db-25cb-45bf-8bab-bc3f62d5ec23 tempest-ServerMetadataTestJSON-1639770617 tempest-ServerMetadataTestJSON-1639770617-project-member] Creating folder: Project (dc12ba52164c4faca8b018c0862df979). Parent ref: group-v292629. {{(pid=61923) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1144.387490] env[61923]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-0f91e2b8-4c34-40a7-81fb-772de7ba4f47 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1144.398081] env[61923]: INFO nova.virt.vmwareapi.vm_util [None req-a84572db-25cb-45bf-8bab-bc3f62d5ec23 tempest-ServerMetadataTestJSON-1639770617 tempest-ServerMetadataTestJSON-1639770617-project-member] Created folder: Project (dc12ba52164c4faca8b018c0862df979) in parent group-v292629. [ 1144.398271] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-a84572db-25cb-45bf-8bab-bc3f62d5ec23 tempest-ServerMetadataTestJSON-1639770617 tempest-ServerMetadataTestJSON-1639770617-project-member] Creating folder: Instances. Parent ref: group-v292757. {{(pid=61923) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1144.398502] env[61923]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-9dbfb8e5-bb0d-4b6e-b086-df9e68f4d72c {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1144.406786] env[61923]: INFO nova.virt.vmwareapi.vm_util [None req-a84572db-25cb-45bf-8bab-bc3f62d5ec23 tempest-ServerMetadataTestJSON-1639770617 tempest-ServerMetadataTestJSON-1639770617-project-member] Created folder: Instances in parent group-v292757. [ 1144.407014] env[61923]: DEBUG oslo.service.loopingcall [None req-a84572db-25cb-45bf-8bab-bc3f62d5ec23 tempest-ServerMetadataTestJSON-1639770617 tempest-ServerMetadataTestJSON-1639770617-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61923) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1144.407204] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b82a241e-db60-40bc-a738-b961c3e38b43] Creating VM on the ESX host {{(pid=61923) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1144.407393] env[61923]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-a110e86a-ed9f-495b-b31c-0d94daba40a0 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1144.424669] env[61923]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1144.424669] env[61923]: value = "task-1378067" [ 1144.424669] env[61923]: _type = "Task" [ 1144.424669] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1144.431406] env[61923]: DEBUG oslo_vmware.api [-] Task: {'id': task-1378067, 'name': CreateVM_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1144.684657] env[61923]: DEBUG nova.compute.manager [req-7830b92a-96e0-420f-868c-b48482956fb4 req-f10b3efb-28f8-4d1b-99aa-94b63d416c23 service nova] [instance: b82a241e-db60-40bc-a738-b961c3e38b43] Received event network-changed-919e54ef-c0f6-4dde-b5e4-be987b7e3823 {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1144.684951] env[61923]: DEBUG nova.compute.manager [req-7830b92a-96e0-420f-868c-b48482956fb4 req-f10b3efb-28f8-4d1b-99aa-94b63d416c23 service nova] [instance: b82a241e-db60-40bc-a738-b961c3e38b43] Refreshing instance network info cache due to event network-changed-919e54ef-c0f6-4dde-b5e4-be987b7e3823. {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1144.685130] env[61923]: DEBUG oslo_concurrency.lockutils [req-7830b92a-96e0-420f-868c-b48482956fb4 req-f10b3efb-28f8-4d1b-99aa-94b63d416c23 service nova] Acquiring lock "refresh_cache-b82a241e-db60-40bc-a738-b961c3e38b43" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1144.685295] env[61923]: DEBUG oslo_concurrency.lockutils [req-7830b92a-96e0-420f-868c-b48482956fb4 req-f10b3efb-28f8-4d1b-99aa-94b63d416c23 service nova] Acquired lock "refresh_cache-b82a241e-db60-40bc-a738-b961c3e38b43" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1144.685465] env[61923]: DEBUG nova.network.neutron [req-7830b92a-96e0-420f-868c-b48482956fb4 req-f10b3efb-28f8-4d1b-99aa-94b63d416c23 service nova] [instance: b82a241e-db60-40bc-a738-b961c3e38b43] Refreshing network info cache for port 919e54ef-c0f6-4dde-b5e4-be987b7e3823 {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1144.694976] env[61923]: DEBUG nova.scheduler.client.report [None req-53588979-1128-4b38-bd80-7fa4534341ad tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1144.934250] env[61923]: DEBUG oslo_vmware.api [-] Task: {'id': task-1378067, 'name': CreateVM_Task, 'duration_secs': 0.317849} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1144.934580] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b82a241e-db60-40bc-a738-b961c3e38b43] Created VM on the ESX host {{(pid=61923) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1144.935057] env[61923]: DEBUG oslo_concurrency.lockutils [None req-a84572db-25cb-45bf-8bab-bc3f62d5ec23 tempest-ServerMetadataTestJSON-1639770617 tempest-ServerMetadataTestJSON-1639770617-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1144.935242] env[61923]: DEBUG oslo_concurrency.lockutils [None req-a84572db-25cb-45bf-8bab-bc3f62d5ec23 tempest-ServerMetadataTestJSON-1639770617 tempest-ServerMetadataTestJSON-1639770617-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1144.935566] env[61923]: DEBUG oslo_concurrency.lockutils [None req-a84572db-25cb-45bf-8bab-bc3f62d5ec23 tempest-ServerMetadataTestJSON-1639770617 tempest-ServerMetadataTestJSON-1639770617-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1144.935816] env[61923]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-23ccd5a4-2101-4665-8ec7-65c9ae3d386f {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1144.939968] env[61923]: DEBUG oslo_vmware.api [None req-a84572db-25cb-45bf-8bab-bc3f62d5ec23 tempest-ServerMetadataTestJSON-1639770617 tempest-ServerMetadataTestJSON-1639770617-project-member] Waiting for the task: (returnval){ [ 1144.939968] env[61923]: value = "session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52514dc0-ae15-970a-43be-12180339e055" [ 1144.939968] env[61923]: _type = "Task" [ 1144.939968] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1144.947189] env[61923]: DEBUG oslo_vmware.api [None req-a84572db-25cb-45bf-8bab-bc3f62d5ec23 tempest-ServerMetadataTestJSON-1639770617 tempest-ServerMetadataTestJSON-1639770617-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52514dc0-ae15-970a-43be-12180339e055, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1145.200091] env[61923]: DEBUG oslo_concurrency.lockutils [None req-53588979-1128-4b38-bd80-7fa4534341ad tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.137s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1145.200091] env[61923]: DEBUG nova.compute.manager [None req-53588979-1128-4b38-bd80-7fa4534341ad tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: 5dfbd79e-d43e-4769-b9f1-e8fdc09de31b] Start building networks asynchronously for instance. {{(pid=61923) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1145.202489] env[61923]: DEBUG oslo_concurrency.lockutils [None req-72bb36a2-c5ce-4bb4-b8a6-7d89be6fdabb tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 2.333s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1145.203196] env[61923]: DEBUG nova.objects.instance [None req-72bb36a2-c5ce-4bb4-b8a6-7d89be6fdabb tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Lazy-loading 'resources' on Instance uuid 6fddc45c-d589-4423-a990-4be24d9dc84e {{(pid=61923) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1145.383610] env[61923]: DEBUG nova.network.neutron [req-7830b92a-96e0-420f-868c-b48482956fb4 req-f10b3efb-28f8-4d1b-99aa-94b63d416c23 service nova] [instance: b82a241e-db60-40bc-a738-b961c3e38b43] Updated VIF entry in instance network info cache for port 919e54ef-c0f6-4dde-b5e4-be987b7e3823. {{(pid=61923) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1145.383955] env[61923]: DEBUG nova.network.neutron [req-7830b92a-96e0-420f-868c-b48482956fb4 req-f10b3efb-28f8-4d1b-99aa-94b63d416c23 service nova] [instance: b82a241e-db60-40bc-a738-b961c3e38b43] Updating instance_info_cache with network_info: [{"id": "919e54ef-c0f6-4dde-b5e4-be987b7e3823", "address": "fa:16:3e:d1:74:b7", "network": {"id": "5371fb9f-d5d8-4ece-b7cc-4b97b7c6da43", "bridge": "br-int", "label": "tempest-ServerMetadataTestJSON-368422525-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "dc12ba52164c4faca8b018c0862df979", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "aa410d21-2141-45bb-8d0b-16c77304605f", "external-id": "nsx-vlan-transportzone-886", "segmentation_id": 886, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap919e54ef-c0", "ovs_interfaceid": "919e54ef-c0f6-4dde-b5e4-be987b7e3823", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1145.450499] env[61923]: DEBUG oslo_vmware.api [None req-a84572db-25cb-45bf-8bab-bc3f62d5ec23 tempest-ServerMetadataTestJSON-1639770617 tempest-ServerMetadataTestJSON-1639770617-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52514dc0-ae15-970a-43be-12180339e055, 'name': SearchDatastore_Task, 'duration_secs': 0.009224} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1145.450763] env[61923]: DEBUG oslo_concurrency.lockutils [None req-a84572db-25cb-45bf-8bab-bc3f62d5ec23 tempest-ServerMetadataTestJSON-1639770617 tempest-ServerMetadataTestJSON-1639770617-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1145.450999] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-a84572db-25cb-45bf-8bab-bc3f62d5ec23 tempest-ServerMetadataTestJSON-1639770617 tempest-ServerMetadataTestJSON-1639770617-project-member] [instance: b82a241e-db60-40bc-a738-b961c3e38b43] Processing image 0f153f63-ae0a-45b1-b7f5-7f9b673c947f {{(pid=61923) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1145.451256] env[61923]: DEBUG oslo_concurrency.lockutils [None req-a84572db-25cb-45bf-8bab-bc3f62d5ec23 tempest-ServerMetadataTestJSON-1639770617 tempest-ServerMetadataTestJSON-1639770617-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1145.451406] env[61923]: DEBUG oslo_concurrency.lockutils [None req-a84572db-25cb-45bf-8bab-bc3f62d5ec23 tempest-ServerMetadataTestJSON-1639770617 tempest-ServerMetadataTestJSON-1639770617-project-member] Acquired lock "[datastore2] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1145.451585] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-a84572db-25cb-45bf-8bab-bc3f62d5ec23 tempest-ServerMetadataTestJSON-1639770617 tempest-ServerMetadataTestJSON-1639770617-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=61923) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1145.451844] env[61923]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-76505350-7f03-4549-a548-55aefd1bda16 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1145.460016] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-a84572db-25cb-45bf-8bab-bc3f62d5ec23 tempest-ServerMetadataTestJSON-1639770617 tempest-ServerMetadataTestJSON-1639770617-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=61923) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1145.460205] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-a84572db-25cb-45bf-8bab-bc3f62d5ec23 tempest-ServerMetadataTestJSON-1639770617 tempest-ServerMetadataTestJSON-1639770617-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=61923) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1145.460869] env[61923]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e6a29123-78ef-4c64-aead-1b177f287fd3 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1145.465431] env[61923]: DEBUG oslo_vmware.api [None req-a84572db-25cb-45bf-8bab-bc3f62d5ec23 tempest-ServerMetadataTestJSON-1639770617 tempest-ServerMetadataTestJSON-1639770617-project-member] Waiting for the task: (returnval){ [ 1145.465431] env[61923]: value = "session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52a420bc-22b8-5a6b-066c-b3be183eb005" [ 1145.465431] env[61923]: _type = "Task" [ 1145.465431] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1145.472740] env[61923]: DEBUG oslo_vmware.api [None req-a84572db-25cb-45bf-8bab-bc3f62d5ec23 tempest-ServerMetadataTestJSON-1639770617 tempest-ServerMetadataTestJSON-1639770617-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52a420bc-22b8-5a6b-066c-b3be183eb005, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1145.705537] env[61923]: DEBUG nova.compute.utils [None req-53588979-1128-4b38-bd80-7fa4534341ad tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Using /dev/sd instead of None {{(pid=61923) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1145.709984] env[61923]: DEBUG nova.compute.manager [None req-53588979-1128-4b38-bd80-7fa4534341ad tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: 5dfbd79e-d43e-4769-b9f1-e8fdc09de31b] Allocating IP information in the background. {{(pid=61923) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1145.709984] env[61923]: DEBUG nova.network.neutron [None req-53588979-1128-4b38-bd80-7fa4534341ad tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: 5dfbd79e-d43e-4769-b9f1-e8fdc09de31b] allocate_for_instance() {{(pid=61923) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1145.746993] env[61923]: DEBUG nova.policy [None req-53588979-1128-4b38-bd80-7fa4534341ad tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6087a385c42e4383aa26a7f586cca87f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5a4db4f8aa194a219e80c538cc715d43', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61923) authorize /opt/stack/nova/nova/policy.py:201}} [ 1145.753440] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-009b12f1-b98f-4582-a672-293e9c54ed80 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1145.760895] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e3245c6-fcea-478a-8f9a-ba4f0d29dc46 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1145.790346] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0eb54235-ac21-4b5f-a5c8-e34b311746c9 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1145.797359] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e86ea9a1-d8d6-493a-826d-06a038120d7c {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1145.810021] env[61923]: DEBUG nova.compute.provider_tree [None req-72bb36a2-c5ce-4bb4-b8a6-7d89be6fdabb tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1145.886289] env[61923]: DEBUG oslo_concurrency.lockutils [req-7830b92a-96e0-420f-868c-b48482956fb4 req-f10b3efb-28f8-4d1b-99aa-94b63d416c23 service nova] Releasing lock "refresh_cache-b82a241e-db60-40bc-a738-b961c3e38b43" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1145.975976] env[61923]: DEBUG oslo_vmware.api [None req-a84572db-25cb-45bf-8bab-bc3f62d5ec23 tempest-ServerMetadataTestJSON-1639770617 tempest-ServerMetadataTestJSON-1639770617-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52a420bc-22b8-5a6b-066c-b3be183eb005, 'name': SearchDatastore_Task, 'duration_secs': 0.00791} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1145.976781] env[61923]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e4d59488-47ba-4d40-96cc-63a0c083acf7 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1145.981714] env[61923]: DEBUG oslo_vmware.api [None req-a84572db-25cb-45bf-8bab-bc3f62d5ec23 tempest-ServerMetadataTestJSON-1639770617 tempest-ServerMetadataTestJSON-1639770617-project-member] Waiting for the task: (returnval){ [ 1145.981714] env[61923]: value = "session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52fa80f6-bda2-e6f3-d039-ed5840e00139" [ 1145.981714] env[61923]: _type = "Task" [ 1145.981714] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1145.982434] env[61923]: DEBUG nova.network.neutron [None req-53588979-1128-4b38-bd80-7fa4534341ad tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: 5dfbd79e-d43e-4769-b9f1-e8fdc09de31b] Successfully created port: 0c9f3dd3-d6f5-47b0-90e7-d5e54eea95e1 {{(pid=61923) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1145.991740] env[61923]: DEBUG oslo_vmware.api [None req-a84572db-25cb-45bf-8bab-bc3f62d5ec23 tempest-ServerMetadataTestJSON-1639770617 tempest-ServerMetadataTestJSON-1639770617-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52fa80f6-bda2-e6f3-d039-ed5840e00139, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1146.210446] env[61923]: DEBUG nova.compute.manager [None req-53588979-1128-4b38-bd80-7fa4534341ad tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: 5dfbd79e-d43e-4769-b9f1-e8fdc09de31b] Start building block device mappings for instance. {{(pid=61923) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1146.313020] env[61923]: DEBUG nova.scheduler.client.report [None req-72bb36a2-c5ce-4bb4-b8a6-7d89be6fdabb tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1146.496324] env[61923]: DEBUG oslo_vmware.api [None req-a84572db-25cb-45bf-8bab-bc3f62d5ec23 tempest-ServerMetadataTestJSON-1639770617 tempest-ServerMetadataTestJSON-1639770617-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52fa80f6-bda2-e6f3-d039-ed5840e00139, 'name': SearchDatastore_Task, 'duration_secs': 0.010035} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1146.496598] env[61923]: DEBUG oslo_concurrency.lockutils [None req-a84572db-25cb-45bf-8bab-bc3f62d5ec23 tempest-ServerMetadataTestJSON-1639770617 tempest-ServerMetadataTestJSON-1639770617-project-member] Releasing lock "[datastore2] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1146.496859] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-a84572db-25cb-45bf-8bab-bc3f62d5ec23 tempest-ServerMetadataTestJSON-1639770617 tempest-ServerMetadataTestJSON-1639770617-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk to [datastore2] b82a241e-db60-40bc-a738-b961c3e38b43/b82a241e-db60-40bc-a738-b961c3e38b43.vmdk {{(pid=61923) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1146.497130] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-33b4345b-ffc8-4c96-b2cf-9e11bc66fd4f {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1146.503172] env[61923]: DEBUG oslo_vmware.api [None req-a84572db-25cb-45bf-8bab-bc3f62d5ec23 tempest-ServerMetadataTestJSON-1639770617 tempest-ServerMetadataTestJSON-1639770617-project-member] Waiting for the task: (returnval){ [ 1146.503172] env[61923]: value = "task-1378068" [ 1146.503172] env[61923]: _type = "Task" [ 1146.503172] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1146.510811] env[61923]: DEBUG oslo_vmware.api [None req-a84572db-25cb-45bf-8bab-bc3f62d5ec23 tempest-ServerMetadataTestJSON-1639770617 tempest-ServerMetadataTestJSON-1639770617-project-member] Task: {'id': task-1378068, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1146.818679] env[61923]: DEBUG oslo_concurrency.lockutils [None req-72bb36a2-c5ce-4bb4-b8a6-7d89be6fdabb tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.616s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1146.848380] env[61923]: INFO nova.scheduler.client.report [None req-72bb36a2-c5ce-4bb4-b8a6-7d89be6fdabb tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Deleted allocations for instance 6fddc45c-d589-4423-a990-4be24d9dc84e [ 1147.013542] env[61923]: DEBUG oslo_vmware.api [None req-a84572db-25cb-45bf-8bab-bc3f62d5ec23 tempest-ServerMetadataTestJSON-1639770617 tempest-ServerMetadataTestJSON-1639770617-project-member] Task: {'id': task-1378068, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.50751} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1147.013946] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-a84572db-25cb-45bf-8bab-bc3f62d5ec23 tempest-ServerMetadataTestJSON-1639770617 tempest-ServerMetadataTestJSON-1639770617-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk to [datastore2] b82a241e-db60-40bc-a738-b961c3e38b43/b82a241e-db60-40bc-a738-b961c3e38b43.vmdk {{(pid=61923) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1147.014050] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-a84572db-25cb-45bf-8bab-bc3f62d5ec23 tempest-ServerMetadataTestJSON-1639770617 tempest-ServerMetadataTestJSON-1639770617-project-member] [instance: b82a241e-db60-40bc-a738-b961c3e38b43] Extending root virtual disk to 1048576 {{(pid=61923) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1147.014279] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-11a18b8a-2ee6-41e0-a7c6-9a27379fe07f {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1147.020368] env[61923]: DEBUG oslo_vmware.api [None req-a84572db-25cb-45bf-8bab-bc3f62d5ec23 tempest-ServerMetadataTestJSON-1639770617 tempest-ServerMetadataTestJSON-1639770617-project-member] Waiting for the task: (returnval){ [ 1147.020368] env[61923]: value = "task-1378069" [ 1147.020368] env[61923]: _type = "Task" [ 1147.020368] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1147.028126] env[61923]: DEBUG oslo_vmware.api [None req-a84572db-25cb-45bf-8bab-bc3f62d5ec23 tempest-ServerMetadataTestJSON-1639770617 tempest-ServerMetadataTestJSON-1639770617-project-member] Task: {'id': task-1378069, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1147.220123] env[61923]: DEBUG nova.compute.manager [None req-53588979-1128-4b38-bd80-7fa4534341ad tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: 5dfbd79e-d43e-4769-b9f1-e8fdc09de31b] Start spawning the instance on the hypervisor. {{(pid=61923) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1147.243992] env[61923]: DEBUG nova.virt.hardware [None req-53588979-1128-4b38-bd80-7fa4534341ad tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-29T20:07:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-29T20:07:35Z,direct_url=,disk_format='vmdk',id=0f153f63-ae0a-45b1-b7f5-7f9b673c947f,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='4e7b5e3b3c3443c8bd5c70f8a15739f5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-29T20:07:36Z,virtual_size=,visibility=), allow threads: False {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1147.244272] env[61923]: DEBUG nova.virt.hardware [None req-53588979-1128-4b38-bd80-7fa4534341ad tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Flavor limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1147.244430] env[61923]: DEBUG nova.virt.hardware [None req-53588979-1128-4b38-bd80-7fa4534341ad tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Image limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1147.244608] env[61923]: DEBUG nova.virt.hardware [None req-53588979-1128-4b38-bd80-7fa4534341ad tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Flavor pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1147.244753] env[61923]: DEBUG nova.virt.hardware [None req-53588979-1128-4b38-bd80-7fa4534341ad tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Image pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1147.244897] env[61923]: DEBUG nova.virt.hardware [None req-53588979-1128-4b38-bd80-7fa4534341ad tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1147.245126] env[61923]: DEBUG nova.virt.hardware [None req-53588979-1128-4b38-bd80-7fa4534341ad tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1147.245294] env[61923]: DEBUG nova.virt.hardware [None req-53588979-1128-4b38-bd80-7fa4534341ad tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1147.245463] env[61923]: DEBUG nova.virt.hardware [None req-53588979-1128-4b38-bd80-7fa4534341ad tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Got 1 possible topologies {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1147.245622] env[61923]: DEBUG nova.virt.hardware [None req-53588979-1128-4b38-bd80-7fa4534341ad tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1147.245792] env[61923]: DEBUG nova.virt.hardware [None req-53588979-1128-4b38-bd80-7fa4534341ad tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1147.246656] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ba85ac8-a33e-4b16-8ef2-b441dde2778b {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1147.254673] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37c9294b-77a8-4884-9b2e-f7c645966c60 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1147.359957] env[61923]: DEBUG oslo_concurrency.lockutils [None req-72bb36a2-c5ce-4bb4-b8a6-7d89be6fdabb tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Lock "6fddc45c-d589-4423-a990-4be24d9dc84e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.504s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1147.379992] env[61923]: DEBUG nova.compute.manager [req-4904a742-88b2-4952-a6a1-1f830313fa77 req-58adca90-e936-49f1-8e36-fc371a342186 service nova] [instance: 5dfbd79e-d43e-4769-b9f1-e8fdc09de31b] Received event network-vif-plugged-0c9f3dd3-d6f5-47b0-90e7-d5e54eea95e1 {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1147.380244] env[61923]: DEBUG oslo_concurrency.lockutils [req-4904a742-88b2-4952-a6a1-1f830313fa77 req-58adca90-e936-49f1-8e36-fc371a342186 service nova] Acquiring lock "5dfbd79e-d43e-4769-b9f1-e8fdc09de31b-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1147.380540] env[61923]: DEBUG oslo_concurrency.lockutils [req-4904a742-88b2-4952-a6a1-1f830313fa77 req-58adca90-e936-49f1-8e36-fc371a342186 service nova] Lock "5dfbd79e-d43e-4769-b9f1-e8fdc09de31b-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1147.380730] env[61923]: DEBUG oslo_concurrency.lockutils [req-4904a742-88b2-4952-a6a1-1f830313fa77 req-58adca90-e936-49f1-8e36-fc371a342186 service nova] Lock "5dfbd79e-d43e-4769-b9f1-e8fdc09de31b-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1147.380900] env[61923]: DEBUG nova.compute.manager [req-4904a742-88b2-4952-a6a1-1f830313fa77 req-58adca90-e936-49f1-8e36-fc371a342186 service nova] [instance: 5dfbd79e-d43e-4769-b9f1-e8fdc09de31b] No waiting events found dispatching network-vif-plugged-0c9f3dd3-d6f5-47b0-90e7-d5e54eea95e1 {{(pid=61923) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1147.381080] env[61923]: WARNING nova.compute.manager [req-4904a742-88b2-4952-a6a1-1f830313fa77 req-58adca90-e936-49f1-8e36-fc371a342186 service nova] [instance: 5dfbd79e-d43e-4769-b9f1-e8fdc09de31b] Received unexpected event network-vif-plugged-0c9f3dd3-d6f5-47b0-90e7-d5e54eea95e1 for instance with vm_state building and task_state spawning. [ 1147.530336] env[61923]: DEBUG oslo_vmware.api [None req-a84572db-25cb-45bf-8bab-bc3f62d5ec23 tempest-ServerMetadataTestJSON-1639770617 tempest-ServerMetadataTestJSON-1639770617-project-member] Task: {'id': task-1378069, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.0567} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1147.530557] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-a84572db-25cb-45bf-8bab-bc3f62d5ec23 tempest-ServerMetadataTestJSON-1639770617 tempest-ServerMetadataTestJSON-1639770617-project-member] [instance: b82a241e-db60-40bc-a738-b961c3e38b43] Extended root virtual disk {{(pid=61923) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1147.531296] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e03e21b3-d000-4602-9646-6a23bc498d9b {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1147.552202] env[61923]: DEBUG nova.virt.vmwareapi.volumeops [None req-a84572db-25cb-45bf-8bab-bc3f62d5ec23 tempest-ServerMetadataTestJSON-1639770617 tempest-ServerMetadataTestJSON-1639770617-project-member] [instance: b82a241e-db60-40bc-a738-b961c3e38b43] Reconfiguring VM instance instance-00000067 to attach disk [datastore2] b82a241e-db60-40bc-a738-b961c3e38b43/b82a241e-db60-40bc-a738-b961c3e38b43.vmdk or device None with type sparse {{(pid=61923) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1147.552427] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ea1a998d-71a2-4fd3-ad68-0c709b88b168 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1147.571155] env[61923]: DEBUG oslo_vmware.api [None req-a84572db-25cb-45bf-8bab-bc3f62d5ec23 tempest-ServerMetadataTestJSON-1639770617 tempest-ServerMetadataTestJSON-1639770617-project-member] Waiting for the task: (returnval){ [ 1147.571155] env[61923]: value = "task-1378070" [ 1147.571155] env[61923]: _type = "Task" [ 1147.571155] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1147.577926] env[61923]: DEBUG oslo_vmware.api [None req-a84572db-25cb-45bf-8bab-bc3f62d5ec23 tempest-ServerMetadataTestJSON-1639770617 tempest-ServerMetadataTestJSON-1639770617-project-member] Task: {'id': task-1378070, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1147.917430] env[61923]: DEBUG nova.network.neutron [None req-53588979-1128-4b38-bd80-7fa4534341ad tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: 5dfbd79e-d43e-4769-b9f1-e8fdc09de31b] Successfully updated port: 0c9f3dd3-d6f5-47b0-90e7-d5e54eea95e1 {{(pid=61923) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1147.941021] env[61923]: DEBUG nova.compute.manager [req-0ca5b2f4-d603-4937-afc4-5e6589c4fa60 req-ac26df99-2983-414c-b8fc-e2e542c41d2f service nova] [instance: 5dfbd79e-d43e-4769-b9f1-e8fdc09de31b] Received event network-changed-0c9f3dd3-d6f5-47b0-90e7-d5e54eea95e1 {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1147.941021] env[61923]: DEBUG nova.compute.manager [req-0ca5b2f4-d603-4937-afc4-5e6589c4fa60 req-ac26df99-2983-414c-b8fc-e2e542c41d2f service nova] [instance: 5dfbd79e-d43e-4769-b9f1-e8fdc09de31b] Refreshing instance network info cache due to event network-changed-0c9f3dd3-d6f5-47b0-90e7-d5e54eea95e1. {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1147.941318] env[61923]: DEBUG oslo_concurrency.lockutils [req-0ca5b2f4-d603-4937-afc4-5e6589c4fa60 req-ac26df99-2983-414c-b8fc-e2e542c41d2f service nova] Acquiring lock "refresh_cache-5dfbd79e-d43e-4769-b9f1-e8fdc09de31b" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1147.941318] env[61923]: DEBUG oslo_concurrency.lockutils [req-0ca5b2f4-d603-4937-afc4-5e6589c4fa60 req-ac26df99-2983-414c-b8fc-e2e542c41d2f service nova] Acquired lock "refresh_cache-5dfbd79e-d43e-4769-b9f1-e8fdc09de31b" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1147.941414] env[61923]: DEBUG nova.network.neutron [req-0ca5b2f4-d603-4937-afc4-5e6589c4fa60 req-ac26df99-2983-414c-b8fc-e2e542c41d2f service nova] [instance: 5dfbd79e-d43e-4769-b9f1-e8fdc09de31b] Refreshing network info cache for port 0c9f3dd3-d6f5-47b0-90e7-d5e54eea95e1 {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1148.081093] env[61923]: DEBUG oslo_vmware.api [None req-a84572db-25cb-45bf-8bab-bc3f62d5ec23 tempest-ServerMetadataTestJSON-1639770617 tempest-ServerMetadataTestJSON-1639770617-project-member] Task: {'id': task-1378070, 'name': ReconfigVM_Task, 'duration_secs': 0.272024} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1148.081418] env[61923]: DEBUG nova.virt.vmwareapi.volumeops [None req-a84572db-25cb-45bf-8bab-bc3f62d5ec23 tempest-ServerMetadataTestJSON-1639770617 tempest-ServerMetadataTestJSON-1639770617-project-member] [instance: b82a241e-db60-40bc-a738-b961c3e38b43] Reconfigured VM instance instance-00000067 to attach disk [datastore2] b82a241e-db60-40bc-a738-b961c3e38b43/b82a241e-db60-40bc-a738-b961c3e38b43.vmdk or device None with type sparse {{(pid=61923) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1148.082016] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-d7dc01cc-f763-4842-80e2-56015d799e74 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1148.087508] env[61923]: DEBUG oslo_vmware.api [None req-a84572db-25cb-45bf-8bab-bc3f62d5ec23 tempest-ServerMetadataTestJSON-1639770617 tempest-ServerMetadataTestJSON-1639770617-project-member] Waiting for the task: (returnval){ [ 1148.087508] env[61923]: value = "task-1378071" [ 1148.087508] env[61923]: _type = "Task" [ 1148.087508] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1148.094775] env[61923]: DEBUG oslo_vmware.api [None req-a84572db-25cb-45bf-8bab-bc3f62d5ec23 tempest-ServerMetadataTestJSON-1639770617 tempest-ServerMetadataTestJSON-1639770617-project-member] Task: {'id': task-1378071, 'name': Rename_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1148.419984] env[61923]: DEBUG oslo_concurrency.lockutils [None req-53588979-1128-4b38-bd80-7fa4534341ad tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Acquiring lock "refresh_cache-5dfbd79e-d43e-4769-b9f1-e8fdc09de31b" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1148.473029] env[61923]: DEBUG nova.network.neutron [req-0ca5b2f4-d603-4937-afc4-5e6589c4fa60 req-ac26df99-2983-414c-b8fc-e2e542c41d2f service nova] [instance: 5dfbd79e-d43e-4769-b9f1-e8fdc09de31b] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1148.541729] env[61923]: DEBUG nova.network.neutron [req-0ca5b2f4-d603-4937-afc4-5e6589c4fa60 req-ac26df99-2983-414c-b8fc-e2e542c41d2f service nova] [instance: 5dfbd79e-d43e-4769-b9f1-e8fdc09de31b] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1148.596532] env[61923]: DEBUG oslo_vmware.api [None req-a84572db-25cb-45bf-8bab-bc3f62d5ec23 tempest-ServerMetadataTestJSON-1639770617 tempest-ServerMetadataTestJSON-1639770617-project-member] Task: {'id': task-1378071, 'name': Rename_Task, 'duration_secs': 0.126749} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1148.596807] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-a84572db-25cb-45bf-8bab-bc3f62d5ec23 tempest-ServerMetadataTestJSON-1639770617 tempest-ServerMetadataTestJSON-1639770617-project-member] [instance: b82a241e-db60-40bc-a738-b961c3e38b43] Powering on the VM {{(pid=61923) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1148.597061] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-19f89826-18b6-411c-8c45-b255d8a5f2d3 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1148.603533] env[61923]: DEBUG oslo_vmware.api [None req-a84572db-25cb-45bf-8bab-bc3f62d5ec23 tempest-ServerMetadataTestJSON-1639770617 tempest-ServerMetadataTestJSON-1639770617-project-member] Waiting for the task: (returnval){ [ 1148.603533] env[61923]: value = "task-1378072" [ 1148.603533] env[61923]: _type = "Task" [ 1148.603533] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1148.610714] env[61923]: DEBUG oslo_vmware.api [None req-a84572db-25cb-45bf-8bab-bc3f62d5ec23 tempest-ServerMetadataTestJSON-1639770617 tempest-ServerMetadataTestJSON-1639770617-project-member] Task: {'id': task-1378072, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1149.044471] env[61923]: DEBUG oslo_concurrency.lockutils [req-0ca5b2f4-d603-4937-afc4-5e6589c4fa60 req-ac26df99-2983-414c-b8fc-e2e542c41d2f service nova] Releasing lock "refresh_cache-5dfbd79e-d43e-4769-b9f1-e8fdc09de31b" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1149.044885] env[61923]: DEBUG oslo_concurrency.lockutils [None req-53588979-1128-4b38-bd80-7fa4534341ad tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Acquired lock "refresh_cache-5dfbd79e-d43e-4769-b9f1-e8fdc09de31b" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1149.045060] env[61923]: DEBUG nova.network.neutron [None req-53588979-1128-4b38-bd80-7fa4534341ad tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: 5dfbd79e-d43e-4769-b9f1-e8fdc09de31b] Building network info cache for instance {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1149.113865] env[61923]: DEBUG oslo_vmware.api [None req-a84572db-25cb-45bf-8bab-bc3f62d5ec23 tempest-ServerMetadataTestJSON-1639770617 tempest-ServerMetadataTestJSON-1639770617-project-member] Task: {'id': task-1378072, 'name': PowerOnVM_Task, 'duration_secs': 0.426718} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1149.114212] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-a84572db-25cb-45bf-8bab-bc3f62d5ec23 tempest-ServerMetadataTestJSON-1639770617 tempest-ServerMetadataTestJSON-1639770617-project-member] [instance: b82a241e-db60-40bc-a738-b961c3e38b43] Powered on the VM {{(pid=61923) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1149.114406] env[61923]: INFO nova.compute.manager [None req-a84572db-25cb-45bf-8bab-bc3f62d5ec23 tempest-ServerMetadataTestJSON-1639770617 tempest-ServerMetadataTestJSON-1639770617-project-member] [instance: b82a241e-db60-40bc-a738-b961c3e38b43] Took 6.69 seconds to spawn the instance on the hypervisor. [ 1149.114647] env[61923]: DEBUG nova.compute.manager [None req-a84572db-25cb-45bf-8bab-bc3f62d5ec23 tempest-ServerMetadataTestJSON-1639770617 tempest-ServerMetadataTestJSON-1639770617-project-member] [instance: b82a241e-db60-40bc-a738-b961c3e38b43] Checking state {{(pid=61923) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1149.115398] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d999e102-8880-41cb-b235-d6ddf47bf5cd {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1149.577077] env[61923]: DEBUG nova.network.neutron [None req-53588979-1128-4b38-bd80-7fa4534341ad tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: 5dfbd79e-d43e-4769-b9f1-e8fdc09de31b] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1149.632015] env[61923]: INFO nova.compute.manager [None req-a84572db-25cb-45bf-8bab-bc3f62d5ec23 tempest-ServerMetadataTestJSON-1639770617 tempest-ServerMetadataTestJSON-1639770617-project-member] [instance: b82a241e-db60-40bc-a738-b961c3e38b43] Took 12.69 seconds to build instance. [ 1149.727639] env[61923]: DEBUG nova.network.neutron [None req-53588979-1128-4b38-bd80-7fa4534341ad tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: 5dfbd79e-d43e-4769-b9f1-e8fdc09de31b] Updating instance_info_cache with network_info: [{"id": "0c9f3dd3-d6f5-47b0-90e7-d5e54eea95e1", "address": "fa:16:3e:87:b1:85", "network": {"id": "77aec6ca-d1be-4b39-9d51-08853a1183a1", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1813085414-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5a4db4f8aa194a219e80c538cc715d43", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "56b944d8-803d-43f2-945d-0f334ee4ea1c", "external-id": "nsx-vlan-transportzone-799", "segmentation_id": 799, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0c9f3dd3-d6", "ovs_interfaceid": "0c9f3dd3-d6f5-47b0-90e7-d5e54eea95e1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1150.133857] env[61923]: DEBUG oslo_concurrency.lockutils [None req-a84572db-25cb-45bf-8bab-bc3f62d5ec23 tempest-ServerMetadataTestJSON-1639770617 tempest-ServerMetadataTestJSON-1639770617-project-member] Lock "b82a241e-db60-40bc-a738-b961c3e38b43" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.202s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1150.230783] env[61923]: DEBUG oslo_concurrency.lockutils [None req-53588979-1128-4b38-bd80-7fa4534341ad tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Releasing lock "refresh_cache-5dfbd79e-d43e-4769-b9f1-e8fdc09de31b" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1150.231134] env[61923]: DEBUG nova.compute.manager [None req-53588979-1128-4b38-bd80-7fa4534341ad tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: 5dfbd79e-d43e-4769-b9f1-e8fdc09de31b] Instance network_info: |[{"id": "0c9f3dd3-d6f5-47b0-90e7-d5e54eea95e1", "address": "fa:16:3e:87:b1:85", "network": {"id": "77aec6ca-d1be-4b39-9d51-08853a1183a1", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1813085414-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5a4db4f8aa194a219e80c538cc715d43", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "56b944d8-803d-43f2-945d-0f334ee4ea1c", "external-id": "nsx-vlan-transportzone-799", "segmentation_id": 799, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0c9f3dd3-d6", "ovs_interfaceid": "0c9f3dd3-d6f5-47b0-90e7-d5e54eea95e1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61923) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1150.231609] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-53588979-1128-4b38-bd80-7fa4534341ad tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: 5dfbd79e-d43e-4769-b9f1-e8fdc09de31b] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:87:b1:85', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '56b944d8-803d-43f2-945d-0f334ee4ea1c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '0c9f3dd3-d6f5-47b0-90e7-d5e54eea95e1', 'vif_model': 'vmxnet3'}] {{(pid=61923) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1150.239305] env[61923]: DEBUG oslo.service.loopingcall [None req-53588979-1128-4b38-bd80-7fa4534341ad tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61923) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1150.239546] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5dfbd79e-d43e-4769-b9f1-e8fdc09de31b] Creating VM on the ESX host {{(pid=61923) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1150.239794] env[61923]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-4d928959-d96a-4fe4-b593-7fe2a8a97261 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1150.259867] env[61923]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1150.259867] env[61923]: value = "task-1378073" [ 1150.259867] env[61923]: _type = "Task" [ 1150.259867] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1150.267405] env[61923]: DEBUG oslo_vmware.api [-] Task: {'id': task-1378073, 'name': CreateVM_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1150.769397] env[61923]: DEBUG oslo_vmware.api [-] Task: {'id': task-1378073, 'name': CreateVM_Task, 'duration_secs': 0.300778} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1150.769575] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5dfbd79e-d43e-4769-b9f1-e8fdc09de31b] Created VM on the ESX host {{(pid=61923) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1150.770344] env[61923]: DEBUG oslo_concurrency.lockutils [None req-53588979-1128-4b38-bd80-7fa4534341ad tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1150.770517] env[61923]: DEBUG oslo_concurrency.lockutils [None req-53588979-1128-4b38-bd80-7fa4534341ad tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1150.770835] env[61923]: DEBUG oslo_concurrency.lockutils [None req-53588979-1128-4b38-bd80-7fa4534341ad tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1150.771103] env[61923]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a89399a3-678e-4c93-8afa-f89c64c0e0e2 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1150.775287] env[61923]: DEBUG oslo_vmware.api [None req-53588979-1128-4b38-bd80-7fa4534341ad tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Waiting for the task: (returnval){ [ 1150.775287] env[61923]: value = "session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52f0cdc2-3b35-ba60-32f4-196ea6f8db35" [ 1150.775287] env[61923]: _type = "Task" [ 1150.775287] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1150.782305] env[61923]: DEBUG oslo_vmware.api [None req-53588979-1128-4b38-bd80-7fa4534341ad tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52f0cdc2-3b35-ba60-32f4-196ea6f8db35, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1151.161060] env[61923]: DEBUG oslo_concurrency.lockutils [None req-96f3922e-3100-4537-9b23-96fbafe6520a tempest-ServerMetadataTestJSON-1639770617 tempest-ServerMetadataTestJSON-1639770617-project-member] Acquiring lock "b82a241e-db60-40bc-a738-b961c3e38b43" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1151.161459] env[61923]: DEBUG oslo_concurrency.lockutils [None req-96f3922e-3100-4537-9b23-96fbafe6520a tempest-ServerMetadataTestJSON-1639770617 tempest-ServerMetadataTestJSON-1639770617-project-member] Lock "b82a241e-db60-40bc-a738-b961c3e38b43" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1151.161558] env[61923]: DEBUG oslo_concurrency.lockutils [None req-96f3922e-3100-4537-9b23-96fbafe6520a tempest-ServerMetadataTestJSON-1639770617 tempest-ServerMetadataTestJSON-1639770617-project-member] Acquiring lock "b82a241e-db60-40bc-a738-b961c3e38b43-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1151.161750] env[61923]: DEBUG oslo_concurrency.lockutils [None req-96f3922e-3100-4537-9b23-96fbafe6520a tempest-ServerMetadataTestJSON-1639770617 tempest-ServerMetadataTestJSON-1639770617-project-member] Lock "b82a241e-db60-40bc-a738-b961c3e38b43-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1151.161933] env[61923]: DEBUG oslo_concurrency.lockutils [None req-96f3922e-3100-4537-9b23-96fbafe6520a tempest-ServerMetadataTestJSON-1639770617 tempest-ServerMetadataTestJSON-1639770617-project-member] Lock "b82a241e-db60-40bc-a738-b961c3e38b43-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1151.164250] env[61923]: INFO nova.compute.manager [None req-96f3922e-3100-4537-9b23-96fbafe6520a tempest-ServerMetadataTestJSON-1639770617 tempest-ServerMetadataTestJSON-1639770617-project-member] [instance: b82a241e-db60-40bc-a738-b961c3e38b43] Terminating instance [ 1151.165962] env[61923]: DEBUG nova.compute.manager [None req-96f3922e-3100-4537-9b23-96fbafe6520a tempest-ServerMetadataTestJSON-1639770617 tempest-ServerMetadataTestJSON-1639770617-project-member] [instance: b82a241e-db60-40bc-a738-b961c3e38b43] Start destroying the instance on the hypervisor. {{(pid=61923) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1151.166174] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-96f3922e-3100-4537-9b23-96fbafe6520a tempest-ServerMetadataTestJSON-1639770617 tempest-ServerMetadataTestJSON-1639770617-project-member] [instance: b82a241e-db60-40bc-a738-b961c3e38b43] Destroying instance {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1151.166984] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-596b741b-1f54-44db-93f4-b213fd951e12 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1151.174555] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-96f3922e-3100-4537-9b23-96fbafe6520a tempest-ServerMetadataTestJSON-1639770617 tempest-ServerMetadataTestJSON-1639770617-project-member] [instance: b82a241e-db60-40bc-a738-b961c3e38b43] Powering off the VM {{(pid=61923) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1151.174773] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9032c088-a932-4d15-835c-51c92823f6e9 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1151.181238] env[61923]: DEBUG oslo_vmware.api [None req-96f3922e-3100-4537-9b23-96fbafe6520a tempest-ServerMetadataTestJSON-1639770617 tempest-ServerMetadataTestJSON-1639770617-project-member] Waiting for the task: (returnval){ [ 1151.181238] env[61923]: value = "task-1378074" [ 1151.181238] env[61923]: _type = "Task" [ 1151.181238] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1151.188532] env[61923]: DEBUG oslo_vmware.api [None req-96f3922e-3100-4537-9b23-96fbafe6520a tempest-ServerMetadataTestJSON-1639770617 tempest-ServerMetadataTestJSON-1639770617-project-member] Task: {'id': task-1378074, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1151.286633] env[61923]: DEBUG oslo_vmware.api [None req-53588979-1128-4b38-bd80-7fa4534341ad tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52f0cdc2-3b35-ba60-32f4-196ea6f8db35, 'name': SearchDatastore_Task, 'duration_secs': 0.010528} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1151.286957] env[61923]: DEBUG oslo_concurrency.lockutils [None req-53588979-1128-4b38-bd80-7fa4534341ad tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1151.287227] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-53588979-1128-4b38-bd80-7fa4534341ad tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: 5dfbd79e-d43e-4769-b9f1-e8fdc09de31b] Processing image 0f153f63-ae0a-45b1-b7f5-7f9b673c947f {{(pid=61923) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1151.287471] env[61923]: DEBUG oslo_concurrency.lockutils [None req-53588979-1128-4b38-bd80-7fa4534341ad tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1151.287638] env[61923]: DEBUG oslo_concurrency.lockutils [None req-53588979-1128-4b38-bd80-7fa4534341ad tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1151.287870] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-53588979-1128-4b38-bd80-7fa4534341ad tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61923) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1151.288191] env[61923]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-9581c880-7dfa-48d7-ae8e-7636438714b2 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1151.296505] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-53588979-1128-4b38-bd80-7fa4534341ad tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61923) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1151.296727] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-53588979-1128-4b38-bd80-7fa4534341ad tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61923) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1151.297547] env[61923]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-71f06b80-f8e9-4480-91d1-e17b7f5d02cf {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1151.302882] env[61923]: DEBUG oslo_vmware.api [None req-53588979-1128-4b38-bd80-7fa4534341ad tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Waiting for the task: (returnval){ [ 1151.302882] env[61923]: value = "session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]5233671a-f2bd-5f18-1e9d-1668313e8603" [ 1151.302882] env[61923]: _type = "Task" [ 1151.302882] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1151.310101] env[61923]: DEBUG oslo_vmware.api [None req-53588979-1128-4b38-bd80-7fa4534341ad tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]5233671a-f2bd-5f18-1e9d-1668313e8603, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1151.692269] env[61923]: DEBUG oslo_vmware.api [None req-96f3922e-3100-4537-9b23-96fbafe6520a tempest-ServerMetadataTestJSON-1639770617 tempest-ServerMetadataTestJSON-1639770617-project-member] Task: {'id': task-1378074, 'name': PowerOffVM_Task, 'duration_secs': 0.193618} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1151.692566] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-96f3922e-3100-4537-9b23-96fbafe6520a tempest-ServerMetadataTestJSON-1639770617 tempest-ServerMetadataTestJSON-1639770617-project-member] [instance: b82a241e-db60-40bc-a738-b961c3e38b43] Powered off the VM {{(pid=61923) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1151.692911] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-96f3922e-3100-4537-9b23-96fbafe6520a tempest-ServerMetadataTestJSON-1639770617 tempest-ServerMetadataTestJSON-1639770617-project-member] [instance: b82a241e-db60-40bc-a738-b961c3e38b43] Unregistering the VM {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1151.693056] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-696b8bf8-4841-4953-9012-f4ccc2b8bf8b {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1151.754836] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-96f3922e-3100-4537-9b23-96fbafe6520a tempest-ServerMetadataTestJSON-1639770617 tempest-ServerMetadataTestJSON-1639770617-project-member] [instance: b82a241e-db60-40bc-a738-b961c3e38b43] Unregistered the VM {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1151.755126] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-96f3922e-3100-4537-9b23-96fbafe6520a tempest-ServerMetadataTestJSON-1639770617 tempest-ServerMetadataTestJSON-1639770617-project-member] [instance: b82a241e-db60-40bc-a738-b961c3e38b43] Deleting contents of the VM from datastore datastore2 {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1151.755322] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-96f3922e-3100-4537-9b23-96fbafe6520a tempest-ServerMetadataTestJSON-1639770617 tempest-ServerMetadataTestJSON-1639770617-project-member] Deleting the datastore file [datastore2] b82a241e-db60-40bc-a738-b961c3e38b43 {{(pid=61923) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1151.755618] env[61923]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-4406c3dd-7207-44e9-9645-9d40d836d317 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1151.761589] env[61923]: DEBUG oslo_vmware.api [None req-96f3922e-3100-4537-9b23-96fbafe6520a tempest-ServerMetadataTestJSON-1639770617 tempest-ServerMetadataTestJSON-1639770617-project-member] Waiting for the task: (returnval){ [ 1151.761589] env[61923]: value = "task-1378076" [ 1151.761589] env[61923]: _type = "Task" [ 1151.761589] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1151.770463] env[61923]: DEBUG oslo_vmware.api [None req-96f3922e-3100-4537-9b23-96fbafe6520a tempest-ServerMetadataTestJSON-1639770617 tempest-ServerMetadataTestJSON-1639770617-project-member] Task: {'id': task-1378076, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1151.813035] env[61923]: DEBUG oslo_vmware.api [None req-53588979-1128-4b38-bd80-7fa4534341ad tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]5233671a-f2bd-5f18-1e9d-1668313e8603, 'name': SearchDatastore_Task, 'duration_secs': 0.007994} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1151.814033] env[61923]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-39df5629-cf28-4636-9294-8f8f7b469f7e {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1151.819273] env[61923]: DEBUG oslo_vmware.api [None req-53588979-1128-4b38-bd80-7fa4534341ad tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Waiting for the task: (returnval){ [ 1151.819273] env[61923]: value = "session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]529fbe40-a874-f2f1-dfc2-9b3faf6a68cb" [ 1151.819273] env[61923]: _type = "Task" [ 1151.819273] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1151.826923] env[61923]: DEBUG oslo_vmware.api [None req-53588979-1128-4b38-bd80-7fa4534341ad tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]529fbe40-a874-f2f1-dfc2-9b3faf6a68cb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1152.270900] env[61923]: DEBUG oslo_vmware.api [None req-96f3922e-3100-4537-9b23-96fbafe6520a tempest-ServerMetadataTestJSON-1639770617 tempest-ServerMetadataTestJSON-1639770617-project-member] Task: {'id': task-1378076, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.156142} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1152.271292] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-96f3922e-3100-4537-9b23-96fbafe6520a tempest-ServerMetadataTestJSON-1639770617 tempest-ServerMetadataTestJSON-1639770617-project-member] Deleted the datastore file {{(pid=61923) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1152.271360] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-96f3922e-3100-4537-9b23-96fbafe6520a tempest-ServerMetadataTestJSON-1639770617 tempest-ServerMetadataTestJSON-1639770617-project-member] [instance: b82a241e-db60-40bc-a738-b961c3e38b43] Deleted contents of the VM from datastore datastore2 {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1152.271559] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-96f3922e-3100-4537-9b23-96fbafe6520a tempest-ServerMetadataTestJSON-1639770617 tempest-ServerMetadataTestJSON-1639770617-project-member] [instance: b82a241e-db60-40bc-a738-b961c3e38b43] Instance destroyed {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1152.271742] env[61923]: INFO nova.compute.manager [None req-96f3922e-3100-4537-9b23-96fbafe6520a tempest-ServerMetadataTestJSON-1639770617 tempest-ServerMetadataTestJSON-1639770617-project-member] [instance: b82a241e-db60-40bc-a738-b961c3e38b43] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1152.271995] env[61923]: DEBUG oslo.service.loopingcall [None req-96f3922e-3100-4537-9b23-96fbafe6520a tempest-ServerMetadataTestJSON-1639770617 tempest-ServerMetadataTestJSON-1639770617-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61923) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1152.272215] env[61923]: DEBUG nova.compute.manager [-] [instance: b82a241e-db60-40bc-a738-b961c3e38b43] Deallocating network for instance {{(pid=61923) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1152.272326] env[61923]: DEBUG nova.network.neutron [-] [instance: b82a241e-db60-40bc-a738-b961c3e38b43] deallocate_for_instance() {{(pid=61923) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1152.330715] env[61923]: DEBUG oslo_vmware.api [None req-53588979-1128-4b38-bd80-7fa4534341ad tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]529fbe40-a874-f2f1-dfc2-9b3faf6a68cb, 'name': SearchDatastore_Task, 'duration_secs': 0.009115} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1152.330972] env[61923]: DEBUG oslo_concurrency.lockutils [None req-53588979-1128-4b38-bd80-7fa4534341ad tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1152.331277] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-53588979-1128-4b38-bd80-7fa4534341ad tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk to [datastore1] 5dfbd79e-d43e-4769-b9f1-e8fdc09de31b/5dfbd79e-d43e-4769-b9f1-e8fdc09de31b.vmdk {{(pid=61923) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1152.331567] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-60d1a95b-047e-4da4-8b39-e721234d5879 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1152.337338] env[61923]: DEBUG oslo_vmware.api [None req-53588979-1128-4b38-bd80-7fa4534341ad tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Waiting for the task: (returnval){ [ 1152.337338] env[61923]: value = "task-1378077" [ 1152.337338] env[61923]: _type = "Task" [ 1152.337338] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1152.344540] env[61923]: DEBUG oslo_vmware.api [None req-53588979-1128-4b38-bd80-7fa4534341ad tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Task: {'id': task-1378077, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1152.530961] env[61923]: DEBUG nova.compute.manager [req-b8ca6132-edba-4148-9373-32c10e890c86 req-562b16b8-ea08-498b-9884-49839f88905b service nova] [instance: b82a241e-db60-40bc-a738-b961c3e38b43] Received event network-vif-deleted-919e54ef-c0f6-4dde-b5e4-be987b7e3823 {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1152.531146] env[61923]: INFO nova.compute.manager [req-b8ca6132-edba-4148-9373-32c10e890c86 req-562b16b8-ea08-498b-9884-49839f88905b service nova] [instance: b82a241e-db60-40bc-a738-b961c3e38b43] Neutron deleted interface 919e54ef-c0f6-4dde-b5e4-be987b7e3823; detaching it from the instance and deleting it from the info cache [ 1152.531406] env[61923]: DEBUG nova.network.neutron [req-b8ca6132-edba-4148-9373-32c10e890c86 req-562b16b8-ea08-498b-9884-49839f88905b service nova] [instance: b82a241e-db60-40bc-a738-b961c3e38b43] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1152.847092] env[61923]: DEBUG oslo_vmware.api [None req-53588979-1128-4b38-bd80-7fa4534341ad tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Task: {'id': task-1378077, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.426959} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1152.847371] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-53588979-1128-4b38-bd80-7fa4534341ad tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk to [datastore1] 5dfbd79e-d43e-4769-b9f1-e8fdc09de31b/5dfbd79e-d43e-4769-b9f1-e8fdc09de31b.vmdk {{(pid=61923) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1152.847591] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-53588979-1128-4b38-bd80-7fa4534341ad tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: 5dfbd79e-d43e-4769-b9f1-e8fdc09de31b] Extending root virtual disk to 1048576 {{(pid=61923) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1152.847842] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-b2bf70db-8ddf-473b-8b01-70ce11a7c27e {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1152.854010] env[61923]: DEBUG oslo_vmware.api [None req-53588979-1128-4b38-bd80-7fa4534341ad tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Waiting for the task: (returnval){ [ 1152.854010] env[61923]: value = "task-1378078" [ 1152.854010] env[61923]: _type = "Task" [ 1152.854010] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1152.862212] env[61923]: DEBUG oslo_vmware.api [None req-53588979-1128-4b38-bd80-7fa4534341ad tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Task: {'id': task-1378078, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1153.006930] env[61923]: DEBUG nova.network.neutron [-] [instance: b82a241e-db60-40bc-a738-b961c3e38b43] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1153.034579] env[61923]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-3f534006-62c6-4e1a-89b5-4dc2c273aaaa {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1153.043524] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b6fbd03-2d6e-43af-89b6-3bf4b217b1c5 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1153.065876] env[61923]: DEBUG nova.compute.manager [req-b8ca6132-edba-4148-9373-32c10e890c86 req-562b16b8-ea08-498b-9884-49839f88905b service nova] [instance: b82a241e-db60-40bc-a738-b961c3e38b43] Detach interface failed, port_id=919e54ef-c0f6-4dde-b5e4-be987b7e3823, reason: Instance b82a241e-db60-40bc-a738-b961c3e38b43 could not be found. {{(pid=61923) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1153.363762] env[61923]: DEBUG oslo_vmware.api [None req-53588979-1128-4b38-bd80-7fa4534341ad tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Task: {'id': task-1378078, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.059258} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1153.364085] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-53588979-1128-4b38-bd80-7fa4534341ad tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: 5dfbd79e-d43e-4769-b9f1-e8fdc09de31b] Extended root virtual disk {{(pid=61923) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1153.364845] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30797cbb-4c33-4a8d-bc0c-c5f92fc3fe81 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1153.386186] env[61923]: DEBUG nova.virt.vmwareapi.volumeops [None req-53588979-1128-4b38-bd80-7fa4534341ad tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: 5dfbd79e-d43e-4769-b9f1-e8fdc09de31b] Reconfiguring VM instance instance-00000068 to attach disk [datastore1] 5dfbd79e-d43e-4769-b9f1-e8fdc09de31b/5dfbd79e-d43e-4769-b9f1-e8fdc09de31b.vmdk or device None with type sparse {{(pid=61923) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1153.386439] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-fb29038c-62e4-4a20-b066-662c21b01dfb {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1153.405524] env[61923]: DEBUG oslo_vmware.api [None req-53588979-1128-4b38-bd80-7fa4534341ad tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Waiting for the task: (returnval){ [ 1153.405524] env[61923]: value = "task-1378079" [ 1153.405524] env[61923]: _type = "Task" [ 1153.405524] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1153.412760] env[61923]: DEBUG oslo_vmware.api [None req-53588979-1128-4b38-bd80-7fa4534341ad tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Task: {'id': task-1378079, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1153.509794] env[61923]: INFO nova.compute.manager [-] [instance: b82a241e-db60-40bc-a738-b961c3e38b43] Took 1.24 seconds to deallocate network for instance. [ 1153.914855] env[61923]: DEBUG oslo_vmware.api [None req-53588979-1128-4b38-bd80-7fa4534341ad tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Task: {'id': task-1378079, 'name': ReconfigVM_Task, 'duration_secs': 0.253375} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1153.915205] env[61923]: DEBUG nova.virt.vmwareapi.volumeops [None req-53588979-1128-4b38-bd80-7fa4534341ad tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: 5dfbd79e-d43e-4769-b9f1-e8fdc09de31b] Reconfigured VM instance instance-00000068 to attach disk [datastore1] 5dfbd79e-d43e-4769-b9f1-e8fdc09de31b/5dfbd79e-d43e-4769-b9f1-e8fdc09de31b.vmdk or device None with type sparse {{(pid=61923) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1153.915921] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-8984f447-93f8-46bf-9625-65a05928b9cc {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1153.922337] env[61923]: DEBUG oslo_vmware.api [None req-53588979-1128-4b38-bd80-7fa4534341ad tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Waiting for the task: (returnval){ [ 1153.922337] env[61923]: value = "task-1378080" [ 1153.922337] env[61923]: _type = "Task" [ 1153.922337] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1153.929673] env[61923]: DEBUG oslo_vmware.api [None req-53588979-1128-4b38-bd80-7fa4534341ad tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Task: {'id': task-1378080, 'name': Rename_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1154.017319] env[61923]: DEBUG oslo_concurrency.lockutils [None req-96f3922e-3100-4537-9b23-96fbafe6520a tempest-ServerMetadataTestJSON-1639770617 tempest-ServerMetadataTestJSON-1639770617-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1154.017686] env[61923]: DEBUG oslo_concurrency.lockutils [None req-96f3922e-3100-4537-9b23-96fbafe6520a tempest-ServerMetadataTestJSON-1639770617 tempest-ServerMetadataTestJSON-1639770617-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1154.017925] env[61923]: DEBUG nova.objects.instance [None req-96f3922e-3100-4537-9b23-96fbafe6520a tempest-ServerMetadataTestJSON-1639770617 tempest-ServerMetadataTestJSON-1639770617-project-member] Lazy-loading 'resources' on Instance uuid b82a241e-db60-40bc-a738-b961c3e38b43 {{(pid=61923) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1154.433017] env[61923]: DEBUG oslo_vmware.api [None req-53588979-1128-4b38-bd80-7fa4534341ad tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Task: {'id': task-1378080, 'name': Rename_Task, 'duration_secs': 0.126282} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1154.433391] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-53588979-1128-4b38-bd80-7fa4534341ad tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: 5dfbd79e-d43e-4769-b9f1-e8fdc09de31b] Powering on the VM {{(pid=61923) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1154.433541] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-4ac78b02-fd65-4a5c-822a-0099d7d8493b {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1154.451237] env[61923]: DEBUG oslo_vmware.api [None req-53588979-1128-4b38-bd80-7fa4534341ad tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Waiting for the task: (returnval){ [ 1154.451237] env[61923]: value = "task-1378081" [ 1154.451237] env[61923]: _type = "Task" [ 1154.451237] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1154.462436] env[61923]: DEBUG oslo_vmware.api [None req-53588979-1128-4b38-bd80-7fa4534341ad tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Task: {'id': task-1378081, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1154.563536] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98fe9832-ccad-4329-8a75-20e4c72cf38f {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1154.571298] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66966f41-8427-4da5-83db-d1994a59e9be {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1154.602100] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-744b67a4-3e94-416d-8827-0e55e0d497d7 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1154.608912] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d9099a2-873d-4288-8113-df8009d7f2b5 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1154.621486] env[61923]: DEBUG nova.compute.provider_tree [None req-96f3922e-3100-4537-9b23-96fbafe6520a tempest-ServerMetadataTestJSON-1639770617 tempest-ServerMetadataTestJSON-1639770617-project-member] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1154.960693] env[61923]: DEBUG oslo_vmware.api [None req-53588979-1128-4b38-bd80-7fa4534341ad tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Task: {'id': task-1378081, 'name': PowerOnVM_Task, 'duration_secs': 0.441703} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1154.960973] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-53588979-1128-4b38-bd80-7fa4534341ad tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: 5dfbd79e-d43e-4769-b9f1-e8fdc09de31b] Powered on the VM {{(pid=61923) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1154.961204] env[61923]: INFO nova.compute.manager [None req-53588979-1128-4b38-bd80-7fa4534341ad tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: 5dfbd79e-d43e-4769-b9f1-e8fdc09de31b] Took 7.74 seconds to spawn the instance on the hypervisor. [ 1154.961395] env[61923]: DEBUG nova.compute.manager [None req-53588979-1128-4b38-bd80-7fa4534341ad tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: 5dfbd79e-d43e-4769-b9f1-e8fdc09de31b] Checking state {{(pid=61923) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1154.962182] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42197a19-ec5d-4c88-a8bf-5192283163a4 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1155.125237] env[61923]: DEBUG nova.scheduler.client.report [None req-96f3922e-3100-4537-9b23-96fbafe6520a tempest-ServerMetadataTestJSON-1639770617 tempest-ServerMetadataTestJSON-1639770617-project-member] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1155.478778] env[61923]: INFO nova.compute.manager [None req-53588979-1128-4b38-bd80-7fa4534341ad tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: 5dfbd79e-d43e-4769-b9f1-e8fdc09de31b] Took 14.20 seconds to build instance. [ 1155.633333] env[61923]: DEBUG oslo_concurrency.lockutils [None req-96f3922e-3100-4537-9b23-96fbafe6520a tempest-ServerMetadataTestJSON-1639770617 tempest-ServerMetadataTestJSON-1639770617-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.616s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1155.654042] env[61923]: INFO nova.scheduler.client.report [None req-96f3922e-3100-4537-9b23-96fbafe6520a tempest-ServerMetadataTestJSON-1639770617 tempest-ServerMetadataTestJSON-1639770617-project-member] Deleted allocations for instance b82a241e-db60-40bc-a738-b961c3e38b43 [ 1155.980775] env[61923]: DEBUG oslo_concurrency.lockutils [None req-53588979-1128-4b38-bd80-7fa4534341ad tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Lock "5dfbd79e-d43e-4769-b9f1-e8fdc09de31b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.707s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1156.162322] env[61923]: DEBUG oslo_concurrency.lockutils [None req-96f3922e-3100-4537-9b23-96fbafe6520a tempest-ServerMetadataTestJSON-1639770617 tempest-ServerMetadataTestJSON-1639770617-project-member] Lock "b82a241e-db60-40bc-a738-b961c3e38b43" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.001s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1157.646216] env[61923]: DEBUG oslo_concurrency.lockutils [None req-6ba47f25-2dd0-44be-a91b-f9f659a3dd74 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Acquiring lock "5dfbd79e-d43e-4769-b9f1-e8fdc09de31b" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1157.646216] env[61923]: DEBUG oslo_concurrency.lockutils [None req-6ba47f25-2dd0-44be-a91b-f9f659a3dd74 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Lock "5dfbd79e-d43e-4769-b9f1-e8fdc09de31b" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1157.646216] env[61923]: DEBUG oslo_concurrency.lockutils [None req-6ba47f25-2dd0-44be-a91b-f9f659a3dd74 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Acquiring lock "5dfbd79e-d43e-4769-b9f1-e8fdc09de31b-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1157.646216] env[61923]: DEBUG oslo_concurrency.lockutils [None req-6ba47f25-2dd0-44be-a91b-f9f659a3dd74 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Lock "5dfbd79e-d43e-4769-b9f1-e8fdc09de31b-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1157.646576] env[61923]: DEBUG oslo_concurrency.lockutils [None req-6ba47f25-2dd0-44be-a91b-f9f659a3dd74 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Lock "5dfbd79e-d43e-4769-b9f1-e8fdc09de31b-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1157.649258] env[61923]: INFO nova.compute.manager [None req-6ba47f25-2dd0-44be-a91b-f9f659a3dd74 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: 5dfbd79e-d43e-4769-b9f1-e8fdc09de31b] Terminating instance [ 1157.650773] env[61923]: DEBUG nova.compute.manager [None req-6ba47f25-2dd0-44be-a91b-f9f659a3dd74 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: 5dfbd79e-d43e-4769-b9f1-e8fdc09de31b] Start destroying the instance on the hypervisor. {{(pid=61923) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1157.650970] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-6ba47f25-2dd0-44be-a91b-f9f659a3dd74 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: 5dfbd79e-d43e-4769-b9f1-e8fdc09de31b] Destroying instance {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1157.652218] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45edeb58-066b-4400-a190-d23d5c684a4d {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1157.659968] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-6ba47f25-2dd0-44be-a91b-f9f659a3dd74 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: 5dfbd79e-d43e-4769-b9f1-e8fdc09de31b] Powering off the VM {{(pid=61923) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1157.659968] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-77148415-ec0b-4cf1-8e34-29e3c4d99f65 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1157.667273] env[61923]: DEBUG oslo_vmware.api [None req-6ba47f25-2dd0-44be-a91b-f9f659a3dd74 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Waiting for the task: (returnval){ [ 1157.667273] env[61923]: value = "task-1378082" [ 1157.667273] env[61923]: _type = "Task" [ 1157.667273] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1157.676176] env[61923]: DEBUG oslo_vmware.api [None req-6ba47f25-2dd0-44be-a91b-f9f659a3dd74 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Task: {'id': task-1378082, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1158.176770] env[61923]: DEBUG oslo_vmware.api [None req-6ba47f25-2dd0-44be-a91b-f9f659a3dd74 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Task: {'id': task-1378082, 'name': PowerOffVM_Task, 'duration_secs': 0.194109} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1158.177063] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-6ba47f25-2dd0-44be-a91b-f9f659a3dd74 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: 5dfbd79e-d43e-4769-b9f1-e8fdc09de31b] Powered off the VM {{(pid=61923) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1158.177250] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-6ba47f25-2dd0-44be-a91b-f9f659a3dd74 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: 5dfbd79e-d43e-4769-b9f1-e8fdc09de31b] Unregistering the VM {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1158.177493] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e6415cb1-a872-4241-a660-c5ebadd1174e {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.239487] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-6ba47f25-2dd0-44be-a91b-f9f659a3dd74 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: 5dfbd79e-d43e-4769-b9f1-e8fdc09de31b] Unregistered the VM {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1158.239726] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-6ba47f25-2dd0-44be-a91b-f9f659a3dd74 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: 5dfbd79e-d43e-4769-b9f1-e8fdc09de31b] Deleting contents of the VM from datastore datastore1 {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1158.239914] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-6ba47f25-2dd0-44be-a91b-f9f659a3dd74 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Deleting the datastore file [datastore1] 5dfbd79e-d43e-4769-b9f1-e8fdc09de31b {{(pid=61923) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1158.240226] env[61923]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-14726b44-9827-4a51-83a6-2d5ed8b7a895 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.246039] env[61923]: DEBUG oslo_vmware.api [None req-6ba47f25-2dd0-44be-a91b-f9f659a3dd74 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Waiting for the task: (returnval){ [ 1158.246039] env[61923]: value = "task-1378084" [ 1158.246039] env[61923]: _type = "Task" [ 1158.246039] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1158.253314] env[61923]: DEBUG oslo_vmware.api [None req-6ba47f25-2dd0-44be-a91b-f9f659a3dd74 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Task: {'id': task-1378084, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1158.756133] env[61923]: DEBUG oslo_vmware.api [None req-6ba47f25-2dd0-44be-a91b-f9f659a3dd74 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Task: {'id': task-1378084, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.126698} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1158.756457] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-6ba47f25-2dd0-44be-a91b-f9f659a3dd74 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Deleted the datastore file {{(pid=61923) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1158.756544] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-6ba47f25-2dd0-44be-a91b-f9f659a3dd74 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: 5dfbd79e-d43e-4769-b9f1-e8fdc09de31b] Deleted contents of the VM from datastore datastore1 {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1158.756722] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-6ba47f25-2dd0-44be-a91b-f9f659a3dd74 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: 5dfbd79e-d43e-4769-b9f1-e8fdc09de31b] Instance destroyed {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1158.756896] env[61923]: INFO nova.compute.manager [None req-6ba47f25-2dd0-44be-a91b-f9f659a3dd74 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] [instance: 5dfbd79e-d43e-4769-b9f1-e8fdc09de31b] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1158.757189] env[61923]: DEBUG oslo.service.loopingcall [None req-6ba47f25-2dd0-44be-a91b-f9f659a3dd74 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61923) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1158.757355] env[61923]: DEBUG nova.compute.manager [-] [instance: 5dfbd79e-d43e-4769-b9f1-e8fdc09de31b] Deallocating network for instance {{(pid=61923) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1158.757448] env[61923]: DEBUG nova.network.neutron [-] [instance: 5dfbd79e-d43e-4769-b9f1-e8fdc09de31b] deallocate_for_instance() {{(pid=61923) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1159.177777] env[61923]: DEBUG nova.compute.manager [req-5385c1fb-3de8-4df1-b13e-b0e88b2a5131 req-12a9d137-8035-4379-ba02-39c2eefde5ea service nova] [instance: 5dfbd79e-d43e-4769-b9f1-e8fdc09de31b] Received event network-vif-deleted-0c9f3dd3-d6f5-47b0-90e7-d5e54eea95e1 {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1159.178096] env[61923]: INFO nova.compute.manager [req-5385c1fb-3de8-4df1-b13e-b0e88b2a5131 req-12a9d137-8035-4379-ba02-39c2eefde5ea service nova] [instance: 5dfbd79e-d43e-4769-b9f1-e8fdc09de31b] Neutron deleted interface 0c9f3dd3-d6f5-47b0-90e7-d5e54eea95e1; detaching it from the instance and deleting it from the info cache [ 1159.178399] env[61923]: DEBUG nova.network.neutron [req-5385c1fb-3de8-4df1-b13e-b0e88b2a5131 req-12a9d137-8035-4379-ba02-39c2eefde5ea service nova] [instance: 5dfbd79e-d43e-4769-b9f1-e8fdc09de31b] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1159.652677] env[61923]: DEBUG nova.network.neutron [-] [instance: 5dfbd79e-d43e-4769-b9f1-e8fdc09de31b] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1159.680909] env[61923]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-dbd4ecd7-7b5a-4a81-8c68-032c4a20ba33 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1159.691297] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7296cd00-f8d8-425b-8294-1672bb1cf097 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1159.716134] env[61923]: DEBUG nova.compute.manager [req-5385c1fb-3de8-4df1-b13e-b0e88b2a5131 req-12a9d137-8035-4379-ba02-39c2eefde5ea service nova] [instance: 5dfbd79e-d43e-4769-b9f1-e8fdc09de31b] Detach interface failed, port_id=0c9f3dd3-d6f5-47b0-90e7-d5e54eea95e1, reason: Instance 5dfbd79e-d43e-4769-b9f1-e8fdc09de31b could not be found. {{(pid=61923) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1160.155858] env[61923]: INFO nova.compute.manager [-] [instance: 5dfbd79e-d43e-4769-b9f1-e8fdc09de31b] Took 1.40 seconds to deallocate network for instance. [ 1160.662475] env[61923]: DEBUG oslo_concurrency.lockutils [None req-6ba47f25-2dd0-44be-a91b-f9f659a3dd74 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1160.662762] env[61923]: DEBUG oslo_concurrency.lockutils [None req-6ba47f25-2dd0-44be-a91b-f9f659a3dd74 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1160.662992] env[61923]: DEBUG nova.objects.instance [None req-6ba47f25-2dd0-44be-a91b-f9f659a3dd74 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Lazy-loading 'resources' on Instance uuid 5dfbd79e-d43e-4769-b9f1-e8fdc09de31b {{(pid=61923) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1161.198877] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f3915f0-2fc5-4ea1-95f0-d5ad1b952da4 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1161.206283] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55fa4610-9c9f-47fe-bee8-b7258ced2b1d {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1161.234978] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bbe07195-2923-4b51-85d4-e40009799d48 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1161.241977] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee258ed9-e92b-41ce-87e8-31181fbdd270 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1161.254972] env[61923]: DEBUG nova.compute.provider_tree [None req-6ba47f25-2dd0-44be-a91b-f9f659a3dd74 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1161.758598] env[61923]: DEBUG nova.scheduler.client.report [None req-6ba47f25-2dd0-44be-a91b-f9f659a3dd74 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1162.265060] env[61923]: DEBUG oslo_concurrency.lockutils [None req-6ba47f25-2dd0-44be-a91b-f9f659a3dd74 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.602s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1162.284186] env[61923]: INFO nova.scheduler.client.report [None req-6ba47f25-2dd0-44be-a91b-f9f659a3dd74 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Deleted allocations for instance 5dfbd79e-d43e-4769-b9f1-e8fdc09de31b [ 1162.793186] env[61923]: DEBUG oslo_concurrency.lockutils [None req-6ba47f25-2dd0-44be-a91b-f9f659a3dd74 tempest-ServerDiskConfigTestJSON-665621459 tempest-ServerDiskConfigTestJSON-665621459-project-member] Lock "5dfbd79e-d43e-4769-b9f1-e8fdc09de31b" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.147s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1168.163344] env[61923]: DEBUG oslo_concurrency.lockutils [None req-f132986c-33f7-4c26-b5b4-9bc035f0576b tempest-ServerGroupTestJSON-946002276 tempest-ServerGroupTestJSON-946002276-project-member] Acquiring lock "3439f60d-7d80-4d02-955c-dbefbdd04a32" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1168.163621] env[61923]: DEBUG oslo_concurrency.lockutils [None req-f132986c-33f7-4c26-b5b4-9bc035f0576b tempest-ServerGroupTestJSON-946002276 tempest-ServerGroupTestJSON-946002276-project-member] Lock "3439f60d-7d80-4d02-955c-dbefbdd04a32" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1168.666092] env[61923]: DEBUG nova.compute.manager [None req-f132986c-33f7-4c26-b5b4-9bc035f0576b tempest-ServerGroupTestJSON-946002276 tempest-ServerGroupTestJSON-946002276-project-member] [instance: 3439f60d-7d80-4d02-955c-dbefbdd04a32] Starting instance... {{(pid=61923) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1169.188628] env[61923]: DEBUG oslo_concurrency.lockutils [None req-f132986c-33f7-4c26-b5b4-9bc035f0576b tempest-ServerGroupTestJSON-946002276 tempest-ServerGroupTestJSON-946002276-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1169.188864] env[61923]: DEBUG oslo_concurrency.lockutils [None req-f132986c-33f7-4c26-b5b4-9bc035f0576b tempest-ServerGroupTestJSON-946002276 tempest-ServerGroupTestJSON-946002276-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1169.190180] env[61923]: INFO nova.compute.claims [None req-f132986c-33f7-4c26-b5b4-9bc035f0576b tempest-ServerGroupTestJSON-946002276 tempest-ServerGroupTestJSON-946002276-project-member] [instance: 3439f60d-7d80-4d02-955c-dbefbdd04a32] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1170.226904] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58a88985-532d-4ea7-811d-54997194bf48 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1170.233832] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad5960ab-c77d-4b92-86a6-e2fbe7d5d7cc {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1170.265648] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96701d00-e33d-4383-b94a-9a9acc2ff1d2 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1170.272994] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8acf3bd4-e352-4a4a-8a3f-826a9f64a9e2 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1170.286383] env[61923]: DEBUG nova.compute.provider_tree [None req-f132986c-33f7-4c26-b5b4-9bc035f0576b tempest-ServerGroupTestJSON-946002276 tempest-ServerGroupTestJSON-946002276-project-member] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1170.789650] env[61923]: DEBUG nova.scheduler.client.report [None req-f132986c-33f7-4c26-b5b4-9bc035f0576b tempest-ServerGroupTestJSON-946002276 tempest-ServerGroupTestJSON-946002276-project-member] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1171.295071] env[61923]: DEBUG oslo_concurrency.lockutils [None req-f132986c-33f7-4c26-b5b4-9bc035f0576b tempest-ServerGroupTestJSON-946002276 tempest-ServerGroupTestJSON-946002276-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.106s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1171.798541] env[61923]: DEBUG oslo_concurrency.lockutils [None req-f132986c-33f7-4c26-b5b4-9bc035f0576b tempest-ServerGroupTestJSON-946002276 tempest-ServerGroupTestJSON-946002276-project-member] Acquiring lock "afa2550b-2e3e-49e2-bde6-95ab600168b3" by "nova.compute.manager.ComputeManager._validate_instance_group_policy.._do_validation" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1171.798682] env[61923]: DEBUG oslo_concurrency.lockutils [None req-f132986c-33f7-4c26-b5b4-9bc035f0576b tempest-ServerGroupTestJSON-946002276 tempest-ServerGroupTestJSON-946002276-project-member] Lock "afa2550b-2e3e-49e2-bde6-95ab600168b3" acquired by "nova.compute.manager.ComputeManager._validate_instance_group_policy.._do_validation" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1172.300966] env[61923]: DEBUG oslo_concurrency.lockutils [None req-f132986c-33f7-4c26-b5b4-9bc035f0576b tempest-ServerGroupTestJSON-946002276 tempest-ServerGroupTestJSON-946002276-project-member] Lock "afa2550b-2e3e-49e2-bde6-95ab600168b3" "released" by "nova.compute.manager.ComputeManager._validate_instance_group_policy.._do_validation" :: held 0.502s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1172.301581] env[61923]: DEBUG nova.compute.manager [None req-f132986c-33f7-4c26-b5b4-9bc035f0576b tempest-ServerGroupTestJSON-946002276 tempest-ServerGroupTestJSON-946002276-project-member] [instance: 3439f60d-7d80-4d02-955c-dbefbdd04a32] Start building networks asynchronously for instance. {{(pid=61923) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1172.806869] env[61923]: DEBUG nova.compute.utils [None req-f132986c-33f7-4c26-b5b4-9bc035f0576b tempest-ServerGroupTestJSON-946002276 tempest-ServerGroupTestJSON-946002276-project-member] Using /dev/sd instead of None {{(pid=61923) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1172.808332] env[61923]: DEBUG nova.compute.manager [None req-f132986c-33f7-4c26-b5b4-9bc035f0576b tempest-ServerGroupTestJSON-946002276 tempest-ServerGroupTestJSON-946002276-project-member] [instance: 3439f60d-7d80-4d02-955c-dbefbdd04a32] Allocating IP information in the background. {{(pid=61923) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1172.808501] env[61923]: DEBUG nova.network.neutron [None req-f132986c-33f7-4c26-b5b4-9bc035f0576b tempest-ServerGroupTestJSON-946002276 tempest-ServerGroupTestJSON-946002276-project-member] [instance: 3439f60d-7d80-4d02-955c-dbefbdd04a32] allocate_for_instance() {{(pid=61923) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1172.854350] env[61923]: DEBUG nova.policy [None req-f132986c-33f7-4c26-b5b4-9bc035f0576b tempest-ServerGroupTestJSON-946002276 tempest-ServerGroupTestJSON-946002276-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e434bf595ec44bea8b8a55b72ef13cab', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '73f80f67dfbb4ce08171c612bb1ba85a', 'project_domain_id': 'default', 'roles': ['reader', 'member'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=61923) authorize /opt/stack/nova/nova/policy.py:201}} [ 1173.092060] env[61923]: DEBUG nova.network.neutron [None req-f132986c-33f7-4c26-b5b4-9bc035f0576b tempest-ServerGroupTestJSON-946002276 tempest-ServerGroupTestJSON-946002276-project-member] [instance: 3439f60d-7d80-4d02-955c-dbefbdd04a32] Successfully created port: 464e0808-f3a6-49bc-b3f3-e343a5371a14 {{(pid=61923) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1173.312096] env[61923]: DEBUG nova.compute.manager [None req-f132986c-33f7-4c26-b5b4-9bc035f0576b tempest-ServerGroupTestJSON-946002276 tempest-ServerGroupTestJSON-946002276-project-member] [instance: 3439f60d-7d80-4d02-955c-dbefbdd04a32] Start building block device mappings for instance. {{(pid=61923) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1174.321528] env[61923]: DEBUG nova.compute.manager [None req-f132986c-33f7-4c26-b5b4-9bc035f0576b tempest-ServerGroupTestJSON-946002276 tempest-ServerGroupTestJSON-946002276-project-member] [instance: 3439f60d-7d80-4d02-955c-dbefbdd04a32] Start spawning the instance on the hypervisor. {{(pid=61923) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1174.346191] env[61923]: DEBUG nova.virt.hardware [None req-f132986c-33f7-4c26-b5b4-9bc035f0576b tempest-ServerGroupTestJSON-946002276 tempest-ServerGroupTestJSON-946002276-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-29T20:07:52Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-29T20:07:35Z,direct_url=,disk_format='vmdk',id=0f153f63-ae0a-45b1-b7f5-7f9b673c947f,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='4e7b5e3b3c3443c8bd5c70f8a15739f5',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-29T20:07:36Z,virtual_size=,visibility=), allow threads: False {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1174.346395] env[61923]: DEBUG nova.virt.hardware [None req-f132986c-33f7-4c26-b5b4-9bc035f0576b tempest-ServerGroupTestJSON-946002276 tempest-ServerGroupTestJSON-946002276-project-member] Flavor limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1174.346596] env[61923]: DEBUG nova.virt.hardware [None req-f132986c-33f7-4c26-b5b4-9bc035f0576b tempest-ServerGroupTestJSON-946002276 tempest-ServerGroupTestJSON-946002276-project-member] Image limits 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1174.346791] env[61923]: DEBUG nova.virt.hardware [None req-f132986c-33f7-4c26-b5b4-9bc035f0576b tempest-ServerGroupTestJSON-946002276 tempest-ServerGroupTestJSON-946002276-project-member] Flavor pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1174.346940] env[61923]: DEBUG nova.virt.hardware [None req-f132986c-33f7-4c26-b5b4-9bc035f0576b tempest-ServerGroupTestJSON-946002276 tempest-ServerGroupTestJSON-946002276-project-member] Image pref 0:0:0 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1174.347102] env[61923]: DEBUG nova.virt.hardware [None req-f132986c-33f7-4c26-b5b4-9bc035f0576b tempest-ServerGroupTestJSON-946002276 tempest-ServerGroupTestJSON-946002276-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=61923) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1174.347642] env[61923]: DEBUG nova.virt.hardware [None req-f132986c-33f7-4c26-b5b4-9bc035f0576b tempest-ServerGroupTestJSON-946002276 tempest-ServerGroupTestJSON-946002276-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1174.347642] env[61923]: DEBUG nova.virt.hardware [None req-f132986c-33f7-4c26-b5b4-9bc035f0576b tempest-ServerGroupTestJSON-946002276 tempest-ServerGroupTestJSON-946002276-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1174.347642] env[61923]: DEBUG nova.virt.hardware [None req-f132986c-33f7-4c26-b5b4-9bc035f0576b tempest-ServerGroupTestJSON-946002276 tempest-ServerGroupTestJSON-946002276-project-member] Got 1 possible topologies {{(pid=61923) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1174.347797] env[61923]: DEBUG nova.virt.hardware [None req-f132986c-33f7-4c26-b5b4-9bc035f0576b tempest-ServerGroupTestJSON-946002276 tempest-ServerGroupTestJSON-946002276-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1174.347966] env[61923]: DEBUG nova.virt.hardware [None req-f132986c-33f7-4c26-b5b4-9bc035f0576b tempest-ServerGroupTestJSON-946002276 tempest-ServerGroupTestJSON-946002276-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=61923) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1174.349895] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-184404fb-f919-4d62-bce6-c53cb550137c {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1174.358786] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60c5809b-749a-4a75-bb41-9c07e8f95bd1 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1174.454531] env[61923]: DEBUG nova.compute.manager [req-6c4d28c2-051b-4a5c-9522-ee6b69d85e36 req-43d66c65-f5f4-483a-ba7b-67dd71b73c39 service nova] [instance: 3439f60d-7d80-4d02-955c-dbefbdd04a32] Received event network-vif-plugged-464e0808-f3a6-49bc-b3f3-e343a5371a14 {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1174.454771] env[61923]: DEBUG oslo_concurrency.lockutils [req-6c4d28c2-051b-4a5c-9522-ee6b69d85e36 req-43d66c65-f5f4-483a-ba7b-67dd71b73c39 service nova] Acquiring lock "3439f60d-7d80-4d02-955c-dbefbdd04a32-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1174.454985] env[61923]: DEBUG oslo_concurrency.lockutils [req-6c4d28c2-051b-4a5c-9522-ee6b69d85e36 req-43d66c65-f5f4-483a-ba7b-67dd71b73c39 service nova] Lock "3439f60d-7d80-4d02-955c-dbefbdd04a32-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1174.455181] env[61923]: DEBUG oslo_concurrency.lockutils [req-6c4d28c2-051b-4a5c-9522-ee6b69d85e36 req-43d66c65-f5f4-483a-ba7b-67dd71b73c39 service nova] Lock "3439f60d-7d80-4d02-955c-dbefbdd04a32-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1174.455353] env[61923]: DEBUG nova.compute.manager [req-6c4d28c2-051b-4a5c-9522-ee6b69d85e36 req-43d66c65-f5f4-483a-ba7b-67dd71b73c39 service nova] [instance: 3439f60d-7d80-4d02-955c-dbefbdd04a32] No waiting events found dispatching network-vif-plugged-464e0808-f3a6-49bc-b3f3-e343a5371a14 {{(pid=61923) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1174.455521] env[61923]: WARNING nova.compute.manager [req-6c4d28c2-051b-4a5c-9522-ee6b69d85e36 req-43d66c65-f5f4-483a-ba7b-67dd71b73c39 service nova] [instance: 3439f60d-7d80-4d02-955c-dbefbdd04a32] Received unexpected event network-vif-plugged-464e0808-f3a6-49bc-b3f3-e343a5371a14 for instance with vm_state building and task_state spawning. [ 1175.008176] env[61923]: DEBUG nova.network.neutron [None req-f132986c-33f7-4c26-b5b4-9bc035f0576b tempest-ServerGroupTestJSON-946002276 tempest-ServerGroupTestJSON-946002276-project-member] [instance: 3439f60d-7d80-4d02-955c-dbefbdd04a32] Successfully updated port: 464e0808-f3a6-49bc-b3f3-e343a5371a14 {{(pid=61923) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1175.031285] env[61923]: DEBUG nova.compute.manager [req-2ec5fc48-255d-498a-9207-e772b66c289e req-69eaca78-354d-4c8e-a6f0-e45761def726 service nova] [instance: 3439f60d-7d80-4d02-955c-dbefbdd04a32] Received event network-changed-464e0808-f3a6-49bc-b3f3-e343a5371a14 {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1175.031504] env[61923]: DEBUG nova.compute.manager [req-2ec5fc48-255d-498a-9207-e772b66c289e req-69eaca78-354d-4c8e-a6f0-e45761def726 service nova] [instance: 3439f60d-7d80-4d02-955c-dbefbdd04a32] Refreshing instance network info cache due to event network-changed-464e0808-f3a6-49bc-b3f3-e343a5371a14. {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1175.031708] env[61923]: DEBUG oslo_concurrency.lockutils [req-2ec5fc48-255d-498a-9207-e772b66c289e req-69eaca78-354d-4c8e-a6f0-e45761def726 service nova] Acquiring lock "refresh_cache-3439f60d-7d80-4d02-955c-dbefbdd04a32" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1175.031807] env[61923]: DEBUG oslo_concurrency.lockutils [req-2ec5fc48-255d-498a-9207-e772b66c289e req-69eaca78-354d-4c8e-a6f0-e45761def726 service nova] Acquired lock "refresh_cache-3439f60d-7d80-4d02-955c-dbefbdd04a32" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1175.031977] env[61923]: DEBUG nova.network.neutron [req-2ec5fc48-255d-498a-9207-e772b66c289e req-69eaca78-354d-4c8e-a6f0-e45761def726 service nova] [instance: 3439f60d-7d80-4d02-955c-dbefbdd04a32] Refreshing network info cache for port 464e0808-f3a6-49bc-b3f3-e343a5371a14 {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1175.510080] env[61923]: DEBUG oslo_concurrency.lockutils [None req-f132986c-33f7-4c26-b5b4-9bc035f0576b tempest-ServerGroupTestJSON-946002276 tempest-ServerGroupTestJSON-946002276-project-member] Acquiring lock "refresh_cache-3439f60d-7d80-4d02-955c-dbefbdd04a32" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1175.564199] env[61923]: DEBUG nova.network.neutron [req-2ec5fc48-255d-498a-9207-e772b66c289e req-69eaca78-354d-4c8e-a6f0-e45761def726 service nova] [instance: 3439f60d-7d80-4d02-955c-dbefbdd04a32] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1175.630518] env[61923]: DEBUG nova.network.neutron [req-2ec5fc48-255d-498a-9207-e772b66c289e req-69eaca78-354d-4c8e-a6f0-e45761def726 service nova] [instance: 3439f60d-7d80-4d02-955c-dbefbdd04a32] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1176.133224] env[61923]: DEBUG oslo_concurrency.lockutils [req-2ec5fc48-255d-498a-9207-e772b66c289e req-69eaca78-354d-4c8e-a6f0-e45761def726 service nova] Releasing lock "refresh_cache-3439f60d-7d80-4d02-955c-dbefbdd04a32" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1176.133629] env[61923]: DEBUG oslo_concurrency.lockutils [None req-f132986c-33f7-4c26-b5b4-9bc035f0576b tempest-ServerGroupTestJSON-946002276 tempest-ServerGroupTestJSON-946002276-project-member] Acquired lock "refresh_cache-3439f60d-7d80-4d02-955c-dbefbdd04a32" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1176.133782] env[61923]: DEBUG nova.network.neutron [None req-f132986c-33f7-4c26-b5b4-9bc035f0576b tempest-ServerGroupTestJSON-946002276 tempest-ServerGroupTestJSON-946002276-project-member] [instance: 3439f60d-7d80-4d02-955c-dbefbdd04a32] Building network info cache for instance {{(pid=61923) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1176.667091] env[61923]: DEBUG nova.network.neutron [None req-f132986c-33f7-4c26-b5b4-9bc035f0576b tempest-ServerGroupTestJSON-946002276 tempest-ServerGroupTestJSON-946002276-project-member] [instance: 3439f60d-7d80-4d02-955c-dbefbdd04a32] Instance cache missing network info. {{(pid=61923) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1176.787499] env[61923]: DEBUG nova.network.neutron [None req-f132986c-33f7-4c26-b5b4-9bc035f0576b tempest-ServerGroupTestJSON-946002276 tempest-ServerGroupTestJSON-946002276-project-member] [instance: 3439f60d-7d80-4d02-955c-dbefbdd04a32] Updating instance_info_cache with network_info: [{"id": "464e0808-f3a6-49bc-b3f3-e343a5371a14", "address": "fa:16:3e:87:f2:78", "network": {"id": "df383700-d37d-4f36-be22-20b5bf2a0e68", "bridge": "br-int", "label": "tempest-ServerGroupTestJSON-1053440393-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "73f80f67dfbb4ce08171c612bb1ba85a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "73f6629b-7f80-4a5b-8f15-c7a1635b3c33", "external-id": "nsx-vlan-transportzone-481", "segmentation_id": 481, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap464e0808-f3", "ovs_interfaceid": "464e0808-f3a6-49bc-b3f3-e343a5371a14", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1177.290270] env[61923]: DEBUG oslo_concurrency.lockutils [None req-f132986c-33f7-4c26-b5b4-9bc035f0576b tempest-ServerGroupTestJSON-946002276 tempest-ServerGroupTestJSON-946002276-project-member] Releasing lock "refresh_cache-3439f60d-7d80-4d02-955c-dbefbdd04a32" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1177.290629] env[61923]: DEBUG nova.compute.manager [None req-f132986c-33f7-4c26-b5b4-9bc035f0576b tempest-ServerGroupTestJSON-946002276 tempest-ServerGroupTestJSON-946002276-project-member] [instance: 3439f60d-7d80-4d02-955c-dbefbdd04a32] Instance network_info: |[{"id": "464e0808-f3a6-49bc-b3f3-e343a5371a14", "address": "fa:16:3e:87:f2:78", "network": {"id": "df383700-d37d-4f36-be22-20b5bf2a0e68", "bridge": "br-int", "label": "tempest-ServerGroupTestJSON-1053440393-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "73f80f67dfbb4ce08171c612bb1ba85a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "73f6629b-7f80-4a5b-8f15-c7a1635b3c33", "external-id": "nsx-vlan-transportzone-481", "segmentation_id": 481, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap464e0808-f3", "ovs_interfaceid": "464e0808-f3a6-49bc-b3f3-e343a5371a14", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=61923) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1177.291085] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-f132986c-33f7-4c26-b5b4-9bc035f0576b tempest-ServerGroupTestJSON-946002276 tempest-ServerGroupTestJSON-946002276-project-member] [instance: 3439f60d-7d80-4d02-955c-dbefbdd04a32] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:87:f2:78', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '73f6629b-7f80-4a5b-8f15-c7a1635b3c33', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '464e0808-f3a6-49bc-b3f3-e343a5371a14', 'vif_model': 'vmxnet3'}] {{(pid=61923) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1177.298358] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-f132986c-33f7-4c26-b5b4-9bc035f0576b tempest-ServerGroupTestJSON-946002276 tempest-ServerGroupTestJSON-946002276-project-member] Creating folder: Project (73f80f67dfbb4ce08171c612bb1ba85a). Parent ref: group-v292629. {{(pid=61923) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1177.298630] env[61923]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-aca884d4-2101-4276-831c-54f519ac77d5 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1177.310358] env[61923]: INFO nova.virt.vmwareapi.vm_util [None req-f132986c-33f7-4c26-b5b4-9bc035f0576b tempest-ServerGroupTestJSON-946002276 tempest-ServerGroupTestJSON-946002276-project-member] Created folder: Project (73f80f67dfbb4ce08171c612bb1ba85a) in parent group-v292629. [ 1177.310577] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-f132986c-33f7-4c26-b5b4-9bc035f0576b tempest-ServerGroupTestJSON-946002276 tempest-ServerGroupTestJSON-946002276-project-member] Creating folder: Instances. Parent ref: group-v292761. {{(pid=61923) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1177.311175] env[61923]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-89ef56aa-881f-44c6-a925-58050c7c0150 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1177.319438] env[61923]: INFO nova.virt.vmwareapi.vm_util [None req-f132986c-33f7-4c26-b5b4-9bc035f0576b tempest-ServerGroupTestJSON-946002276 tempest-ServerGroupTestJSON-946002276-project-member] Created folder: Instances in parent group-v292761. [ 1177.319660] env[61923]: DEBUG oslo.service.loopingcall [None req-f132986c-33f7-4c26-b5b4-9bc035f0576b tempest-ServerGroupTestJSON-946002276 tempest-ServerGroupTestJSON-946002276-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=61923) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1177.319836] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3439f60d-7d80-4d02-955c-dbefbdd04a32] Creating VM on the ESX host {{(pid=61923) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1177.320036] env[61923]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-9b96d49a-12c5-439b-af1d-dcfe9a8faeeb {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1177.338496] env[61923]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1177.338496] env[61923]: value = "task-1378087" [ 1177.338496] env[61923]: _type = "Task" [ 1177.338496] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1177.345532] env[61923]: DEBUG oslo_vmware.api [-] Task: {'id': task-1378087, 'name': CreateVM_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1177.848464] env[61923]: DEBUG oslo_vmware.api [-] Task: {'id': task-1378087, 'name': CreateVM_Task} progress is 25%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1178.349087] env[61923]: DEBUG oslo_vmware.api [-] Task: {'id': task-1378087, 'name': CreateVM_Task, 'duration_secs': 0.729693} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1178.349276] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3439f60d-7d80-4d02-955c-dbefbdd04a32] Created VM on the ESX host {{(pid=61923) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1178.349943] env[61923]: DEBUG oslo_concurrency.lockutils [None req-f132986c-33f7-4c26-b5b4-9bc035f0576b tempest-ServerGroupTestJSON-946002276 tempest-ServerGroupTestJSON-946002276-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1178.350132] env[61923]: DEBUG oslo_concurrency.lockutils [None req-f132986c-33f7-4c26-b5b4-9bc035f0576b tempest-ServerGroupTestJSON-946002276 tempest-ServerGroupTestJSON-946002276-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1178.350515] env[61923]: DEBUG oslo_concurrency.lockutils [None req-f132986c-33f7-4c26-b5b4-9bc035f0576b tempest-ServerGroupTestJSON-946002276 tempest-ServerGroupTestJSON-946002276-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1178.350776] env[61923]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5be0893d-9d53-46e3-a637-ad647dae302e {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1178.354939] env[61923]: DEBUG oslo_vmware.api [None req-f132986c-33f7-4c26-b5b4-9bc035f0576b tempest-ServerGroupTestJSON-946002276 tempest-ServerGroupTestJSON-946002276-project-member] Waiting for the task: (returnval){ [ 1178.354939] env[61923]: value = "session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52fff530-99f7-5eae-6f16-c8df95b217e0" [ 1178.354939] env[61923]: _type = "Task" [ 1178.354939] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1178.361858] env[61923]: DEBUG oslo_vmware.api [None req-f132986c-33f7-4c26-b5b4-9bc035f0576b tempest-ServerGroupTestJSON-946002276 tempest-ServerGroupTestJSON-946002276-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52fff530-99f7-5eae-6f16-c8df95b217e0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1178.865562] env[61923]: DEBUG oslo_vmware.api [None req-f132986c-33f7-4c26-b5b4-9bc035f0576b tempest-ServerGroupTestJSON-946002276 tempest-ServerGroupTestJSON-946002276-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52fff530-99f7-5eae-6f16-c8df95b217e0, 'name': SearchDatastore_Task, 'duration_secs': 0.010404} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1178.865921] env[61923]: DEBUG oslo_concurrency.lockutils [None req-f132986c-33f7-4c26-b5b4-9bc035f0576b tempest-ServerGroupTestJSON-946002276 tempest-ServerGroupTestJSON-946002276-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1178.866099] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-f132986c-33f7-4c26-b5b4-9bc035f0576b tempest-ServerGroupTestJSON-946002276 tempest-ServerGroupTestJSON-946002276-project-member] [instance: 3439f60d-7d80-4d02-955c-dbefbdd04a32] Processing image 0f153f63-ae0a-45b1-b7f5-7f9b673c947f {{(pid=61923) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1178.866344] env[61923]: DEBUG oslo_concurrency.lockutils [None req-f132986c-33f7-4c26-b5b4-9bc035f0576b tempest-ServerGroupTestJSON-946002276 tempest-ServerGroupTestJSON-946002276-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1178.866489] env[61923]: DEBUG oslo_concurrency.lockutils [None req-f132986c-33f7-4c26-b5b4-9bc035f0576b tempest-ServerGroupTestJSON-946002276 tempest-ServerGroupTestJSON-946002276-project-member] Acquired lock "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1178.866672] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-f132986c-33f7-4c26-b5b4-9bc035f0576b tempest-ServerGroupTestJSON-946002276 tempest-ServerGroupTestJSON-946002276-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=61923) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1178.866935] env[61923]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-570aa04b-95e9-46be-9eb3-0268ddfb8de8 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1178.874843] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-f132986c-33f7-4c26-b5b4-9bc035f0576b tempest-ServerGroupTestJSON-946002276 tempest-ServerGroupTestJSON-946002276-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=61923) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1178.875032] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-f132986c-33f7-4c26-b5b4-9bc035f0576b tempest-ServerGroupTestJSON-946002276 tempest-ServerGroupTestJSON-946002276-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=61923) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1178.875692] env[61923]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-86dbe428-847c-4e52-aa2a-d5aa23abc697 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1178.880187] env[61923]: DEBUG oslo_vmware.api [None req-f132986c-33f7-4c26-b5b4-9bc035f0576b tempest-ServerGroupTestJSON-946002276 tempest-ServerGroupTestJSON-946002276-project-member] Waiting for the task: (returnval){ [ 1178.880187] env[61923]: value = "session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52ecbfc8-a536-d516-44b7-513616b6b58a" [ 1178.880187] env[61923]: _type = "Task" [ 1178.880187] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1178.887207] env[61923]: DEBUG oslo_vmware.api [None req-f132986c-33f7-4c26-b5b4-9bc035f0576b tempest-ServerGroupTestJSON-946002276 tempest-ServerGroupTestJSON-946002276-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52ecbfc8-a536-d516-44b7-513616b6b58a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1179.395452] env[61923]: DEBUG oslo_vmware.api [None req-f132986c-33f7-4c26-b5b4-9bc035f0576b tempest-ServerGroupTestJSON-946002276 tempest-ServerGroupTestJSON-946002276-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]52ecbfc8-a536-d516-44b7-513616b6b58a, 'name': SearchDatastore_Task, 'duration_secs': 0.007567} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1179.396722] env[61923]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-45661c1c-a8d6-4797-b10a-61a094f109b0 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1179.403392] env[61923]: DEBUG oslo_vmware.api [None req-f132986c-33f7-4c26-b5b4-9bc035f0576b tempest-ServerGroupTestJSON-946002276 tempest-ServerGroupTestJSON-946002276-project-member] Waiting for the task: (returnval){ [ 1179.403392] env[61923]: value = "session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]5238af5a-8ed7-305a-3632-ce4e9c10be1b" [ 1179.403392] env[61923]: _type = "Task" [ 1179.403392] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1179.414333] env[61923]: DEBUG oslo_vmware.api [None req-f132986c-33f7-4c26-b5b4-9bc035f0576b tempest-ServerGroupTestJSON-946002276 tempest-ServerGroupTestJSON-946002276-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]5238af5a-8ed7-305a-3632-ce4e9c10be1b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1179.913466] env[61923]: DEBUG oslo_vmware.api [None req-f132986c-33f7-4c26-b5b4-9bc035f0576b tempest-ServerGroupTestJSON-946002276 tempest-ServerGroupTestJSON-946002276-project-member] Task: {'id': session[52cfc8d9-5655-6e0a-dab6-913fb74e9e85]5238af5a-8ed7-305a-3632-ce4e9c10be1b, 'name': SearchDatastore_Task, 'duration_secs': 0.008522} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1179.913788] env[61923]: DEBUG oslo_concurrency.lockutils [None req-f132986c-33f7-4c26-b5b4-9bc035f0576b tempest-ServerGroupTestJSON-946002276 tempest-ServerGroupTestJSON-946002276-project-member] Releasing lock "[datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk" {{(pid=61923) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1179.914013] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-f132986c-33f7-4c26-b5b4-9bc035f0576b tempest-ServerGroupTestJSON-946002276 tempest-ServerGroupTestJSON-946002276-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk to [datastore1] 3439f60d-7d80-4d02-955c-dbefbdd04a32/3439f60d-7d80-4d02-955c-dbefbdd04a32.vmdk {{(pid=61923) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1179.914276] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-679e6c8e-6a6a-41fa-a846-eee279151335 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1179.920876] env[61923]: DEBUG oslo_vmware.api [None req-f132986c-33f7-4c26-b5b4-9bc035f0576b tempest-ServerGroupTestJSON-946002276 tempest-ServerGroupTestJSON-946002276-project-member] Waiting for the task: (returnval){ [ 1179.920876] env[61923]: value = "task-1378088" [ 1179.920876] env[61923]: _type = "Task" [ 1179.920876] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1179.927720] env[61923]: DEBUG oslo_vmware.api [None req-f132986c-33f7-4c26-b5b4-9bc035f0576b tempest-ServerGroupTestJSON-946002276 tempest-ServerGroupTestJSON-946002276-project-member] Task: {'id': task-1378088, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1180.430374] env[61923]: DEBUG oslo_vmware.api [None req-f132986c-33f7-4c26-b5b4-9bc035f0576b tempest-ServerGroupTestJSON-946002276 tempest-ServerGroupTestJSON-946002276-project-member] Task: {'id': task-1378088, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.435589} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1180.430662] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-f132986c-33f7-4c26-b5b4-9bc035f0576b tempest-ServerGroupTestJSON-946002276 tempest-ServerGroupTestJSON-946002276-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/0f153f63-ae0a-45b1-b7f5-7f9b673c947f/0f153f63-ae0a-45b1-b7f5-7f9b673c947f.vmdk to [datastore1] 3439f60d-7d80-4d02-955c-dbefbdd04a32/3439f60d-7d80-4d02-955c-dbefbdd04a32.vmdk {{(pid=61923) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1180.430886] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-f132986c-33f7-4c26-b5b4-9bc035f0576b tempest-ServerGroupTestJSON-946002276 tempest-ServerGroupTestJSON-946002276-project-member] [instance: 3439f60d-7d80-4d02-955c-dbefbdd04a32] Extending root virtual disk to 1048576 {{(pid=61923) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1180.431158] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-1028afc9-c6fc-4401-bfba-b24e5d6669c2 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1180.437924] env[61923]: DEBUG oslo_vmware.api [None req-f132986c-33f7-4c26-b5b4-9bc035f0576b tempest-ServerGroupTestJSON-946002276 tempest-ServerGroupTestJSON-946002276-project-member] Waiting for the task: (returnval){ [ 1180.437924] env[61923]: value = "task-1378089" [ 1180.437924] env[61923]: _type = "Task" [ 1180.437924] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1180.445263] env[61923]: DEBUG oslo_vmware.api [None req-f132986c-33f7-4c26-b5b4-9bc035f0576b tempest-ServerGroupTestJSON-946002276 tempest-ServerGroupTestJSON-946002276-project-member] Task: {'id': task-1378089, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1180.947307] env[61923]: DEBUG oslo_vmware.api [None req-f132986c-33f7-4c26-b5b4-9bc035f0576b tempest-ServerGroupTestJSON-946002276 tempest-ServerGroupTestJSON-946002276-project-member] Task: {'id': task-1378089, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.071064} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1180.947675] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-f132986c-33f7-4c26-b5b4-9bc035f0576b tempest-ServerGroupTestJSON-946002276 tempest-ServerGroupTestJSON-946002276-project-member] [instance: 3439f60d-7d80-4d02-955c-dbefbdd04a32] Extended root virtual disk {{(pid=61923) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1180.948308] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b70b3029-aa3e-4de6-8a8c-190623858441 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1180.969074] env[61923]: DEBUG nova.virt.vmwareapi.volumeops [None req-f132986c-33f7-4c26-b5b4-9bc035f0576b tempest-ServerGroupTestJSON-946002276 tempest-ServerGroupTestJSON-946002276-project-member] [instance: 3439f60d-7d80-4d02-955c-dbefbdd04a32] Reconfiguring VM instance instance-00000069 to attach disk [datastore1] 3439f60d-7d80-4d02-955c-dbefbdd04a32/3439f60d-7d80-4d02-955c-dbefbdd04a32.vmdk or device None with type sparse {{(pid=61923) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1180.969301] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-82992794-94c6-4d4c-a9af-8511ff008f3d {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1180.988386] env[61923]: DEBUG oslo_vmware.api [None req-f132986c-33f7-4c26-b5b4-9bc035f0576b tempest-ServerGroupTestJSON-946002276 tempest-ServerGroupTestJSON-946002276-project-member] Waiting for the task: (returnval){ [ 1180.988386] env[61923]: value = "task-1378090" [ 1180.988386] env[61923]: _type = "Task" [ 1180.988386] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1180.995500] env[61923]: DEBUG oslo_vmware.api [None req-f132986c-33f7-4c26-b5b4-9bc035f0576b tempest-ServerGroupTestJSON-946002276 tempest-ServerGroupTestJSON-946002276-project-member] Task: {'id': task-1378090, 'name': ReconfigVM_Task} progress is 5%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1181.498175] env[61923]: DEBUG oslo_vmware.api [None req-f132986c-33f7-4c26-b5b4-9bc035f0576b tempest-ServerGroupTestJSON-946002276 tempest-ServerGroupTestJSON-946002276-project-member] Task: {'id': task-1378090, 'name': ReconfigVM_Task, 'duration_secs': 0.29722} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1181.498462] env[61923]: DEBUG nova.virt.vmwareapi.volumeops [None req-f132986c-33f7-4c26-b5b4-9bc035f0576b tempest-ServerGroupTestJSON-946002276 tempest-ServerGroupTestJSON-946002276-project-member] [instance: 3439f60d-7d80-4d02-955c-dbefbdd04a32] Reconfigured VM instance instance-00000069 to attach disk [datastore1] 3439f60d-7d80-4d02-955c-dbefbdd04a32/3439f60d-7d80-4d02-955c-dbefbdd04a32.vmdk or device None with type sparse {{(pid=61923) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1181.499086] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-6bd0c828-41aa-4b90-9715-6fab42641b3e {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1181.505195] env[61923]: DEBUG oslo_vmware.api [None req-f132986c-33f7-4c26-b5b4-9bc035f0576b tempest-ServerGroupTestJSON-946002276 tempest-ServerGroupTestJSON-946002276-project-member] Waiting for the task: (returnval){ [ 1181.505195] env[61923]: value = "task-1378091" [ 1181.505195] env[61923]: _type = "Task" [ 1181.505195] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1181.512264] env[61923]: DEBUG oslo_vmware.api [None req-f132986c-33f7-4c26-b5b4-9bc035f0576b tempest-ServerGroupTestJSON-946002276 tempest-ServerGroupTestJSON-946002276-project-member] Task: {'id': task-1378091, 'name': Rename_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1182.014623] env[61923]: DEBUG oslo_vmware.api [None req-f132986c-33f7-4c26-b5b4-9bc035f0576b tempest-ServerGroupTestJSON-946002276 tempest-ServerGroupTestJSON-946002276-project-member] Task: {'id': task-1378091, 'name': Rename_Task, 'duration_secs': 0.131013} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1182.015037] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-f132986c-33f7-4c26-b5b4-9bc035f0576b tempest-ServerGroupTestJSON-946002276 tempest-ServerGroupTestJSON-946002276-project-member] [instance: 3439f60d-7d80-4d02-955c-dbefbdd04a32] Powering on the VM {{(pid=61923) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1182.015189] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-04a1202f-176f-4657-ad2f-f2757168bea0 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1182.021851] env[61923]: DEBUG oslo_vmware.api [None req-f132986c-33f7-4c26-b5b4-9bc035f0576b tempest-ServerGroupTestJSON-946002276 tempest-ServerGroupTestJSON-946002276-project-member] Waiting for the task: (returnval){ [ 1182.021851] env[61923]: value = "task-1378092" [ 1182.021851] env[61923]: _type = "Task" [ 1182.021851] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1182.029709] env[61923]: DEBUG oslo_vmware.api [None req-f132986c-33f7-4c26-b5b4-9bc035f0576b tempest-ServerGroupTestJSON-946002276 tempest-ServerGroupTestJSON-946002276-project-member] Task: {'id': task-1378092, 'name': PowerOnVM_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1182.531837] env[61923]: DEBUG oslo_vmware.api [None req-f132986c-33f7-4c26-b5b4-9bc035f0576b tempest-ServerGroupTestJSON-946002276 tempest-ServerGroupTestJSON-946002276-project-member] Task: {'id': task-1378092, 'name': PowerOnVM_Task, 'duration_secs': 0.425992} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1182.532127] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-f132986c-33f7-4c26-b5b4-9bc035f0576b tempest-ServerGroupTestJSON-946002276 tempest-ServerGroupTestJSON-946002276-project-member] [instance: 3439f60d-7d80-4d02-955c-dbefbdd04a32] Powered on the VM {{(pid=61923) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1182.532339] env[61923]: INFO nova.compute.manager [None req-f132986c-33f7-4c26-b5b4-9bc035f0576b tempest-ServerGroupTestJSON-946002276 tempest-ServerGroupTestJSON-946002276-project-member] [instance: 3439f60d-7d80-4d02-955c-dbefbdd04a32] Took 8.21 seconds to spawn the instance on the hypervisor. [ 1182.532556] env[61923]: DEBUG nova.compute.manager [None req-f132986c-33f7-4c26-b5b4-9bc035f0576b tempest-ServerGroupTestJSON-946002276 tempest-ServerGroupTestJSON-946002276-project-member] [instance: 3439f60d-7d80-4d02-955c-dbefbdd04a32] Checking state {{(pid=61923) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1182.533326] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f31d08e-0492-4d77-80d3-a7fb4bf13b26 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1183.049883] env[61923]: INFO nova.compute.manager [None req-f132986c-33f7-4c26-b5b4-9bc035f0576b tempest-ServerGroupTestJSON-946002276 tempest-ServerGroupTestJSON-946002276-project-member] [instance: 3439f60d-7d80-4d02-955c-dbefbdd04a32] Took 13.88 seconds to build instance. [ 1183.371455] env[61923]: DEBUG oslo_concurrency.lockutils [None req-02132831-acd0-4b17-a64d-d97539211890 tempest-ServerGroupTestJSON-946002276 tempest-ServerGroupTestJSON-946002276-project-member] Acquiring lock "3439f60d-7d80-4d02-955c-dbefbdd04a32" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1183.552064] env[61923]: DEBUG oslo_concurrency.lockutils [None req-f132986c-33f7-4c26-b5b4-9bc035f0576b tempest-ServerGroupTestJSON-946002276 tempest-ServerGroupTestJSON-946002276-project-member] Lock "3439f60d-7d80-4d02-955c-dbefbdd04a32" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.388s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1183.552297] env[61923]: DEBUG oslo_concurrency.lockutils [None req-02132831-acd0-4b17-a64d-d97539211890 tempest-ServerGroupTestJSON-946002276 tempest-ServerGroupTestJSON-946002276-project-member] Lock "3439f60d-7d80-4d02-955c-dbefbdd04a32" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.181s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1183.552598] env[61923]: DEBUG oslo_concurrency.lockutils [None req-02132831-acd0-4b17-a64d-d97539211890 tempest-ServerGroupTestJSON-946002276 tempest-ServerGroupTestJSON-946002276-project-member] Acquiring lock "3439f60d-7d80-4d02-955c-dbefbdd04a32-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1183.552809] env[61923]: DEBUG oslo_concurrency.lockutils [None req-02132831-acd0-4b17-a64d-d97539211890 tempest-ServerGroupTestJSON-946002276 tempest-ServerGroupTestJSON-946002276-project-member] Lock "3439f60d-7d80-4d02-955c-dbefbdd04a32-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1183.552984] env[61923]: DEBUG oslo_concurrency.lockutils [None req-02132831-acd0-4b17-a64d-d97539211890 tempest-ServerGroupTestJSON-946002276 tempest-ServerGroupTestJSON-946002276-project-member] Lock "3439f60d-7d80-4d02-955c-dbefbdd04a32-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1183.555209] env[61923]: INFO nova.compute.manager [None req-02132831-acd0-4b17-a64d-d97539211890 tempest-ServerGroupTestJSON-946002276 tempest-ServerGroupTestJSON-946002276-project-member] [instance: 3439f60d-7d80-4d02-955c-dbefbdd04a32] Terminating instance [ 1183.556961] env[61923]: DEBUG nova.compute.manager [None req-02132831-acd0-4b17-a64d-d97539211890 tempest-ServerGroupTestJSON-946002276 tempest-ServerGroupTestJSON-946002276-project-member] [instance: 3439f60d-7d80-4d02-955c-dbefbdd04a32] Start destroying the instance on the hypervisor. {{(pid=61923) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1183.557179] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-02132831-acd0-4b17-a64d-d97539211890 tempest-ServerGroupTestJSON-946002276 tempest-ServerGroupTestJSON-946002276-project-member] [instance: 3439f60d-7d80-4d02-955c-dbefbdd04a32] Destroying instance {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1183.558013] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67434ead-cf1a-4de1-ad7b-2da671bfc590 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1183.565308] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-02132831-acd0-4b17-a64d-d97539211890 tempest-ServerGroupTestJSON-946002276 tempest-ServerGroupTestJSON-946002276-project-member] [instance: 3439f60d-7d80-4d02-955c-dbefbdd04a32] Powering off the VM {{(pid=61923) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1183.565533] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4efaebda-5608-4d29-87d2-a6958f8d893c {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1183.571485] env[61923]: DEBUG oslo_vmware.api [None req-02132831-acd0-4b17-a64d-d97539211890 tempest-ServerGroupTestJSON-946002276 tempest-ServerGroupTestJSON-946002276-project-member] Waiting for the task: (returnval){ [ 1183.571485] env[61923]: value = "task-1378093" [ 1183.571485] env[61923]: _type = "Task" [ 1183.571485] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1183.578746] env[61923]: DEBUG oslo_vmware.api [None req-02132831-acd0-4b17-a64d-d97539211890 tempest-ServerGroupTestJSON-946002276 tempest-ServerGroupTestJSON-946002276-project-member] Task: {'id': task-1378093, 'name': PowerOffVM_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1184.081184] env[61923]: DEBUG oslo_vmware.api [None req-02132831-acd0-4b17-a64d-d97539211890 tempest-ServerGroupTestJSON-946002276 tempest-ServerGroupTestJSON-946002276-project-member] Task: {'id': task-1378093, 'name': PowerOffVM_Task, 'duration_secs': 0.197772} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1184.081483] env[61923]: DEBUG nova.virt.vmwareapi.vm_util [None req-02132831-acd0-4b17-a64d-d97539211890 tempest-ServerGroupTestJSON-946002276 tempest-ServerGroupTestJSON-946002276-project-member] [instance: 3439f60d-7d80-4d02-955c-dbefbdd04a32] Powered off the VM {{(pid=61923) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1184.081633] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-02132831-acd0-4b17-a64d-d97539211890 tempest-ServerGroupTestJSON-946002276 tempest-ServerGroupTestJSON-946002276-project-member] [instance: 3439f60d-7d80-4d02-955c-dbefbdd04a32] Unregistering the VM {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1184.081884] env[61923]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-1a295a00-3d91-4c07-a75a-ebf1a90494ca {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1184.141624] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-02132831-acd0-4b17-a64d-d97539211890 tempest-ServerGroupTestJSON-946002276 tempest-ServerGroupTestJSON-946002276-project-member] [instance: 3439f60d-7d80-4d02-955c-dbefbdd04a32] Unregistered the VM {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1184.141847] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-02132831-acd0-4b17-a64d-d97539211890 tempest-ServerGroupTestJSON-946002276 tempest-ServerGroupTestJSON-946002276-project-member] [instance: 3439f60d-7d80-4d02-955c-dbefbdd04a32] Deleting contents of the VM from datastore datastore1 {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1184.142055] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-02132831-acd0-4b17-a64d-d97539211890 tempest-ServerGroupTestJSON-946002276 tempest-ServerGroupTestJSON-946002276-project-member] Deleting the datastore file [datastore1] 3439f60d-7d80-4d02-955c-dbefbdd04a32 {{(pid=61923) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1184.142349] env[61923]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-5da8e0d1-4362-4121-8805-3df7633f3270 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1184.149152] env[61923]: DEBUG oslo_vmware.api [None req-02132831-acd0-4b17-a64d-d97539211890 tempest-ServerGroupTestJSON-946002276 tempest-ServerGroupTestJSON-946002276-project-member] Waiting for the task: (returnval){ [ 1184.149152] env[61923]: value = "task-1378095" [ 1184.149152] env[61923]: _type = "Task" [ 1184.149152] env[61923]: } to complete. {{(pid=61923) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1184.156652] env[61923]: DEBUG oslo_vmware.api [None req-02132831-acd0-4b17-a64d-d97539211890 tempest-ServerGroupTestJSON-946002276 tempest-ServerGroupTestJSON-946002276-project-member] Task: {'id': task-1378095, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1184.659210] env[61923]: DEBUG oslo_vmware.api [None req-02132831-acd0-4b17-a64d-d97539211890 tempest-ServerGroupTestJSON-946002276 tempest-ServerGroupTestJSON-946002276-project-member] Task: {'id': task-1378095, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.174478} completed successfully. {{(pid=61923) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1184.659480] env[61923]: DEBUG nova.virt.vmwareapi.ds_util [None req-02132831-acd0-4b17-a64d-d97539211890 tempest-ServerGroupTestJSON-946002276 tempest-ServerGroupTestJSON-946002276-project-member] Deleted the datastore file {{(pid=61923) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1184.659677] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-02132831-acd0-4b17-a64d-d97539211890 tempest-ServerGroupTestJSON-946002276 tempest-ServerGroupTestJSON-946002276-project-member] [instance: 3439f60d-7d80-4d02-955c-dbefbdd04a32] Deleted contents of the VM from datastore datastore1 {{(pid=61923) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1184.659856] env[61923]: DEBUG nova.virt.vmwareapi.vmops [None req-02132831-acd0-4b17-a64d-d97539211890 tempest-ServerGroupTestJSON-946002276 tempest-ServerGroupTestJSON-946002276-project-member] [instance: 3439f60d-7d80-4d02-955c-dbefbdd04a32] Instance destroyed {{(pid=61923) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1184.660042] env[61923]: INFO nova.compute.manager [None req-02132831-acd0-4b17-a64d-d97539211890 tempest-ServerGroupTestJSON-946002276 tempest-ServerGroupTestJSON-946002276-project-member] [instance: 3439f60d-7d80-4d02-955c-dbefbdd04a32] Took 1.10 seconds to destroy the instance on the hypervisor. [ 1184.660310] env[61923]: DEBUG oslo.service.loopingcall [None req-02132831-acd0-4b17-a64d-d97539211890 tempest-ServerGroupTestJSON-946002276 tempest-ServerGroupTestJSON-946002276-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=61923) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1184.660528] env[61923]: DEBUG nova.compute.manager [-] [instance: 3439f60d-7d80-4d02-955c-dbefbdd04a32] Deallocating network for instance {{(pid=61923) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1184.660623] env[61923]: DEBUG nova.network.neutron [-] [instance: 3439f60d-7d80-4d02-955c-dbefbdd04a32] deallocate_for_instance() {{(pid=61923) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1184.916904] env[61923]: DEBUG nova.compute.manager [req-872fe5f2-4a71-4e39-9dc8-c3e99852aba4 req-0a80c6bf-b408-4f91-bf74-0304904e3c07 service nova] [instance: 3439f60d-7d80-4d02-955c-dbefbdd04a32] Received event network-vif-deleted-464e0808-f3a6-49bc-b3f3-e343a5371a14 {{(pid=61923) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1184.917121] env[61923]: INFO nova.compute.manager [req-872fe5f2-4a71-4e39-9dc8-c3e99852aba4 req-0a80c6bf-b408-4f91-bf74-0304904e3c07 service nova] [instance: 3439f60d-7d80-4d02-955c-dbefbdd04a32] Neutron deleted interface 464e0808-f3a6-49bc-b3f3-e343a5371a14; detaching it from the instance and deleting it from the info cache [ 1184.917360] env[61923]: DEBUG nova.network.neutron [req-872fe5f2-4a71-4e39-9dc8-c3e99852aba4 req-0a80c6bf-b408-4f91-bf74-0304904e3c07 service nova] [instance: 3439f60d-7d80-4d02-955c-dbefbdd04a32] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1185.394999] env[61923]: DEBUG nova.network.neutron [-] [instance: 3439f60d-7d80-4d02-955c-dbefbdd04a32] Updating instance_info_cache with network_info: [] {{(pid=61923) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1185.422519] env[61923]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-60e76525-df55-4eb4-a81c-40cd60b4bfbc {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1185.432694] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6f2ff1f-4785-4b06-b586-d782c53b91a9 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1185.455046] env[61923]: DEBUG nova.compute.manager [req-872fe5f2-4a71-4e39-9dc8-c3e99852aba4 req-0a80c6bf-b408-4f91-bf74-0304904e3c07 service nova] [instance: 3439f60d-7d80-4d02-955c-dbefbdd04a32] Detach interface failed, port_id=464e0808-f3a6-49bc-b3f3-e343a5371a14, reason: Instance 3439f60d-7d80-4d02-955c-dbefbdd04a32 could not be found. {{(pid=61923) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1185.897800] env[61923]: INFO nova.compute.manager [-] [instance: 3439f60d-7d80-4d02-955c-dbefbdd04a32] Took 1.24 seconds to deallocate network for instance. [ 1186.405486] env[61923]: DEBUG oslo_concurrency.lockutils [None req-02132831-acd0-4b17-a64d-d97539211890 tempest-ServerGroupTestJSON-946002276 tempest-ServerGroupTestJSON-946002276-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1186.405847] env[61923]: DEBUG oslo_concurrency.lockutils [None req-02132831-acd0-4b17-a64d-d97539211890 tempest-ServerGroupTestJSON-946002276 tempest-ServerGroupTestJSON-946002276-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1186.405915] env[61923]: DEBUG nova.objects.instance [None req-02132831-acd0-4b17-a64d-d97539211890 tempest-ServerGroupTestJSON-946002276 tempest-ServerGroupTestJSON-946002276-project-member] Lazy-loading 'resources' on Instance uuid 3439f60d-7d80-4d02-955c-dbefbdd04a32 {{(pid=61923) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1186.940705] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-337945f2-f29e-48d2-b35e-4d9db231b68d {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1186.947605] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-535a1f41-9e16-4c83-bae9-a5b38f82f3e6 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1186.976751] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c8f657b-4203-45c7-bf36-93a53c26cfe2 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1186.983235] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5977ec59-8170-44ba-afc5-48e49a6428ed {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1186.995531] env[61923]: DEBUG nova.compute.provider_tree [None req-02132831-acd0-4b17-a64d-d97539211890 tempest-ServerGroupTestJSON-946002276 tempest-ServerGroupTestJSON-946002276-project-member] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1187.498610] env[61923]: DEBUG nova.scheduler.client.report [None req-02132831-acd0-4b17-a64d-d97539211890 tempest-ServerGroupTestJSON-946002276 tempest-ServerGroupTestJSON-946002276-project-member] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1188.005715] env[61923]: DEBUG oslo_concurrency.lockutils [None req-02132831-acd0-4b17-a64d-d97539211890 tempest-ServerGroupTestJSON-946002276 tempest-ServerGroupTestJSON-946002276-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.600s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1188.033129] env[61923]: INFO nova.scheduler.client.report [None req-02132831-acd0-4b17-a64d-d97539211890 tempest-ServerGroupTestJSON-946002276 tempest-ServerGroupTestJSON-946002276-project-member] Deleted allocations for instance 3439f60d-7d80-4d02-955c-dbefbdd04a32 [ 1188.542319] env[61923]: DEBUG oslo_concurrency.lockutils [None req-02132831-acd0-4b17-a64d-d97539211890 tempest-ServerGroupTestJSON-946002276 tempest-ServerGroupTestJSON-946002276-project-member] Lock "3439f60d-7d80-4d02-955c-dbefbdd04a32" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 4.990s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1192.354943] env[61923]: DEBUG oslo_service.periodic_task [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=61923) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1192.355330] env[61923]: DEBUG nova.compute.manager [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Starting heal instance info cache {{(pid=61923) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 1192.355330] env[61923]: DEBUG nova.compute.manager [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Rebuilding the list of instances to heal {{(pid=61923) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 1192.858501] env[61923]: DEBUG nova.compute.manager [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Didn't find any instances for network info cache update. {{(pid=61923) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10014}} [ 1194.355496] env[61923]: DEBUG oslo_service.periodic_task [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=61923) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1194.355856] env[61923]: DEBUG oslo_service.periodic_task [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=61923) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1194.355914] env[61923]: DEBUG oslo_service.periodic_task [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=61923) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1195.355319] env[61923]: DEBUG oslo_service.periodic_task [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=61923) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1195.355567] env[61923]: DEBUG nova.compute.manager [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=61923) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 1196.355542] env[61923]: DEBUG oslo_service.periodic_task [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=61923) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1197.350627] env[61923]: DEBUG oslo_service.periodic_task [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=61923) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1197.355318] env[61923]: DEBUG oslo_service.periodic_task [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=61923) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1198.354889] env[61923]: DEBUG oslo_service.periodic_task [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Running periodic task ComputeManager.update_available_resource {{(pid=61923) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1198.858029] env[61923]: DEBUG oslo_concurrency.lockutils [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1198.858314] env[61923]: DEBUG oslo_concurrency.lockutils [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1198.858488] env[61923]: DEBUG oslo_concurrency.lockutils [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1198.858644] env[61923]: DEBUG nova.compute.resource_tracker [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=61923) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1198.859606] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9943d663-32ae-475c-9636-ea09581a7367 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1198.868215] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6cc0da06-9b53-4d5d-8fdf-8c83afa40fb6 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1198.882588] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a8a69a3-97ee-4471-ab95-284c4d922eae {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1198.888933] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8af40a0e-6166-4e71-baec-cbc9f9495a80 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1198.918300] env[61923]: DEBUG nova.compute.resource_tracker [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181437MB free_disk=178GB free_vcpus=48 pci_devices=None {{(pid=61923) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1198.918438] env[61923]: DEBUG oslo_concurrency.lockutils [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1198.918627] env[61923]: DEBUG oslo_concurrency.lockutils [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1199.939437] env[61923]: DEBUG nova.compute.resource_tracker [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Total usable vcpus: 48, total allocated vcpus: 0 {{(pid=61923) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1199.939749] env[61923]: DEBUG nova.compute.resource_tracker [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=512MB phys_disk=200GB used_disk=0GB total_vcpus=48 used_vcpus=0 pci_stats=[] {{(pid=61923) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1199.954761] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6cf66ad9-991d-498c-bb65-22e4cf5dc3d4 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1199.965304] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38537d96-7364-4f10-9fed-7dfa7c15918b {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1200.012979] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1e85ac1-8596-4922-86a4-bd40b4729d50 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1200.023420] env[61923]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8d7eccc-651c-4142-af4b-140816b7df78 {{(pid=61923) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1200.043442] env[61923]: DEBUG nova.compute.provider_tree [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Inventory has not changed in ProviderTree for provider: f81803f2-f7f8-4939-a757-a77d34a1d0a2 {{(pid=61923) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1200.547086] env[61923]: DEBUG nova.scheduler.client.report [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Inventory has not changed for provider f81803f2-f7f8-4939-a757-a77d34a1d0a2 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 178, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=61923) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1201.053371] env[61923]: DEBUG nova.compute.resource_tracker [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=61923) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1201.053793] env[61923]: DEBUG oslo_concurrency.lockutils [None req-f62e2b35-cc90-406e-aab5-380a4d1a502e None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.135s {{(pid=61923) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}}